Nothing Special   »   [go: up one dir, main page]

US20060265703A1 - Computer architecture and method of operation for multi-computer distributed processing with replicated memory - Google Patents

Computer architecture and method of operation for multi-computer distributed processing with replicated memory Download PDF

Info

Publication number
US20060265703A1
US20060265703A1 US11/259,634 US25963405A US2006265703A1 US 20060265703 A1 US20060265703 A1 US 20060265703A1 US 25963405 A US25963405 A US 25963405A US 2006265703 A1 US2006265703 A1 US 2006265703A1
Authority
US
United States
Prior art keywords
computers
computer
application program
memory
code
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/259,634
Inventor
John Holt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Waratek Pty Ltd
Original Assignee
Waratek Pty Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from AU2005902027A external-priority patent/AU2005902027A0/en
Application filed by Waratek Pty Ltd filed Critical Waratek Pty Ltd
Assigned to WARATEK PTY LTD reassignment WARATEK PTY LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HOLT, JOHN M
Publication of US20060265703A1 publication Critical patent/US20060265703A1/en
Priority to US12/396,446 priority Critical patent/US7860829B2/en
Priority to US12/820,758 priority patent/US20100262590A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F15/00Digital computers in general; Data processing equipment in general
    • G06F15/16Combinations of two or more digital computers each having at least an arithmetic unit, a program unit and a register, e.g. for a simultaneous processing of several programs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/40Transformation of program code
    • G06F8/41Compilation
    • G06F8/45Exploiting coarse grain parallelism in compilation, i.e. parallelism between groups of instructions
    • G06F8/456Parallelism detection

Definitions

  • the present invention relates to computers and other computing machines and information appliances, in particular, to a modified computer architecture and program structure which enables the operation of an application program concurrently or simultaneously on a plurality of computers interconnected via a communications link using a distributed runtime and enables improved performance to be achieved.
  • single prior art machine 1 is made up from a central processing unit, or CPU, 2 which is connected to a memory 3 via a bus 4 . Also connected to the bus 4 are various other functional units of the single machine 1 such as a screen 5 , keyboard 6 and mouse 7 .
  • a fundamental limit to the performance of the machine 1 is that the data to be manipulated by the CPU 2 , and the results of those manipulations, must be moved by the bus 4 .
  • the bus 4 suffers from a number of problems including so called bus “queues” formed by units wishing to gain an access to the bus, contention problems, and the like. These problems can, to some extent, be alleviated by various stratagems including cache memory, however, such stratagems invariably increase the administrative overhead of the machine 1 .
  • FIG. 3 A further possibility of increased computer power through the use of a plural number of machines arises from the prior art concept of distributed computing which is schematically illustrated in FIG. 3 .
  • a single application program (Ap) is partitioned by its author (or another programmer who has become familiar with the application program) into various discrete tasks so as to run upon, say, three machines in which case n in FIG. 3 is the integer 3 .
  • the intention here is that each of the machines M 1 . . . M 3 runs a different third of the entire application and the intention is that the loads applied to the various machines be approximately equal.
  • the machines communicate via a network 14 which can be provided in various forms such as a communications link, the internet, intranets, local area networks, and the like.
  • the speed of operation of such networks 14 is an order of magnitude slower than the speed of operation of the bus 4 in each of the individual machines M 1 , M 2 , . . . , Mn.
  • Distributed computing suffers from a number of disadvantages. Firstly, it is a difficult job to partition the application and this must be done manually. Secondly, communicating data, partial results, results and the like over the network 14 is an administrative overhead. Thirdly, the need for partitioning makes it extremely difficult to scale upwardly by utilising more machines since the application having been partitioned into, say three, does not run well upon four machines. Fourthly, in the event that one of the machines should become disabled, the overall performance of the entire system is substantially degraded.
  • a further prior art arrangement is known as network computing via “clusters” as is schematically illustrated in FIG. 4 .
  • the entire application is loaded onto each of the machines M 1 , M 2 , . . . , Mn.
  • Each machine communicates with a common database but does not communicate directly with the other machines.
  • each machine runs the same application, each machine is doing a different “job” and uses only its own memory. This is somewhat analogous to a number of windows each of which sell train tickets to the public.
  • This approach does operate, is scalable and mainly suffers from the disadvantage that it is difficult to administer the network.
  • Synchronization more generally refers to the exclusive use of an object, class, resource, structure, or other asset to avoid contention between and among computers or machines. This is achieved in JAVA by the “monitor enter” and “monitor exit” instructions or routines. Other languages use different terms but utilize a similar concept.
  • the present invention discloses a computing environment in which an application program operates simultaneously on a plurality of computers.
  • an application program operates simultaneously on a plurality of computers.
  • a multiple computer system having at least one application program each written to operate on only a single computer but running simultaneously on a plurality of computers interconnected by a communications network, wherein different portions of said application program(s) execute substantially simultaneously on different ones of said computers, wherein each computer has an independent local memory accessible only by the corresponding portion of said application program(s) and wherein for each said portion a like plurality of substantially identical objects are created, each in the corresponding computer.
  • a plurality of computers interconnected via a communications link and each having an independent local memory and substantially simultaneously operating a different portion at least one application program each written to operate on only a single computer, each local memory being accessible only by the corresponding portion of said application program.
  • a seventh aspect of the present invention there is disclosed in a multiple thread processing computer operation in which individual threads of a single application program written to operate on only a single computer are simultaneously being processed each on a different corresponding one of a plurality of computers each having an independent local memory accessible only by the corresponding thread and each being interconnected via a communications link, the improvement comprising communicating changes in the contents of local memory physically associated with the computer processing each thread to the local memory of each other said computer via said communications link.
  • a computer program product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the above-mentioned methods.
  • a distributed run time and distributed run time system adapted to enable communications between a plurality of computers, computing machines, or information appliances.
  • a modifier, modifier means, and modifier routine for modifying an application program written to execute on a single computer or computing machine at a time to execute simultaneously on a plurality of networked computers or computing machines.
  • distributed run time and distributed run time system adapted to enable communications between a plurality of computers, computing machines, or information appliances.
  • a computer program and computer program product written to operate on only a single computer but product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the abovementioned procedures, routines, and methods.
  • FIG. 1 is a schematic view of the internal architecture of a conventional computer
  • FIG. 2 is a schematic illustration showing the internal architecture of known symmetric multiple processors
  • FIG. 3 is a schematic representation of prior art distributed computing
  • FIG. 4 is a schematic representation of a prior art network computing using clusters
  • FIG. 5 is a schematic block diagram of a plurality of machines operating the same application program in accordance with a first embodiment of the present invention
  • FIG. 6 is a schematic illustration of a prior art computer arranged to operate JAVA code and thereby constitute a JAVA virtual machine
  • FIG. 7 is a drawing similar to FIG. 6 but illustrating the initial loading of code in accordance with the preferred embodiment
  • FIG. 8 is a drawing similar to FIG. 5 but illustrating the interconnection of a plurality of computers each operating JAVA code in the manner illustrated in FIG. 7 ;
  • FIG. 9 is a flow chart of the procedure followed during loading of the same application on each machine in the network.
  • FIG. 10 is a flow chart showing a modified procedure similar to that of FIG. 9 ;
  • FIG. 11 is a schematic representation of multiple thread processing carried out on the machines of FIG. 8 utilizing a first embodiment of memory updating;
  • FIG. 12 is a schematic representation similar to FIG. 11 but illustrating an alternative embodiment
  • FIG. 13 illustrates multi-thread memory updating for the computers of FIG. 8 ;
  • FIG. 14 is a schematic representation of two laptop computers interconnected to simultaneously run a plurality of applications, with both applications running on a single computer;
  • FIG. 15 is a view similar to FIG. 14 but showing the FIG. 14 apparatus with one application operating on each computer;
  • FIG. 16 is a view similar to FIGS. 14 and 15 but showing the FIG. 14 apparatus with both applications operating simultaneously on both computers;
  • FIG. 17 is a drawing showing an alternative to similar to FIG. 7 .
  • Annexure A provides exemplary actual program or code fragments which implement various aspects of the described embodiments. Although aspects of the invention are described throughout the specification including the Annexes, drawings, and claims, it may be appreciated that Annexure A relates primarily to fields. More particularly, the accompanying Annexure is provided in which: Annexures A1-A10 illustrate exemplary code to illustrate embodiments of the invention in relation to fields.
  • the present invention discloses a modified computer architecture which enables an applications program to be run simultaneously on a plurality of computers in a manner that overcomes the limitations of the aforedescribed conventional architectures, systems, methods, and computer programs.
  • shared memory at each computer may be updated with amendments and/or overwrites so that all memory read requests are satisfied locally.
  • instructions which result in memory being re-written or manipulated are identified. Additional instructions are inserted into the program code (or other modification made) to cause the equivalent memory locations at all computers to be updated. While the invention is not limited to JAVA language or virtual machines, exemplary embodiments are described relative to the JAVA language and standards.
  • the initialization of JAVA language classes and objects are provided for so all memory locations for all computers are initialized in the same manner.
  • the finalization of JAVA language classes and objects is also provide so finalization only occurs when the last class or object present on all machines is no longer required.
  • synchronization is provided such that instructions which result in the application program acquiring (or releasing) a lock on a particular asset (synchronization) are identified. Additional instructions are inserted (or other code modifications performed) to result in a modified synchronization routine with which all computers are updated.
  • one of the features of the invention is to make it appear that one common application program or application code and its executable version (with likely modification) is simultaneously or concurrently executing across a plurality of computers or machines M 1 , . . . , Mn.
  • the instant invention achieves this by running the same application program (for example, Microsoft Word or Adobe Photoshop CS2) on each machine, but modifying the executable code of that application program on each machine as necessary such that each executing instance (‘copy’) on each machine coordinates its local operations on any particular machine with the operations of the respective instances on the other machines such that they all function together in a consistent, coherent and coordinated manner and give the appearance of being one global instance of the application (i.e., a “meta-application”).
  • an application program for example, Microsoft Word or Adobe Photoshop CS2
  • a single application code 50 (sometimes more informally referred to as the application or the application program) can be operated simultaneously on a number of machines M 1 , M 2 . . . Mn interconnected via a communications network or other communications link or path 53 .
  • the communications network or path may be any electronic signaling, data, or digital communications network or path and may advantageously be a relatively slow speed communications path, such as a network connection over the Internet or any common networking configurations known or available as of the date or this applications, and extensions and improvements, thereto.
  • one application code or program 50 may be a single application on the machines, such as Microsoft Word, as opposed to different applications on each machine, such as Microsoft Word on machine M 1 , and Microsoft PowerPoint on machine M 2 , and Netscape Navigator on machine M 3 and so forth. Therefore the terminology “one” application code or program and a “common” application code or program is used to try and capture this situation where all machines M 1 , . . . , Mn are operating or executing the same program or code and not different (and unrelated) programs. In other words copies or replicas of same or substantially the same application code is loaded onto each of the interoperating and connected machines or computers.
  • each of the machines M 1 , M 2 . . . Mn operates with the same application code 50 on each machine M 1 , M 2 . . . , Mn and thus all of the machines M 1 , M 2 , . . . , Mn have the same or substantially the same application code 50 usually with a modification that may be machine specific.
  • each of the machines M 1 , M 2 , . . . , Mn operates with the same (or substantially the same or similar) modifier 51 (in some embodiments implemented as a distributed run time or DRT 71 ) on each machine M 1 , M 2 , . . . , Mn and thus all of the machines M 1 , M 2 . . . Mn have the same (or substantially the same or similar) modifier 51 for each modification required.
  • Different modification for example may be required for memory management and replication, initialization, finalization, and/or synchronization (though not all of these modification types may be required for all embodiments).
  • each application code 50 has been modified by the corresponding modifier 51 according to the same rules (or substantially the same rules since minor optimizing changes are permitted within each modifier 51 / 1 , 51 / 2 , . . . , 51 /n).
  • such memory management modifier 51 -M or DRT 71 -M or other code modifying means component of the overall modifier or distributed run time means is responsible for creating or replicating a memory structure and contents on each of the individual machines M 1 , M 2 . . . Mn that permits the plurality of machines to interoperate.
  • this replicated memory structure will be identical, in other embodiments this memory structure will have portions that are identical and other portions that are not, and in still other embodiments the memory structures are or may not be identical.
  • a single application code 50 (sometimes more informally referred to as the application or the application program) can be operated simultaneously on a number of machines M 1 , M 2 . . . Mn interconnected via a communications network or other communications link or path 53 .
  • one application code or program 50 would be a single common application program on the machines, such as Microsoft Word, as opposed to different applications on each machine, such as Microsoft Word on machine M 1 , and Microsoft PowerPoint on machine M 2 , and Netscape Navigator on machine M 3 and so forth. Therefore the terminology “one”, “single”, and “common” application code or program is used to try and capture this situation where all machines M 1 , .
  • each of the machines M 1 , M 2 . . . Mn operates with the same application code 50 on each machine M 1 , M 2 . . . Mn and thus all of the machines M 1 , M 2 , . . . , Mn have the same or substantially the same application code 50 usually with a modification that may be machine specific.
  • each of the machines M 1 , M 2 , . . . , Mn operates with the same (or substantially the same or similar) modifier 51 on each machine M 1 , M 2 , . . . , Mn and thus all of the machines M 1 , M 2 . . . Mn have the same (or substantially the same or similar) modifier 51 with the modifier of machine M 1 being designated 51 / 1 and the modifier of machine M 2 being designated 51 / 2 , etc.
  • the application code 50 on each machine M 1 , M 2 . . . Mn is modified by the corresponding modifier 51 according to the same rules (or substantially the same rules since minor optimizing changes are permitted within each modifier 51 / 1 , 51 / 2 , . . . , 51 /n).
  • one of the features of the invention is to make it appear that one application program instance of application code 50 is executing simultaneously across all of the plurality of machines M 1 , M 2 , . . . , Mn.
  • the instant invention achieves this by running the same application program code (for example, Microsoft Word or Adobe Photoshop CS2) on each machine, but modifying the executable code of that application program on each machine such that each executing occurrence (or ‘local instance’) on each one of the machines M 1 . . .
  • Mn coordinates its local operations with the operations of the respective occurrences on each one of the other machines such that each occurrence on each one of the plurality of machines function together in a consistent, coherent and coordinated manner so as to give the appearance of being one global instance (or occurrence) of the application program and program code (i.e., a “meta-application”).
  • each of the machines M 1 , M 2 , . . . , Mn has, say, an internal memory capability of 10 MB
  • the total memory available to each application code 50 is not necessarily, as one might expect the number of machines (n) times 10 MB, or alternatively the additive combination of the internal memory capability of all n machines, but rather or still may only be 10 MB.
  • the size of the smallest memory of any of the machines may be used as the maximum memory capacity of the machines when such memory (or a portion thereof) is to be treated as a ‘common’ memory (i.e. similar equivalent memory on each of the machines M 1 . . . Mn) or otherwise used to execute the common application code.
  • each machine M 1 , M 2 . . . Mn has an private (i.e. ‘non-common’) internal memory capability.
  • the private internal memory capability of the machines M 1 , M 2 , . . . , Mn are normally approximately equal but need not be. It may also be advantageous to select the amounts of internal memory in each machine to achieve a desired performance level in each machine and across a constellation or network of connected or coupled plurality of machines, computers, or information appliances M 1 , M 2 , . . . , Mn. Having described these internal and common memory considerations, it will be apparent in light of the description provided herein that the amount of memory that can be common between machines is not a limitation of the invention.
  • FIG. 6 It is known from the prior art to operate a single computer or machine (produced by one of various manufacturers and having an operating system operating in one of various different languages) in a particular language of the application, by creating a virtual machine as schematically illustrated in FIG. 6 .
  • the code and data and virtual machine configuration or arrangement of FIG. 6 takes the form of the application code 50 written in the Java language and executing within a Java Virtual Machine 61 .
  • the intended language of the application is the language JAVA
  • a JAVA virtual machine is used which is able to operate code in JAVA irrespective of the machine manufacturer and internal details of the machine.
  • the JAVA Virtual Machine Specification 2 nd Edition by T. Lindholm & F. Yellin of Sun Microsystems Inc. of the USA, which is incorporated by reference herein.
  • FIG. 6 This conventional art arrangement of FIG. 6 is modified in accordance with embodiments of the present invention by the provision of an additional facility which is conveniently termed “distributed run time” or “distributed run time system” DRT 71 and as seen in FIG. 7 .
  • An alternative showing the modifier 171 and loading 175 code 50 is in FIG. 17 .
  • the application code 50 is loaded onto the Java Virtual Machine 72 in cooperation with the distributed runtime system 71 , through the loading procedure indicated by arrow 75 .
  • distributed runtime and the distributed run time system are essentially synonymous, and by means of illustration but not limitation are generally understood to include library code and processes which support software written in a particular language running on a particular platform. Additionally, a distributed runtime system may also include library code and processes which support software written in a particular language running within a particular distributed computing environment.
  • the runtime system typically deals with the details of the interface between the program and the operation system such as system calls, program start-up and termination, and memory management.
  • a conventional Distributed Computing Environment that does not provide the capabilities of the inventive distributed run time or distributed run time system 71 required in the invention is available from the Open Software Foundation.
  • This Distributed Computing Environment performs a form of computer-to-computer communication for software running on the machines, but among its many limitations, it is not able to implement the modification or communication operations of this invention.
  • the inventive DRT 71 coordinates the particular communications between the plurality of machines M 1 , M 2 , . . . , Mn.
  • inventive distributed runtime 71 comes into operation during the loading procedure indicated by arrow 75 of the JAVA application 50 on each JAVA virtual machine 72 of machines JVM# 1 , JVM# 2 , . . . JVM#n.
  • the sequence of operations during loading will be described hereafter in relation to FIG. 9 . It will be appreciated in light of the description provided herein that although many examples and descriptions are provided relative to the JAVA language and JAVA virtual machines so that the reader may get the benefit of specific examples, the invention is not restricted to either the JAVA language or JAVA virtual machines, or to any other language, virtual machine, machine, or operating environment.
  • FIG. 8 shows in modified form the arrangement of FIG. 5 utilising JAVA virtual machines, each as illustrated in FIG. 7 .
  • the same application code 50 is loaded onto each machine M 1 , M 2 . . . Mn.
  • the communications between each machine M 1 , M 2 , . . . , Mn, and indicated by arrows 83 are advantageously controlled by the individual DRT's 71 / 1 . . . 71 /n within each machine.
  • this may be conceptionalised as the DRT's 71 / 1 , . . .
  • the inventive DRT 71 provides communication that is transport, protocol, and link independent.
  • the modifier 51 may be implemented as a component of or within the distributed run time 71 , and therefore the DRT 71 may implement the functions and operations of the modifier 51 .
  • the function and operation of the modifier 51 may be implemented outside of the structure, software, firmware, or other means used to implement the DRT 71 .
  • the modifier 51 and DRT 71 are implemented or written in a single piece of computer program code that provides the functions of the DRT and modifier. The modifier function and structure therefore maybe subsumed into the DRT and considered to be an optional component.
  • the modifier function and structure is responsible for modifying the executable code of the application code program
  • the distributed run time function and structure is responsible for implementing communications between and among the computers or machines.
  • the communications functionality in one embodiment is implemented via an intermediary protocol layer within the computer program code of the DRT on each machine.
  • the DRT may for example implement a communications stack in the JAVA language and use the Transmission Control Protocol/Internet Protocol (TCP/IP) to provide for communications or talking between the machines.
  • TCP/IP Transmission Control Protocol/Internet Protocol
  • a plurality of individual computers or machines M 1 , M 2 , . . . , Mn are provided, each of which are interconnected via a communications network 53 or other communications link and each of which individual computers or machines provided with a modifier 51 (See in FIG. 5 ) and realised by or in for example the distributed run time (DRT) 71 (See FIG. 8 ) and loaded with a common application code 50 .
  • the term common application program is to be understood to mean an application program or application program code written to operate on a single machine, and loaded and/or executed in whole or in part on each one of the plurality of computers or machines M 1 , M 2 . . .
  • the modifier 51 or DRT 71 or other code modifying means is responsible for modifying the application code 50 so that it may execute memory manipulation operations, such as memory putstatic and puffield instructions in the JAVA language and virtual machine environment, in a coordinated, consistent, and coherent manner across and between the plurality of individual machines M 1 . . . Mn. It follows therefore that in such a computing environment it is necessary to ensure that each of memory location is manipulated in a consistent fashion (with respect to the others).
  • some or all of the plurality of individual computers or machines may be contained within a single housing or chassis (such as so-called “blade servers” manufactured by Hewlett-Packard Development Company, Intel Corporation, IBM Corporation and others) or implemented on a single printed circuit board or even within a single chip or chip set.
  • blade servers manufactured by Hewlett-Packard Development Company, Intel Corporation, IBM Corporation and others
  • a machine (produced by any one of various manufacturers and having an operating system operating in any one of various different languages) can operate in the particular language of the application program code 50 , in this instance the JAVA language. That is, a JAVA virtual machine 72 is able to operate application code 50 in the JAVA language, and utilize the JAVA architecture irrespective of the machine manufacturer and the internal details of the machine.
  • the generalized platform, and/or virtual machine and/or machine and/or runtime system is able to operate application code 50 in the language(s) (possibly including for example, but not limited to any one or more of source-code languages, intermediate-code languages, object-code languages, machine-code languages, and any other code languages) of that platform, and/or virtual machine and/or machine and/or runtime system environment, and utilize the platform, and/or virtual machine and/or machine and/or runtime system and/or language architecture irrespective of the machine manufacturer and the internal details of the machine.
  • platform and/or runtime system may include virtual machine and non-virtual machine software and/or firmware architectures, as well as hardware and direct hardware coded applications and implementations.
  • computers and/or computing machines and/or information appliances or processing systems are still applicable.
  • computers and/or computing machines that do not utilize either classes and/or objects include for example, the x86 computer architecture manufactured by Intel Corporation and others, the SPARC computer architecture manufactured by Sun Microsystems, Inc and others, the PowerPC computer architecture manufactured by International Business Machines Corporation and others, and the personal computer products made by Apple Computer, Inc., and others.
  • primitive data types such as integer data types, floating point data types, long data types, double data types, string data types, character data types and Boolean data types
  • structured data types such as arrays and records
  • code or data structures of procedural languages or other languages and environments such as functions, pointers, components, modules, structures, references and unions.
  • Step 90 commences at Step 90 in FIG. 9 and involves the initial step 91 of preferably scrutinizing or analysing the code and detecting all memory locations addressable by the application code 50 , or optionally some subset of all memory locations addressable by the application code 50 ; such as for example named and unnamed memory locations, variables (such as local variables, global variables, and formal arguments to subroutines or functions), fields, registers, or any other address space or range of addresses which application code 50 may access.
  • variables such as local variables, global variables, and formal arguments to subroutines or functions
  • fields registers, or any other address space or range of addresses which application code 50 may access.
  • Such memory locations in some instances need to be identified for subsequent processing at steps 92 and 93 .
  • the DRT 71 during the loading procedure 75 creates a list of all the memory locations thus identified.
  • the memory locations in the form of JAVA fields are listed by object and class, however, the memory locations, fields, or the like may be listed or organized in any manner so long as they comport with the architectural and programming requirements of the system on which the program is to be used and the principles of the invention described herein. This detection is optional and not required in all embodiments of the invention. It may be noted that the DRT is at least in part fulfilling the roll of the modifier 51 .
  • Step 92 The next phase (designated Step 92 in FIG. 9 ) [Step 92 ] of the modification procedure is to search through the application code 50 in order to locate processing activity or activities that manipulate or change values or contents of any listed memory location (for example, but not limited to JAVA fields) corresponding to the list generated at step 91 when required. Preferably, all processing activities that manipulate or change any one or more values or contents of any one or more listed memory locations, are located.
  • processing activity or operation typically “putstatic” or “puffield” in the JAVA language, or for example, a memory assignment operation, or a memory write operation, or a memory manipulation operation, or more generally operations that otherwise manipulate or change value(s) or content(s) of memory or other addressable areas
  • an “updating propagation routine” is inserted by step 93 in the application code 50 corresponding to the detected memory manipulation operation, to communicate with all other machines in order to notify all other machines of the identity of the manipulated memory location, and the updated, manipulated or changed value(s) or content(s) of the manipulated memory location.
  • the inserted “updating propagation routine” preferably takes the form of a method, function, procedure, or similar subroutine call or operation to a network communications library of DRT 71 .
  • the “updating propagation routine” may take the optional form of a code-block (or other inline code form) inserted into the application code instruction stream at, after, before, or otherwise corresponding to the detected manipulation instruction or operation.
  • the “updating propagation routine” may execute on the same thread or process or processor as the detected memory manipulation operation of step 92 . Thereafter, the loading procedure continues, by loading the modified application code 50 on the machine 72 in place of the unmodified application code 50 , as indicated by step 94 in FIG. 9 .
  • FIG. 10 An alternative form of modification during loading is illustrated in the illustration of FIG. 10 .
  • the start and listing steps 90 and 91 and the searching step 92 are the same as in FIG. 9 .
  • the “updating propagation routine” is inserted into the application code 50 corresponding to the detected memory manipulation operation identified in step 92 , as is indicated in step 93 , in which the application code 50 , or network communications library code 71 of the DRT executing on the same thread or process or processor as the detected memory manipulation operation, carries out the updating, instead an “alert routine” is inserted corresponding to the detected memory manipulation operation, at step 103 .
  • the “alert routine” instructs, notifies or otherwise requests a different and potentially simultaneously or concurrently executing thread or process or processor not used to perform the memory manipulation operation (that is, a different thread or process or processor than the thread or process or processor which manipulated the memory location), such as a different thread or process allocated to the DRT 71 , to carry out the notification, propagation, or communication of all other machines of the identity of the manipulated memory location, and the updated, manipulated or changed value(s) or content(s) of the manipulated memory location.
  • FIG. 11 corresponds to the execution and operation of the modified application code 50 when modified in accordance with the procedures set forth in and described relative to FIG. 9 .
  • FIG. 12 on the other hand (and the steps 112 , 113 , 125 , 127 , and 115 therein) set forth therein correspond to the execution and operation of the modified application code 50 when modified in accordance with FIG. 10 .
  • This analysis or scrutiny of the application code 50 can take place either prior to loading the application program code 50 , or during the application program code 50 loading procedure, or even after the application program code 50 loading procedure. It may be likened to an instrumentation, program transformation, translation, or compilation procedure in that the application code may be instrumented with additional instructions, and/or otherwise modified by meaning-preserving program manipulations, and/or optionally translated from an input code language to a different code language (such as for example from source-code language or intermediate-code language to object-code language or machine-code language), and with the understanding that the term compilation normally or conventionally involves a change in code or language, for example, from source code to object code or from one language to another language.
  • compilation (and its grammatical equivalents) is not so restricted and can also include or embrace modifications within the same code or language.
  • the compilation and its equivalents are understood to encompass both ordinary compilation (such as for example by way of illustration but not limitation, from source-code to object-code), and compilation from source-code to source-code, as well as compilation from object-code to object-code, and any altered combinations therein. It is also inclusive of so-called “intermediary-code languages” which are a form of “pseudo object-code”.
  • the analysis or scrutiny of the application code 50 may take place during the loading of the application program code such as by the operating system reading the application code from the hard disk or other storage device or source and copying it into memory and preparing to begin execution of the application program code.
  • the analysis or scrutiny may take place during the class loading procedure of the java.lang.ClassLoader loadClass method (e.g., “java.lang.ClassLoader.loadClass( )”).
  • the analysis or scrutiny of the application code 50 may take place even after the application program code loading procedure, such as after the operating system has loaded the application code into memory, or optionally even after execution of the relevant corresponding portion of the application program code has started, such as for example after the JAVA virtual machine has loaded the application code into the virtual machine via the “java.lang.ClassLoader.loadClass( )” method and optionally commenced execution.
  • a multiple thread processing machine environment 110 on each one of the machines M 1 , . . . , Mn and consisting of threads 111 / 1 . . . 111 / 4 exists.
  • the processing and execution of the second thread 111 / 2 results in that thread 111 / 2 manipulating a memory location at step 113 , by writing to a listed memory location.
  • the application code 50 is modified at a point corresponding to the write to the memory location of step 113 , so that it propagates, notifies, or communicates the identity and changed value of the manipulated memory location of step 113 to the other machines M 2 , . . . , Mn via network 53 or other communication link or path, as indicated at step 114 .
  • the processing of the application code 50 of that thread 111 / 2 is or may be altered and in some instances interrupted at step 114 by the executing of the inserted “updating propagation routine”, and the same thread 111 / 2 notifies, or propagates, or communicates to all other machines M 2 , . . .
  • the thread 111 / 2 then resumes or continues the processing or the execution of the modified application code 50 at step 115 .
  • a multiple thread processing machine environment 110 comprising or consisting of threads 111 / 1 , . . . , 111 / 3 , and a simultaneously or concurrently executing DRT processing environment 120 consisting of the thread 121 / 1 as illustrated, or optionally a plurality of threads, is executing on each one of the machines M 1 , . . . Mn.
  • the processing and execution of the modified application code 50 on thread 111 / 2 results in a memory manipulation operation of step 113 , which in this instance is a write to a listed memory location.
  • the application code 50 is modified at a point corresponding to the write to the memory location of step 113 , so that it requests or otherwise notifies the threads of the DRT processing environment 120 to notify, or propagate, or communicate to the other machines M 2 , . . . , Mn of the identity and changed value of the manipulated memory location of step 113 , as indicated at steps 125 and 128 and arrow 127 .
  • the thread 111 / 2 processing and executing the modified application code 50 requests a different and potentially simultaneously or concurrently executing thread or process (such as thread 121 / 1 ) of the DRT processing environment 120 to notify the machines M 2 , . . .
  • step 125 and arrow 127 Mn via network 53 or other communications link or path of the identity and changed value of the manipulated memory location of step 113 , as indicated in step 125 and arrow 127 .
  • a different and potentially simultaneously or concurrently executing thread or process 121 / 1 of the DRT processing environment 120 notifies the machines M 2 , . . . , Mn via network 53 or other communications link or path of the identity and changed value of the manipulated memory location of step 113 , as requested of it by the modified application code 50 executing on thread 111 / 2 of step 125 and arrow 127 .
  • step 125 of thread 111 / 2 of FIG. 12 can be carried out quickly, because step 114 of thread 111 / 2 must notify and communicate with machines M 2 , . . . , Mn via the relatively slow network 53 (relatively slow for example when compared to the internal memory bus 4 of FIG. 1 or the global memory 13 of FIG. 2 ) of the identity and changed value of the manipulated memory location of step 113 , whereas step 125 of thread 111 / 2 does not communicate with machines M 2 , . . . , Mn via the relatively slow network 53 .
  • step 125 of thread 111 / 2 requests or otherwise notifies a different and potentially simultaneously or concurrently executing thread 121 / 1 of the DRT processing environment 120 to perform the notification and communication with machines M 2 , . . . , Mn via the relatively slow network 53 of the identify and changed value of the manipulated memory location of step 113 , as indicated by arrow 127 .
  • thread 111 / 2 carrying out step 125 is only interrupted momentarily before the thread 111 / 2 resumes or continues processing or execution of modified application code in step 115 .
  • the other thread 121 / 1 of the DRT processing environment 120 then communicates the identity and changed value of the manipulated memory location of step 113 to machines M 2 , . . . , Mn via the relatively slow network 53 or other relatively slow communications link or path.
  • This second arrangement of FIG. 12 makes better utilisation of the processing power of the various threads 111 / 1 . . . 111 / 3 and 121 / 1 (which are not, in general, subject to equal demands). Irrespective of which arrangement is used, the identity and change value of the manipulated memory location(s) of step 113 is (are) propagated to all the other machines M 2 . . . Mn on the network 53 or other communications link or path.
  • step 114 of FIG. 11 or the DRT 71 / 1 (corresponding to the DRT processing environment 120 of FIG. 12 ) and its thread 121 / 1 of FIG. 12 (represented by step 128 in FIG. 13 ), send, via the network 53 or other communications link or path, the identity and changed value of the manipulated memory location of step 113 of FIGS. 11 and 12 , to each of the other machines M 2 , . . . , Mn.
  • each of the other machines M 2 , . . . , Mn carries out the action of receiving from the network 53 the identity and changed value of, for example, the manipulated memory location of step 113 from machine M 1 , indicated by step 135 , and writes the value received at step 135 to the local memory location corresponding to the identified memory location received at step 135 , indicated by step 136 .
  • Such local memory read and write processing operation as performed according to the invention can typically be satisfied within 10 2 -10 3 cycles of the central processing unit.
  • a relatively slow network communication link or path 53 may advantageously be used because it provides the desired performance and low cost
  • the invention is not limited to a relatively low speed network connection and may be used with any communication link or path.
  • the invention is transport, network, and communications path independent, and does not depend on how the communication between machines or DRTs takes place. In one embodiment, even electronic mail (email) exchanges between machines or DRTs may suffice for the communications.
  • the identity and changed value pair of a manipulated memory location sent over network 53 each pair typically sent as the sole contents of a single packet, frame or cell for example, can be grouped into batches of multiple pairs of identities and changed values corresponding to multiple manipulated memory locations, and sent together over network 53 or other communications link or path in a single packet, frame, or cell.
  • This further modification further reduces the demands on the communication speed of the network 53 or other communications link or path interconnecting the various machines, as each packet, cell or frame may contain multiple identity and changed value pairs, and therefore fewer packets, frames, or cells require to be sent.
  • the embodiment illustrated of FIG. 11 of step 114 sends an updating and propagation message to all machines corresponding to every performed memory manipulation operation.
  • the DRT thread 121 / 1 of FIG. 12 does not need to perform an updating and propagation operation corresponding to every local memory manipulation operation, but instead may send fewer updating and propagation messages than memory manipulation operations, each message containing the last or latest changed value or content of the manipulated memory location, or optionally may only send a single updating and propagation message corresponding to the last memory manipulation operation.
  • This further improvement reduces the demands on the network 53 or other communications link or path, as fewer packets, frames, or cells require to be sent.
  • each DRT 71 when initially recording or creating the list of all, or some subset of all, memory locations (or fields), for each such recorded memory location on each machine M 1 , . . . , Mn there is a name or identity which is common or similar on each of the machines M 1 , . . . , Mn.
  • the local memory location corresponding to a given name or identity (listed for example, during step 91 of FIG. 9 ) will or may vary over time since each machine may and generally will store changed memory values or contents at different memory locations according to its own internal processes.
  • each of the DRTs will have, in general, different local memory locations corresponding to a single memory name or identity, but each global “memory name” or identity will have the same “memory value” stored in the different local memory locations.
  • a particular machine say machine M 2 , loads the asset (such as class or object) inclusive of memory manipulation operation(s), modifies it, and then loads each of the other machines M 1 , M 3 , . . . , Mn (either sequentially or simultaneously or according to any other order, routine or procedure) with the modified object (or class or other asset or resource) inclusive of the new modified memory manipulation operation.
  • asset such as class or object
  • Mn either sequentially or simultaneously or according to any other order, routine or procedure
  • the memory manipulation operation(s) that is (are) loaded is binary executable object code.
  • the memory manipulation operation(s) that is (are) loaded is executable intermediary code.
  • each of the slave (or secondary) machines M 1 , M 3 , . . . , Mn loads the modified object (or class), and inclusive of the new modified memory manipulation operation(s), that was sent to it over the computer communications network or other communications link or path by the master (or primary) machine, such as machine M 2 , or some other machine such as a machine X of FIG. 15 .
  • the computer communications network can be replaced by a shared storage device such as a shared file system, or a shared document/file repository such as a shared database.
  • each machine or computer need not and frequently will not be the same or identical. What is required is that they are modified in a similar enough way that in accordance with the inventive principles described herein, each of the plurality of machines behaves consistently and coherently relative to the other machines to accomplish the operations and objectives described herein.
  • modifications may for example depend on the particular hardware, architecture, operating system, application program code, or the like or different factors. It will also be appreciated that embodiments of the invention may be implemented within an operating system, outside of or without the benefit of any operating system, inside the virtual machine, in an EPROM, in software, in firmware, or in any combination of these.
  • each machine M 1 , . . . , Mn receives the unmodified asset (such as class or object) inclusive of one or more memory manipulation operation(s), but modifies the operations and then loads the asset (such as class or object) consisting of the now modified operations.
  • asset such as class or object
  • one machine such as the master or primary machine may customize or perform a different modification to the memory manipulation operation(s) sent to each machine, this embodiment more readily enables the modification carried out by each machine to be slightly different and to be enhanced, customized, and/or optimized based upon its particular machine architecture, hardware, processor, memory, configuration, operating system, or other factors, yet still similar, coherent and consistent with other machines with all other similar modifications and characteristics that may not need to be similar or identical.
  • the supply or the communication of the asset code (such as class code or object code) to the machines M 1 , . . . , Mn, and optionally inclusive of a machine X of FIG. 15 can be branched, distributed or communicated among and between the different machines in any combination or permutation; such as by providing direct machine to machine communication (for example, M 2 supplies each of M 1 , M 3 , M 4 , etc. directly), or by providing or using cascaded or sequential communication (for example, M 2 supplies M 1 which then supplies M 3 which then supplies M 4 , and so on), or a combination of the direct and cascaded and/or sequential.
  • direct machine to machine communication for example, M 2 supplies each of M 1 , M 3 , M 4 , etc. directly
  • cascaded or sequential communication for example, M 2 supplies M 1 which then supplies M 3 which then supplies M 4 , and so on
  • Annexure A5 is a typical code fragment from a memory manipulation operation prior to modification (e.g., an exemplary unmodified routine with a memory manipulation operation)
  • Annexure A6 is the same routine with a memory manipulation operation after modification (e.g., an exemplary modified routine with a memory manipulation operation).
  • These code fragments are exemplary only and identify one software code means for performing the modification in an exemplary language. It will be appreciated that other software/firmware or computer program code may be used to accomplish the same or analogous function or operation without departing from the invention.
  • Annexures A5 and A6 are exemplary code listings that set forth the conventional or unmodified computer program software code (such as may be used in a single machine or computer environment) of a routine with a memory manipulation operation of application program code 50 and a post-modification excerpt of the same routine such as may be used in embodiments of the present invention having multiple machines.
  • the modified code that is added to the routine is highlighted in bold text.
  • TABLE I Summary Listing of Contents of Annexure A Annexure A includes exemplary program listings in the JAVA language to further illustrate features, aspects, methods, and procedures of described in the detailed description A1. This first excerpt is part of an illustration of the modification code of the modifier 51 in accordance with steps 92 and 103 of FIG. 10 .
  • This second excerpt is part of the DRT.alert( ) method and implements the step of 125 and arrow of 127 of FIG. 12 .
  • This DRT.alert( ) method requests one or more threads of the DRT processing environment of FIG. 12 to update and propagate the value and identity of the changed memory location corresponding to the operation of Annexure A1.
  • This third excerpt is part of the DRT 71, and corresponds to step 128 of FIG. 12 .
  • This code fragment shows the DRT in a separate thread, such as thread 121/1 of FIG. 12 , after being notified or requested by step 125 and array 127, and sending the changed value and changed value location/identity across the network 53 to the other of the plurality of machines M1...Mn.
  • the fourth excerpt is part of the DRT 71, and corresponds to steps 135 and 136 of FIG. 13 .
  • the fifth excerpt is an disassembled compiled form of the example.java application of Annexure A7, which performs a memory manipulation operation (putstatic and putfield).
  • A6 is an disassembled compiled form of the example.java application of Annexure A7, which performs a memory manipulation operation (putstatic and putfield).
  • the sixth excerpt is the disassembled compiled form of the same example application in Annexure A5 after modification has been performed by FieldLoader.java of Annexure A11, in accordance with FIG. 9 of this invention. The modifications are highlighted in bold.
  • the seventh excerpt is the source-code of the example.java application used in excerpt A5 and A6. This example application has two memory locations (staticValue and instanceValue) and performs two memory manipulation operations.
  • the eighth excerpt is the source-code of FieldAlert.java which corresponds to step 125 and arrow 127 of FIG.
  • the ninth excerpt is the source-code of FieldSend.java which corresponds to step 128 of FIG. 12 , and waits for a request/notification generated by FieldAlert.java of A8 corresponding to step 125 and arrow 127, and which propagates a changed value/identity pair requested of it by FieldAlert.java, via network 53.
  • the tenth excerpt is the source-code of FieldReceive.java, which corresponds to steps 135 and 136 of FIG.
  • FieldLoader.java This excerpt is the source-code of FieldLoader.java, which modifies an application program code, such as the example.java application code of Annexure A7, as it is being loaded into a JAVA virtual machine in accordance with steps 90, 91, 92, 103, and 94 of FIG. 10 .
  • FieldLoader.java makes use of the convenience classes of Annexures A12 through to A36 during the modification of a compiled JAVA A12. Attribute_info.java Convience class for representing attribute_info structures within ClassFiles. A13.
  • ClassFile.java Convience class for representing ClassFile structures.
  • A14. Code_attribute.java Convience class for representing Code_attribute structures within ClassFiles.
  • CONSTANT_Class_info.java Convience class for representing CONSTANT_Class_info structures within ClassFiles.
  • CONSTANT_Double_info.java Convience class for representing CONSTANT_Double_info structures within ClassFiles.
  • CONSTANT_Fieldref_info.java Convience class for representing CONSTANT_Fieldref_info structures within ClassFiles.
  • A18. CONSTANT_Float_info.java Convience class for representing CONSTANT_Float_info structures within ClassFiles. A19.
  • CONSTANT_Integer_info.java Convience class for representing CONSTANT_Integer_info structures within ClassFiles.
  • A20. CONSTANT_InterfaceMethodref_info.java Convience class for representing CONSTANT_InterfaceMethodref_info structures within ClassFiles.
  • A21. CONSTANT_Long_info.java Convience class for representing CONSTANT_Long_info structures within ClassFiles.
  • CONSTANT_Methodref_info.java Convience class for representing CONSTANT_Methodref_info structures within ClassFiles.
  • A23. CONSTANT_NameAndType_info.java Convience class for representing CONSTANT_NameAndType_info structures within ClassFiles.
  • CONSTANT_String_info.java Convience class for representing CONSTANT_String_info structures within ClassFiles.
  • CONSTANT_Utf8_info.java Convience class for representing CONSTANT_Utf8_info structures within ClassFiles.
  • ConstantValue_attribute.java Convience class for representing ConstantValue_attribute structures within ClassFiles.
  • cp_info.java Convience class for representing cp_info structures within ClassFiles.
  • Deprecated_attribute.java Convience class for representing Deprecated_attribute structures within ClassFiles.
  • A29 for representing Deprecated_attribute structures within ClassFiles.
  • Exceptions_attribute.java Convience class for representing Exceptions_attribute structures within ClassFiles.
  • A31. InnerClasses_attribute.java Convience class for representing InnerClasses_attribute structures within ClassFiles.
  • A32. LineNumberTable_attribute.java Convience class for representing LineNumberTable_attribute structures within ClassFiles.
  • LocalVariableTable_attribute.java Convience class for representing LocalVariableTable_attribute structures within ClassFiles.
  • A35. SourceFile_attribute.java Convience class for representing SourceFile_attribute structures within ClassFiles.
  • A36. Synthetic_attribute.java Convience class for representing Synthetic_attribu
  • This first excerpt is part of an illustration of the modification code of the modifier 51 in accordance with steps 92 and 103 of FIG. 10 . It searches through the code array of the application program code 50, and when it detects a memory manipulation instruction (i.e. a putstatic instruction (opcode 178) in the JAVA language and virtual machine environment) it modifies the application program code by the insertion of an “alert” routine.
  • a memory manipulation instruction i.e. a putstatic instruction (opcode 178) in the JAVA language and virtual machine environment
  • This second excerpt is part of the DRT.alert( ) method and implements the step of 125 and arrow of 127 of FIG. 12 .
  • This DRT.alert( ) method requests one or more threads of the DRT processing environment of FIG. 12 to update and propagate the value and identity of the changed memory location corresponding to the operation of Annexure A1.
  • This third excerpt is part of the DRT 71, and corresponds to step 128 of FIG. 12 .
  • This code fragment shows the DRT in a separate thread, such as thread 121/1 of FIG. 12 , after being notified or requested by step 125 and array 127, and sending the changed value and changed value location/identity across the network 53 to the other of the plurality of machines M1...Mn.
  • START MulticastSocket ms DRT.getMulticastSocket( ); // The multicast socket // used by the DRT for // communication.
  • byte nameTag 33; // This is the “name tag” on the network for this // field.
  • Field field modifiedClass.getDeclaredField(“myField1”); // Stores // the field // from the // modified // class. // In this example, the field is a byte field. while (DRT.isRunning( )) ⁇ synchronized (ALERT_LOCK) ⁇ ALERT_LOCK.wait( ); // The DRT thread is waiting for the alert // method to be called.
  • byte[ ] b new byte[ ] ⁇ nameTag, field.getByte(null) ⁇ ; // Stores // the // nameTag // and the // value // of the // field from // the // modified // class in a buffer.
  • DatagramPacket dp new DatagramPacket(b, 0, b.length); ms.send(dp); // Send the buffer out across the network. ⁇ ⁇ // END
  • the compiled code in the annexure and portion repeated in the table is taken from the source-code of the file “example.java” which is included in the Annexure A7 (Table VIII).
  • the procedure name “Method void setValues(int, int)” of Step 001 is the name of the displayed disassembled output of the setValues method of the compiled application code of “example.java”.
  • the name “Method void setvalues(int, int)” is arbitrary and selected for this example to indicate a typical JAVA method inclusive of a memory manipulation operation. Overall the method is responsible for writing two values to two different memory locations through the use of an memory manipulation assignment statement (being “putstatic” and “putfield” in this example) and the steps to accomplish this are described in turn.
  • the Java Virtual Machine instruction “iload — 1” causes the Java Virtual Machine to load the integer value in the local variable array at index 1 of the current method frame and store this item on the top of the stack of the current method frame and results in the integer value passed to this method as the first argument and stored in the local variable array at index 1 being pushed onto the stack.
  • the Java Virtual Machine instruction “putstatic #3 ⁇ Field int staticValue>” causes the Java Virtual Machine to pop the topmost value off the stack of the current method frame and store the value in the static field indicated by the CONSTANT_Fieldref_info constant-pool item stored in the 3 rd index of the classfile structure of the application program containing this example setValues( ) method and results in the topmost integer value of the stack of the current method frame being stored in the integer field named “staticValue”.
  • the Java Virtual Machine instruction “aload — 0” causes the Java Virtual Machine to load the item in the local variable array at index 0 of the current method frame and store this item on the top of the stack of the current method frame and results in the ‘this’ object reference stored in the local variable array at index 0 being pushed onto the stack.
  • the Java Virtual Machine instruction “iload — 2” causes the Java Virtual Machine to load the integer value in the local variable array at index 2 of the current method frame and store this item on the top of the stack of the current method frame and results in the integer value passed to this method as the first argument and stored in the local variable array at index 2 being pushed onto the stack.
  • the Java Virtual Machine instruction “puffield #2 ⁇ Field int instanceValue>” causes the Java Virtual Machine to pop the two topmost values off the stack of the current method frame and store the topmost value in the object instance field of the second popped value, indicated by the CONSTANT_Fieldref_info constant-pool item stored in the 2 nd index of the classfile structure of the application program containing this example setValues method and results in the integer value on the top of the stack of the current method frame being stored in the instance field named “instanceValue” of the object reference below the integer value on the stack.
  • Step 007 causes the JAVA virtual machine to cease executing this setValues( ) method by returning control to the previous method frame and results in termination of execution of this setValues( ) method.
  • the JAVA virtual machine manipulates (i.e. writes to) the staticValue and instanceValue memory locations, and in executing the setValues( ) method containing the memory manipulation operation(s) is able to ensure that memory is and remains consistent between multiple threads of a single application instance, and therefore ensure that unwanted behaviour, such as for example inconsistent or incoherent memory between multiple threads of a single application instance (such inconsistent or incoherent memory being for example incorrect or different values or contents with respect to a single memory location) does not occur.
  • unwanted behaviour such as for example inconsistent or incoherent memory between multiple threads of a single application instance (such inconsistent or incoherent memory being for example incorrect or different values or contents with respect to a single memory location) does not occur.
  • the JAVA virtual machine instruction “iconst — 0” is inserted after the “ldc #4” instruction so that the JAVA virtual machine loads an integer value of “0” onto the stack of the current method frame and results in the integer value of “0” loaded onto the top of the stack of the current method frame.
  • This change is significant because it modifies the setValues( ) method to load an integer value, which in this example is “0”, which represents the identity of the memory location (field) manipulated by the preceding “putstatic #3” operation. It is to be noted that the choice or particular form of the memory identifier used for the implementation of this invention is for illustration purposes only.
  • the integer value of “0” is the identifier used of the manipulated memory location, and corresponds to the “staticValue” field as the first field of the “example.java” application, as shown in Annexure A7. Therefore, corresponding to the “putstatic #3” instruction, the “iconst — 0” instruction loads the integer value “0” corresponding to the index of the manipulated field of the “putstatic #3” instruction, and which in this case is the first field of “example.java” hence the “0” integer index value, onto the stack.
  • the JAVA virtual machine instruction “invokestatic #5 ⁇ Method boolean alert(java.lang.Object, int)>” is inserted after the “iconst — 0” instruction so that the JAVA virtual machine pops the two topmost items off the stack of the current method frame (which in accordance with the preceding “ldc #4” instruction is a reference to the String object with the value “example” corresponding to the name of the class to which manipulated field belongs, and the integer “0” corresponding to the index of the manipulated field in the example.java application) and invokes the “alert” method, passing the two topmost items popped off the stack to the new method frame as its first two arguments.
  • This change is significant because it modifies the setValues( ) method to execute the “alert” method and associated operations, corresponding to the preceding memory manipulation operation (that is, the “putstatic #3” instruction) of the setValues( ) method.
  • an “aload — 0” instruction is inserted after the “puffield #2” instruction in order to be the first instruction following the execution of the “puffield #2” instruction.
  • This change is significant because it modifies the setValues( ) method to load a reference to the object corresponding to the manipulated field onto the stack.
  • the JAVA virtual machine instruction “iconst — 1” is inserted after the “aload — 0” instruction so that the JAVA virtual machine loads an integer value of “1” onto the stack of the current method frame and results in the integer value of “1” loaded onto the top of the stack of the current method frame.
  • This change is significant because it modifies the setValues( ) method to load an integer value, which in this example is “1”, which represents the identity of the memory location (field) manipulated by the preceding “puffield #2” operation. It is to be noted that the choice or particular form of the identifier used for the implementation of this invention is for illustration purposes only.
  • the integer value of “1” corresponds to the “instanceValue” field as the second field of the “example.java” application, as shown in Annexure A7. Therefore, corresponding to the “puffield #2” instruction, the “iconst — 1” instruction loads the integer value “1” corresponding to the index of the manipulated field of the “puffield #2” instruction, and which in this case is the second field of “example.java” hence the “1” integer index value, onto the stack.
  • the JAVA virtual machine instruction “invokestatic #5 ⁇ Method boolean alert(java.lang.Object, int)>” is inserted after the “iconst — 1” instruction so that the JAVA virtual machine pops the two topmost item off the stack of the current method frame (which in accordance with the preceding “aload — 0” instruction is a reference to the object corresponding to the object to which the manipulated instance field belongs, and the integer “1” corresponding to the index of the manipulated field in the example.java application) and invokes the “alert” method, passing the two topmost items popped off the stack to the new method frame as its first two arguments.
  • This change is significant because it modifies the setValues( ) method to execute the “alert” method and associated operations, corresponding to the preceding memory manipulation operation (that is, the “puffield #2” instruction) of the setValues( ) method.
  • the method void alert(java.lang.Object, int), part of the FieldAlert code of Annexure A8 and part of the distributed runtime system (DRT) 71 , requests or otherwise notifies a DRT thread 121 / 1 executing the FieldSend.java code of Annexure A9 to update and propagate the changed identity and value of the manipulated memory location to the plurality of machines M 1 . . . Mn.
  • the modified code permits, in a distributed computing environment having a plurality of computers or computing machines, the coordinated operation of memory manipulation operations so that the problems associated with the operation of the unmodified code or procedure on a plurality of machines M 1 . . . Mn (such as for example inconsistent and incoherent memory state and manipulation and updating operation) does not occur when applying the modified code or procedure.
  • FIGS. 14-16 two laptop computers 101 and 102 are illustrated.
  • the computers 101 and 102 are not necessarily identical and indeed, one can be an IBM or IBM-clone and the other can be an APPLE computer.
  • the computers 101 and 102 have two screens 105 , 115 two keyboards 106 , 116 but a single mouse 107 .
  • the two machines 101 , 102 are interconnected by a means of a single coaxial cable or twisted pair cable 314 .
  • Two simple application programs are downloaded onto each of the machines 101 , 102 , the programs being modified as they are being loaded as described above.
  • the first application is a simple calculator program and results in the image of a calculator 108 being displayed on the screen 105 .
  • the second program is a graphics program which displays four coloured blocks 109 which are of different colours and which move about at random within a rectangular box 310 . Again, after loading, the box 310 is displayed on the screen 105 .
  • Each application operates independently so that the blocks 109 are in random motion on the screen 105 whilst numerals within the calculator 108 can be selected (with the mouse 107 ) together with a mathematical operator (such as addition or multiplication) so that the calculator 108 displays the result.
  • the mouse 107 can be used to “grab” the box 310 and move same to the right across the screen 105 and onto the screen 115 so as to arrive at the situation illustrated in FIG. 15 .
  • the calculator application is being conducted on machine 101 whilst the graphics application resulting in display of box 310 is being conducted on machine 102 .
  • JAVA includes both the JAVA language and also JAVA platform and architecture.
  • the unmodified application code may either be replaced with the modified application code in whole, corresponding to the modifications being performed, or alternatively, the unmodified application code may be replaced in part or incrementally as the modifications are performed incrementally on the executing unmodified application code. Regardless of which such modification routes are used, the modifications subsequent to being performed execute in place of the unmodified application code.
  • An advantage of using a global identifier in the invention described is as a form of ‘meta-name’ or ‘meta-identity’ for all the similar equivalent local objects (or classes, or assets or resources or the like) on each one of the plurality of machines M 1 , . . . , Mn.
  • each machine may instead define or use a global name corresponding to the plurality of similar equivalent objects on each machine (e.g., “globalname7787”), and with the understanding that each machine relates the global name to a specific local name or object (e.g., “globalname7787” corresponds to object “localobject456” on machine M 1 , and “globalname7787” corresponds to object “localobject885” on machine M 2 , and “globalname7787” corresponds to object “localobject111” on machine M 3 , and so forth).
  • a global name corresponding to the plurality of similar equivalent objects on each machine e.g., “globalname7787”
  • each machine relates the global name to a specific local name or object (e.g., “globalname7787” corresponds to object “localobject456” on machine M 1 , and “globalname7787” corresponds to object “localobject885” on machine M 2 , and “globalname7787” corresponds to object “local
  • memory locations include, for example, both fields and array types.
  • the above description deals with fields and the changes required for array types are essentially the same mutatis mutandis.
  • the present invention is equally applicable to similar programming languages (including procedural, declarative and object orientated) to JAVA including Micrsoft.NET platform and architecture (Visual Basic, Visual C/C ++ , and C#) FORTRAN, C/C ++ , COBOL, BASIC etc.
  • the abovementioned arrangement, in which the JAVA code which updates memory locations or field values is modified, is based on the assumption that either the runtime system (say, JAVA HOTSPOT VIRTUAL MACHINE written in C and Java) or the operating system (LINUX written in C and Assembler, for example) of each machine M 1 . . . Mn will ordinarily update memory on the local machine (say M 2 ) but not on any corresponding other machines (M 1 , M 3 . . . Mn). It is possible to leave the JAVA code which updates memory locations or field values unamended and instead amend the LINUX or HOTSPOT routine which updates memory locally, so that it correspondingly updates memory on all other machines as well.
  • the term “updating propagation routine” used herein in conjunction with maintaining the memory of all machines M 1 . . . Mn essentially the same, is to be understood to include within its scope both the JAVA putfield and putstatic instructions and related operations and the “combination” of the JAVA putfield and putstatic operations and the LINUX or HOTSPOT code fragments which perform memory updating.
  • object and class used herein are derived from the JAVA environment and are intended to embrace similar terms derived from different environments such as dynamically linked libraries (DLL), or object code packages, or function unit or memory locations.
  • DLL dynamically linked libraries
  • any one or each of these various means may be implemented by computer program code statements or instructions (possibly including by a plurality of computer program code statements or instructions) that execute within computer logic circuits, processors, ASICs, microprocessors, microcontrollers, or other logic to modify the operation of such logic or circuits to accomplish the recited operation or function.
  • any one or each of these various means may be implemented in firmware and in other embodiments such may be implemented in hardware.
  • any one or each of these various means may be implemented by an combination of computer program software, firmware, and/or hardware.
  • any and each of the aforedescribed methods, procedures, and/or routines may advantageously be implemented as a computer program and/or computer program product stored on any tangible media or existing in electronic, signal, or digital form.
  • Such computer program or computer program products comprising instructions separately and/or organized as modules, programs, subroutines, or in any other way for execution in processing logic such as in a processor or microprocessor of a computer, computing machine, or information appliance; the computer program or computer program products modifying the operation of the computer on which it executes or on a computer coupled with, connected to, or otherwise in signal communications with the computer on which the computer program or computer program product is present or executing.
  • Such computer program or computer program product modifying the operation and architectural structure of the computer, computing machine, and/or information appliance to alter the technical operation of the computer and realize the technical effects described herein.
  • the invention may therefore includes a computer program product comprising a set of program instructions stored in a storage medium or exiting electronically in any form and operable to permit a plurality of computers to carry out any of the methods, procedures, routines, or the like as described herein including in any of the claims.
  • the invention may include a plurality of computers interconnected via a communication network or other communications ink or path and each operable to substantially simultaneously or concurrently execute the same or a different portion of an application program code written to operate on only a single computer on a corresponding different one of computers, wherein the computers being programmed to carry out any of the methods, procedures, or routines described in the specification or set forth in any of the claims, or being loaded with a computer program product.
  • This first excerpt is part of an illustration of the modification code of the modifier 51 in accordance with steps 92 and 103 of FIG. 10 . It searches through the code array of the application program code 50 , and when it detects a memory manipulation instruction (i.e. a putstatic instruction (opcode 178 ) in the JAVA language and virtual machine environment) it modifies the application program code by the insertion of an “alert” routine.
  • a memory manipulation instruction i.e. a putstatic instruction (opcode 178 ) in the JAVA language and virtual machine environment
  • This second excerpt is part of the DRT.alert( ) method and implements the step of 125 and arrow of 127 of FIG. 12 .
  • This DRT.alert( ) method requests one or more threads of the DRT processing environment of FIG. 12 to update and propagate the value and identity of the changed memory location corresponding to the operation of Annexure A1.
  • This third excerpt is part of the DRT 71 , and corresponds to step 128 of FIG. 12 .
  • This code fragment shows the DRT in a separate thread, such as thread 121 / 1 of FIG. 12 , after being notified or requested by step 125 and array 127 , and sending the changed value and changed value location/identity across the network 53 to the other of the plulality of machines M 1 . . . Mn.
  • START MulticastSocket ms DRT.getMulticastSocket( ); // The multicast socket // used by the DRT for // communication.
  • byte nameTag 33; // This is the “name tag” on the network for this // field.
  • Field field modifiedClass.getDeclaredField(“myField1”); // Stores // the field // from the // modified // class. // In this example, the field is a byte field. while (DRT.isRunning( )) ⁇ synchronized (ALERT_LOCK) ⁇ ALERT_LOCK.wait( ); // The DRT thread is waiting for the alert // method to be called.
  • byte[ ] b new byte[ ] ⁇ nameTag, field.getByte(null) ⁇ ; // Stores // the // nameTag // and the // value // of the // field from // the // modified // class in a buffer.
  • DatagramPacket dp new DatagramPacket(b, 0, b.length); ms.send(dp); // Send the buffer out across the network. ⁇ ⁇ // END
  • the fourth excerpt is part of the DRT 71 , and corresponds to steps 135 and 136 of FIG. 13 .
  • This is a fragment of code to receive a propagated identity and value pair sent by another DRT 71 over the network, and write the changed value to the idenitified memory location.
  • START MulticastSocket ms DRT.getMulticastSocket( ); // The multicast socket // used by the DRT for // communication.
  • the fifth excerpt is an dissassembled compiled form of the example.java application of Annexure A7, which performs a memory manipulation operation (putstatic and putfield).
  • the sixth excerpt is the dissassembled compiled form of the same example application in Annexure A5 after modification has been performed by FieldLoader.java of Annexure A11, in accordance with FIG. 9 of this invention. The modifications are highlighted in bold.
  • the seventh excerpt is the source-code of the example.java application used in excerpt A5 and A6.
  • This example application has two memory locations (staticValue and instanceValue) and performs two memory manipulation operations.
  • */ public static int staticValue 0; /** Shared instance field.
  • */ public int instanceValue 0; /** Example method that writes to memory (instance field).
  • the ninth excerpt is the source-code of FieldSend.java which corresponds to step 128 of FIG. 12 , and waits for a request/notification generated by FieldAlert.java of A8 corresponding to step 125 and arrow 127 , and which propagates a changed value/identity pair requested of it by FieldAlert.java, via network 53 .
  • the tenth excerpt is the source-code of FieldReceive.java, which corresponds to steps 135 and 136 of FIG. 13 , and which receives a propagated changed value and identity pair sent to it over the network 53 via FieldSend.java of annexure A9.
  • int command (int) (((buffer[index++] & 0xff) ⁇ 24)
  • int globalID (int) (((buffer[index++] & 0xff) ⁇ 24)
  • Object reference globalIDToObject.get( new Integer(globalID)); // Next, get the array of fields for this object.
  • Field[ ] fields reference.getClass( ).getDeclaredFields( ); while (index ⁇ length) ⁇ // Decode the field id.
  • int fieldID (int) (((buffer[index++] & 0xff) ⁇ 24)
  • FieldLoader.java modifies an application program code, such as the example.java application code of Annexure A7, as it is being loaded into a JAVA virtual machine in accordance with steps 90 , 91 , 92 , 103 , and 94 of FIG. 10 .
  • FieldLoader.java makes use of the convenience classes of Annexures A12 through to A36 during the modification of a compiled JAVA classfile.
  • CONSTANT_Fieldref_info fi (CONSTANT_Fieldref_info) cf.constant_pool[(int) (((ca.code[z][1] & 0xff) ⁇ 8)
  • CONSTANT_Class_info ci (CONSTANT_Class_info) cf.constant_pool[fi.class_index];
  • className cf.constant_pool[ci.name_index].toString( ); if (!name.equals(className)) ⁇ throw new AssertionError(“This code only supports fields “ “local to this class”); ⁇ // Ok, now search for the fields name and index.
  • Convience class for representing attribute_info structures within ClassFiles import java.lang.*; import java.io.*; /** This abstract class represents all types of attribute_info * that are used in the JVM specifications. * * All new attribute_info subclasses are to always inherit from this * class. */ public abstract class attribute_info ⁇ public int attribute_name_index; public int attribute_length; /** This is used by subclasses to register themselves * to their parent classFile. */ attribute_info(ClassFile cf) ⁇ ⁇ /** Used during input serialization by ClassFile only.
  • ClassFile for representing ClassFile structures. import java.lang.*; import java.io.*; import java.util.*; /**
  • the ClassFile follows verbatim from the JVM specification. */ public final class ClassFile ⁇ public int magic; public int minor_version; public int major_version; public int constant_pool_count; public cp_info[ ] constant_pool; public int access_flags; public int this_class; public int super_class; public int interfaces_count; public int[ ] interfaces; public int fields_count; public field_info[ ] fields; public int methods_count; public method_info[ ] methods; public int attributes_count; public attribute_info[ ] attributes; /** Constructor.
  • Convience class for representing Code_attribute structures within ClassFiles import java.util.*; import java.lang.*; import java.io.*; /** * The code[ ] is stored as a 2D array. */ public final class Code_attribute extends attribute_info ⁇ public int max_stack; public int max_locals; public int code_length; public byte[ ][ ] code; public int exception_table_length; public exception_table[ ] exception_table; public int attributes_count; public attribute_info[ ] attributes; /** Internal class that handles the exception table.
  • Convience class for representing CONSTANT_Class_info structures within ClassFiles import java.lang.*; import java.io.*; /** Class subtype of a constant pool entry.
  • CONSTANT_Class_info extends cp_info ⁇ /** The index to the name of this class.
  • */ public int name_index 0; /** Convenience constructor.
  • Convience class for representing CONSTANT_Fieldref_info structures within ClassFiles import java.lang.*; import java.io.*; /** Fieldref subtype of a constant pool entry. */ public final class CONSTANT_Fieldref_info extends cp_info ⁇ /** The index to the class that this field is referencing to. */ public int class_index; /** The name and type index this field if referencing to. */ public int name_and_type_index; /** Convenience constructor.
  • Convience class for representing CONSTANT_InterfaceMethodref_info structures within ClassFiles import java.lang.*; import java.io.*; /** InterfaceMethodref subtype of a constant pool entry. */ public final class CONSTANT_InterfaceMethodref_info extends cp_info ⁇ /** The index to the class that this field is referencing to. */ public int class_index; /** The name and type index this field if referencing to.
  • Convience class for representing CONSTANT_Methodref_info structures within ClassFiles import java.lang.*; import java.io.*; /** Methodref subtype of a constant pool entry. */ public final class CONSTANT_Methodref_info extends cp_info ⁇ /** The index to the class that this field is referencing to. */ public int class_index; /** The name and type index this field if referencing to.
  • Convience class for representing CONSTANT_NameAndType_info structures within ClassFiles. import java.io.*; import java.lang.*; /** NameAndType subtype of a constant pool entry. */ public final class CONSTANT_NameAndType_info extends cp_info ⁇ /** The index to the Utf8 that contains the name. */ public int name_index; /** The index fo the Utf8 that constains the signature.
  • Convience class for representing CONSTANT_Utf8_info structures within ClassFiles import java.io.*; import java.lang.*; /** Utf8 subtype of a constant pool entry. * We internally represent the Utf8 info byte array * as a String. */ public final class CONSTANT_Utf8_info extends cp_info ⁇ /** Length of the byte array. */ public int length; /** The actual bytes, represented by a String. */ public String bytes; /** This constructor should be used for the purpose * of part creation. It does not set the parent * ClassFile reference.
  • Convience class for representing cp_info structures within ClassFiles import java.lang.*; import java.io.*; /** Represents the common interface of all constant pool parts * that all specific constant pool items must inherit from.
  • Convience class for representing field_info structures within ClassFiles import java.lang.*; import java.io.*; /** Represents the field_info structure as specified in the JVM specification. */ public final class field_info ⁇ public int access_flags; public int name_index; public int descriptor_index; public int attributes_count; public attribute_info[ ] attributes; /** Convenience constructor.
  • Convience class for representing LocalVariableTable_attribute structures within ClassFiles import java.lang.*; import java.io.*; /** Used by debugger to find out how the source file line number is linked * to the binary code. It has many to one correspondence and is found in * the Code_attribute.
  • Convience class for representing method_info structures within ClassFiles import java.lang.*; import java.io.*; /** This follows the method_info in the JVM specification. */ public final class method_info ⁇ public int access_flags; public int name_index; public int descriptor_index; public int attributes_count; public attribute_info[ ] attributes; /** Constructor. Creates a method_info, initializes it with * the flags set, and the name and descriptor indexes given.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multi Processors (AREA)
  • Computer And Data Communications (AREA)

Abstract

The present invention discloses a modified computer architecture (50, 71, 72) which enables an applications program (50) to be run simultaneously on a plurality of computers (M1, . . . Mn). Shared memory at each computer is updated with amendments and/or overwrites so that all memory read requests are satisfied locally. During initial program loading (75), or similar, instructions which result in memory being re-written or manipulated are identified (92). Additional instructions are inserted (103) to cause the equivalent memory locations at all computers to be updated.

Description

    RELATED APPLICATIONS
  • This application claims the benefit of priority under ore or more of 35 U.S.C. 119 and/or 35 U.S.C. 120 to the following Australian Patent Applications, U.S. Utility Patent Applications and PCT International Patent Applications, each of which is also a related application and each is incorporated herein by reference in its entirety:
  • U.S. patent application Ser. No. ______ filed 25 Oct. 2005 entitled “Computer Architecture And Method Of Operation For Multi-Computer Distributed Processing With Replicated Memory”;
  • U.S. patent application Ser. No. ______ filed 25 Oct. 2005 entitled “Computer Architecture And Method Of Operation For Multi-Computer Distributed Processing With Initilization Of Objects”;
  • U.S. patent application Ser. No. ______ filed 25 Oct. 2005 entitled “Computer Architecture And Method Of Operation For Multi-Computer Distributed Processing With Finalization Of Objects”;
  • U.S. patent application Ser. No. ______ filed 25 Oct. 2005 entitled “Computer Architecture And Method Of Operation For Multi-Computer Distributed Processing With Synchronization”;
  • U.S. patent application Ser. No. ______ filed 25 Oct. 2005 entitled “Computer Architecture And Method Of Operation For Multi-Computer Distributed Processing And Coordinated Memory And Asset Handling”;
  • Australian Provisional Patent Application No. 2005 902 023 filed 21 Apr. 2005 entitled “Multiple Computer Architecture with Replicated Memory Fields”;
  • Australian Provisional Patent Application No. 2005 902 024 filed 21 Apr. 2005 entitled “Modified Computer Architecture with Initialization of Objects”;
  • Australian Provisional Patent Application No. 2005 902 025 filed 21 Apr. 2005 entitled “Modified Computer Architecture with Finalization of Objects”;
  • Australian Provisional Patent Application No. 2005 902 026 filed 21 Apr. 2005 entitled “Modified Computer Architecture with Synchronization of Objects”;
  • Australian Provisional Patent Application No. 2004 902 027 filed 21 Apr. 2005 entitled “Modified Computer Architecture with Coordinated Objects”;
  • U.S. patent application Ser. No. 11/111,757 filed 22 Apr. 2005 entitled “Multiple Computer Architecture with Replicated Memory Fields”;
  • U.S. patent application Ser. No. 11/111,781 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Initialization of Objects”;
  • U.S. patent application Ser. No. 11/111,778 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Finalization of Objects”;
  • U.S. patent application Ser. No. 11/111,779 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Synchronization of Objects”;
  • U.S. patent application Ser. No. 11/111,946 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Coordinated Objects”;
  • PCT International Application No. PCT/AU05/000/582 filed 22 Apr. 2005 entitled “Multiple Computer Architecture with Replicated Memory Fields”;
  • PCT International Application No. PCT/AU05/000/578 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Initialization of Objects”;
  • PCT International Application No. PCT/AU05/000/581 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Finalization of Objects”;
  • PCT International Application No. PCT/AU05/000/579 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Synchronization of Objects”; and
  • PCT International Application No. PCT/AU05/000/580 filed 22 Apr. 2005 entitled “Modified Computer Architecture with Coordinated Objects”.
  • A further related patent application that is hereby incorporated by reference is U.S. patent application Ser. No. 10/830,042 filed 23 Apr. 2004 entitled “Modified Computer Architecture”.
  • FIELD OF THE INVENTION
  • The present invention relates to computers and other computing machines and information appliances, in particular, to a modified computer architecture and program structure which enables the operation of an application program concurrently or simultaneously on a plurality of computers interconnected via a communications link using a distributed runtime and enables improved performance to be achieved.
  • BACKGROUND OF THE INVENTION
  • Ever since the advent of computers, and computing, software for computers has been written to be operated upon a single machine. As indicated in FIG. 1, that single prior art machine 1 is made up from a central processing unit, or CPU, 2 which is connected to a memory 3 via a bus 4. Also connected to the bus 4 are various other functional units of the single machine 1 such as a screen 5, keyboard 6 and mouse 7.
  • A fundamental limit to the performance of the machine 1 is that the data to be manipulated by the CPU 2, and the results of those manipulations, must be moved by the bus 4. The bus 4 suffers from a number of problems including so called bus “queues” formed by units wishing to gain an access to the bus, contention problems, and the like. These problems can, to some extent, be alleviated by various stratagems including cache memory, however, such stratagems invariably increase the administrative overhead of the machine 1.
  • Naturally, over the years various attempts have been made to increase machine performance. One approach is to use symmetric multi-processors. This prior art approach has been used in so called “super” computers and is schematically indicated in FIG. 2. Here a plurality of CPU's 12 are connected to global memory 13. Again, a bottleneck arises in the communications between the CPU's 12 and the memory 13. This process has been termed “Single System Image”. There is only one application and one whole copy of the memory for the application which is distributed over the global memory. The single application can read from and write to, (i.e. share) any memory location completely transparently.
  • Where there are a number of such machines interconnected via a network, this is achieved by taking the single application written for a single machine and partitioning the required memory resources into parts. These parts are then distributed across a number of computers to form the global memory 13 accessible by all CPU's 12. This procedure relies on masking, or hiding, the memory partition from the single running application program. The performance degrades when one CPU on one machine must access (via a network) a memory location physically located in a different machine.
  • Although super computers have been technically successful in achieving high computational rates, they are not commercially successful in that their inherent complexity makes them extremely expensive not only to manufacture but to administer. In particular, the single system image concept has never been able to scale over “commodity” (or mass produced) computers and networks. In particular, the Single System Image concept has only found practical application on very fast (and hence very expensive) computers interconnected by very fast (and similarly expensive) networks.
  • A further possibility of increased computer power through the use of a plural number of machines arises from the prior art concept of distributed computing which is schematically illustrated in FIG. 3. In this known arrangement, a single application program (Ap) is partitioned by its author (or another programmer who has become familiar with the application program) into various discrete tasks so as to run upon, say, three machines in which case n in FIG. 3 is the integer 3. The intention here is that each of the machines M1 . . . M3 runs a different third of the entire application and the intention is that the loads applied to the various machines be approximately equal. The machines communicate via a network 14 which can be provided in various forms such as a communications link, the internet, intranets, local area networks, and the like. Typically the speed of operation of such networks 14 is an order of magnitude slower than the speed of operation of the bus 4 in each of the individual machines M1, M2, . . . , Mn.
  • Distributed computing suffers from a number of disadvantages. Firstly, it is a difficult job to partition the application and this must be done manually. Secondly, communicating data, partial results, results and the like over the network 14 is an administrative overhead. Thirdly, the need for partitioning makes it extremely difficult to scale upwardly by utilising more machines since the application having been partitioned into, say three, does not run well upon four machines. Fourthly, in the event that one of the machines should become disabled, the overall performance of the entire system is substantially degraded.
  • A further prior art arrangement is known as network computing via “clusters” as is schematically illustrated in FIG. 4. In this approach, the entire application is loaded onto each of the machines M1, M2, . . . , Mn. Each machine communicates with a common database but does not communicate directly with the other machines. Although each machine runs the same application, each machine is doing a different “job” and uses only its own memory. This is somewhat analogous to a number of windows each of which sell train tickets to the public. This approach does operate, is scalable and mainly suffers from the disadvantage that it is difficult to administer the network.
  • In computer languages such as for example JAVA and MICROSOFT.NET there are two major types of constructs with which programmers deal. In the JAVA language these are known as objects and classes. More generally they may be referred to as assets. Every time an object (or other asset) is created there is an initialization routine run known as an object initialization (e.g., “<nit>”) routine. Similarly, every time a class is loaded there is a class initialization routine known as “<clinit>”. Other languages use different terms but utilize a similar concept. In either case, however, there is no equivalent “clean up” or deletion routine to delete an object or class (or other asset) once it is no longer required. Instead, this “clean up” happens unobtrusively in a background mode.
  • Furthermore, in any computer environment it is necessary to acquire and release a lock to enable the use of such objects, classes, assets, resources or structures to avoid different parts of the application program from attempting to use the same objects, classes, assets, resources or structures at the one time. In the JAVA environment this is known as synchronization. Synchronization more generally refers to the exclusive use of an object, class, resource, structure, or other asset to avoid contention between and among computers or machines. This is achieved in JAVA by the “monitor enter” and “monitor exit” instructions or routines. Other languages use different terms but utilize a similar concept.
  • Unfortunately, conventional computing systems, architectures, and operating schemes do not provide for computing environments and methods in which an application program can operate simultaneously on an arbitrary plurality of computers where the environment and operating scheme ensure that the abovementioned memory management, initialization, clean up and synchronization procedures operate in a consistent and coordinated fashion across all the computing machines.
  • SUMMARY
  • The present invention discloses a computing environment in which an application program operates simultaneously on a plurality of computers. In such an environment it is advantageous to ensure that the abovementioned asset initialization, clean-up and synchronization procedures operate in a consistent and coordinated fashion across all the machines.
  • In accordance with a first aspect of the present invention there is disclosed a multiple computer system having at least one application program each written to operate on only a single computer but running simultaneously on a plurality of computers interconnected by a communications network, wherein different portions of said application program(s) execute substantially simultaneously on different ones of said computers, wherein each computer has an independent local memory accessible only by the corresponding portion of said application program(s) and wherein for each said portion a like plurality of substantially identical objects are created, each in the corresponding computer.
  • In accordance with a second aspect of the present invention there is disclosed A plurality of computers interconnected via a communications link and each having an independent local memory and substantially simultaneously operating a different portion at least one application program each written to operate on only a single computer, each local memory being accessible only by the corresponding portion of said application program.
  • In accordance with a third aspect of the present invention there is disclosed a method of running simultaneously on a plurality of computers at least one application program each written to operate on only a single computer, said computers being interconnected by means of a communications network and each having an independent local memory, said method comprising the step of: (i) executing different portions of said application program(s) on different ones of said computers and for each said portion creating a like plurality of substantially identical objects each in the corresponding computer and each accessible only by the corresponding portion of said application program.
  • In accordance with a fourth aspect of the present invention there is disclosed a method of loading an application program written to operate only on a single computer onto each of a plurality of computers, the computers being interconnected via a communications link, and different portions of said application program(s) being substantially simultaneously executable on different computers with each computer having an independent local memory accessible only by the corresponding portion of said application program(s), the method comprising the step of modifying the application before, during, or after loading and before execution of the relevant portion of the application program.
  • In accordance with a fifth aspect of the present invention there is disclosed a method of operating simultaneously on a plurality of computers all interconnected via a communications link at least one application program each written to operate on only a single computer, each of said computers having at least a minimum predetermined local memory capacity, different portions of said application program(s) being substantially simultaneously executed on different ones of said computers with the local memory of each computer being only accessible by the corresponding portion of said application program(s), said method comprising the steps of: (i) initially providing each local memory in substantially identical condition, (ii) satisfying all memory reads and writes generated by each said application program portion from said corresponding local memory, and (iii) communicating via said communications link all said memory writes at each said computer which take place locally to all the remainder of said plurality of computers whereby the contents of the local memory utilised by each said computer, subject to an updating data transmission delay, remains substantially identical.
  • In accordance with a sixth aspect of the present invention there is disclosed A method of compiling or modifying an application program written to operate on only a single computer but to run simultaneously on a plurality of computers interconnected via a communications link, with different portions of said application program(s) executing substantially simultaneously on different ones of said computers each of which has an independent local memory accessible only by the corresponding portion of said application program, said method comprising the steps of: (i) detecting instructions which share memory records utilizing one of said computers, (ii) listing all such shared memory records and providing a naming tag for each listed memory record, (iii) detecting those instructions which write to, or manipulate the contents of, any of said listed memory records, and (iv) activating an updating propagation routine following each said detected write or manipulate instruction, said updating propagation routine forwarding the re-written or manipulated contents and name tag of each said re-written or manipulated listed memory record to the remainder of said computers.
  • In accordance with a seventh aspect of the present invention there is disclosed in a multiple thread processing computer operation in which individual threads of a single application program written to operate on only a single computer are simultaneously being processed each on a different corresponding one of a plurality of computers each having an independent local memory accessible only by the corresponding thread and each being interconnected via a communications link, the improvement comprising communicating changes in the contents of local memory physically associated with the computer processing each thread to the local memory of each other said computer via said communications link.
  • In accordance with a eighth aspect of the present invention there is disclosed a computer program product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the above-mentioned methods.
  • In accordance with a ninth aspect of the invention there is disclosed a distributed run time and distributed run time system adapted to enable communications between a plurality of computers, computing machines, or information appliances.
  • In accordance with a tenth aspect of the invention there is disclosed a modifier, modifier means, and modifier routine for modifying an application program written to execute on a single computer or computing machine at a time to execute simultaneously on a plurality of networked computers or computing machines. distributed run time and distributed run time system adapted to enable communications between a plurality of computers, computing machines, or information appliances.
  • In accordance with an eleventh aspect of the present invention there is disclosed a computer program and computer program product written to operate on only a single computer but product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the abovementioned procedures, routines, and methods.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Embodiments of the present invention are now described with reference to the drawings in which:
  • FIG. 1 is a schematic view of the internal architecture of a conventional computer;
  • FIG. 2 is a schematic illustration showing the internal architecture of known symmetric multiple processors;
  • FIG. 3 is a schematic representation of prior art distributed computing;
  • FIG. 4 is a schematic representation of a prior art network computing using clusters;
  • FIG. 5 is a schematic block diagram of a plurality of machines operating the same application program in accordance with a first embodiment of the present invention;
  • FIG. 6 is a schematic illustration of a prior art computer arranged to operate JAVA code and thereby constitute a JAVA virtual machine;
  • FIG. 7 is a drawing similar to FIG. 6 but illustrating the initial loading of code in accordance with the preferred embodiment;
  • FIG. 8 is a drawing similar to FIG. 5 but illustrating the interconnection of a plurality of computers each operating JAVA code in the manner illustrated in FIG. 7;
  • FIG. 9 is a flow chart of the procedure followed during loading of the same application on each machine in the network;
  • FIG. 10 is a flow chart showing a modified procedure similar to that of FIG. 9;
  • FIG. 11 is a schematic representation of multiple thread processing carried out on the machines of FIG. 8 utilizing a first embodiment of memory updating;
  • FIG. 12 is a schematic representation similar to FIG. 11 but illustrating an alternative embodiment;
  • FIG. 13 illustrates multi-thread memory updating for the computers of FIG. 8;
  • FIG. 14 is a schematic representation of two laptop computers interconnected to simultaneously run a plurality of applications, with both applications running on a single computer;
  • FIG. 15 is a view similar to FIG. 14 but showing the FIG. 14 apparatus with one application operating on each computer;
  • FIG. 16 is a view similar to FIGS. 14 and 15 but showing the FIG. 14 apparatus with both applications operating simultaneously on both computers; and
  • FIG. 17 is a drawing showing an alternative to similar to FIG. 7.
  • The specification includes an Annexure A which provides actual program fragments which implement various aspects of the described embodiments.
  • REFERENCE TO ANNEXES
  • Although the specification provides a complete and detailed description of the several embodiments of the invention such that the invention may be understood and implemented without reference to other materials, the specification does includes Annexure A which provide exemplary actual program or code fragments which implement various aspects of the described embodiments. Although aspects of the invention are described throughout the specification including the Annexes, drawings, and claims, it may be appreciated that Annexure A relates primarily to fields. More particularly, the accompanying Annexure is provided in which: Annexures A1-A10 illustrate exemplary code to illustrate embodiments of the invention in relation to fields.
  • It will be appreciated in light of the description provided here that the categorization of the Annexures as well as the use of other headings and subheadings in this description is intended as an aid to the reader and is not to be used to limit the scope of the invention in any way.
  • DETAILED DESCRIPTION OF EMBODIMENTS OF THE INVENTION
  • The present invention discloses a modified computer architecture which enables an applications program to be run simultaneously on a plurality of computers in a manner that overcomes the limitations of the aforedescribed conventional architectures, systems, methods, and computer programs.
  • In one aspect, shared memory at each computer may be updated with amendments and/or overwrites so that all memory read requests are satisfied locally. Before, during or after program loading, but before execution of relevant portions of the program code are executed, or similar, instructions which result in memory being re-written or manipulated are identified. Additional instructions are inserted into the program code (or other modification made) to cause the equivalent memory locations at all computers to be updated. While the invention is not limited to JAVA language or virtual machines, exemplary embodiments are described relative to the JAVA language and standards.
  • In another aspect, the initialization of JAVA language classes and objects (or other assets) are provided for so all memory locations for all computers are initialized in the same manner. In another aspect, the finalization of JAVA language classes and objects is also provide so finalization only occurs when the last class or object present on all machines is no longer required. In still another aspect, synchronization is provided such that instructions which result in the application program acquiring (or releasing) a lock on a particular asset (synchronization) are identified. Additional instructions are inserted (or other code modifications performed) to result in a modified synchronization routine with which all computers are updated.
  • As will become more apparent in light of the further description provided herein, one of the features of the invention is to make it appear that one common application program or application code and its executable version (with likely modification) is simultaneously or concurrently executing across a plurality of computers or machines M1, . . . , Mn. As will be described in considerable detail hereinafter, the instant invention achieves this by running the same application program (for example, Microsoft Word or Adobe Photoshop CS2) on each machine, but modifying the executable code of that application program on each machine as necessary such that each executing instance (‘copy’) on each machine coordinates its local operations on any particular machine with the operations of the respective instances on the other machines such that they all function together in a consistent, coherent and coordinated manner and give the appearance of being one global instance of the application (i.e., a “meta-application”).
  • In accordance with embodiments of the present invention a single application code 50 (sometimes more informally referred to as the application or the application program) can be operated simultaneously on a number of machines M1, M2 . . . Mn interconnected via a communications network or other communications link or path 53. The communications network or path may be any electronic signaling, data, or digital communications network or path and may advantageously be a relatively slow speed communications path, such as a network connection over the Internet or any common networking configurations known or available as of the date or this applications, and extensions and improvements, thereto.
  • By way of example but not limitation, one application code or program 50 may be a single application on the machines, such as Microsoft Word, as opposed to different applications on each machine, such as Microsoft Word on machine M1, and Microsoft PowerPoint on machine M2, and Netscape Navigator on machine M3 and so forth. Therefore the terminology “one” application code or program and a “common” application code or program is used to try and capture this situation where all machines M1, . . . , Mn are operating or executing the same program or code and not different (and unrelated) programs. In other words copies or replicas of same or substantially the same application code is loaded onto each of the interoperating and connected machines or computers. As the characteristics of each machine or computer may differ, the application code 50 may be modified before loading, during the loading process, and with some restrictions after the loading process to provide a customization or modification of the code on each machine. Some dissimilarity between the programs may be permitted so long as the other requirements for interoperability, consistency, and coherency as described herein can be maintain. As it will become apparent hereafter, each of the machines M1, M2 . . . Mn operates with the same application code 50 on each machine M1, M2 . . . , Mn and thus all of the machines M1, M2, . . . , Mn have the same or substantially the same application code 50 usually with a modification that may be machine specific.
  • Similarly, each of the machines M1, M2, . . . , Mn operates with the same (or substantially the same or similar) modifier 51 (in some embodiments implemented as a distributed run time or DRT 71) on each machine M1, M2, . . . , Mn and thus all of the machines M1, M2 . . . Mn have the same (or substantially the same or similar) modifier 51 for each modification required. Different modification for example may be required for memory management and replication, initialization, finalization, and/or synchronization (though not all of these modification types may be required for all embodiments).
  • In addition, during the loading of, or at any time preceding the execution of, the application code 50 (or relevant portion thereof) on each machine M1, M2 . . . Mn, each application code 50 has been modified by the corresponding modifier 51 according to the same rules (or substantially the same rules since minor optimizing changes are permitted within each modifier 51/1, 51/2, . . . , 51/n).
  • With specific reference to any memory management modifier that may be provided, such memory management modifier 51-M or DRT 71-M or other code modifying means component of the overall modifier or distributed run time means is responsible for creating or replicating a memory structure and contents on each of the individual machines M1, M2 . . . Mn that permits the plurality of machines to interoperate. In some embodiments this replicated memory structure will be identical, in other embodiments this memory structure will have portions that are identical and other portions that are not, and in still other embodiments the memory structures are or may not be identical.
  • These structures and procedures when applied in combination when required, maintain a computing environment where memory locations, address ranges, objects, classes, assets, resources, or any other procedural or structural aspect of a computer or computing environment are where required created, maintained, operated, and deactivated or deleted in a coordinated, coherent, and consistent manner across the plurality of individual machines M1, M2 . . . Mn.
  • Attention is now directed to the particulars of several aspects of the invention that may be utilised alone or in any combination.
  • In connection with FIG. 5, in accordance with a preferred embodiment of the present invention a single application code 50 (sometimes more informally referred to as the application or the application program) can be operated simultaneously on a number of machines M1, M2 . . . Mn interconnected via a communications network or other communications link or path 53. By way of example but not limitation, one application code or program 50 would be a single common application program on the machines, such as Microsoft Word, as opposed to different applications on each machine, such as Microsoft Word on machine M1, and Microsoft PowerPoint on machine M2, and Netscape Navigator on machine M3 and so forth. Therefore the terminology “one”, “single”, and “common” application code or program is used to try and capture this situation where all machines M1, . . . , Mn are operating or executing the same program or code and not different (and unrelated) programs. In other words copies or replicas of same or substantially the same application code is loaded onto each of the interoperating and connected machines or computers. As the characteristics of each machine or computer may differ, the application code 50 may be modified before loading, during the loading process, or after the loading process to provide a customization or modification of the code on each machine. Some dissimilarity between the programs may be permitted so long as the other requirements for interoperability, consistency, and coherency as described herein can be maintain. As it will become apparent hereafter, each of the machines M1, M2 . . . Mn operates with the same application code 50 on each machine M1, M2 . . . Mn and thus all of the machines M1, M2, . . . , Mn have the same or substantially the same application code 50 usually with a modification that may be machine specific.
  • Similarly, each of the machines M1, M2, . . . , Mn operates with the same (or substantially the same or similar) modifier 51 on each machine M1, M2, . . . , Mn and thus all of the machines M1, M2 . . . Mn have the same (or substantially the same or similar) modifier 51 with the modifier of machine M1 being designated 51/1 and the modifier of machine M2 being designated 51/2, etc. In addition, before or during the loading of, or preceding the execution of, or even after execution has commenced, the application code 50 on each machine M1, M2 . . . Mn is modified by the corresponding modifier 51 according to the same rules (or substantially the same rules since minor optimizing changes are permitted within each modifier 51/1, 51/2, . . . , 51/n).
  • As will become more apparent in light of the further description provided herein, one of the features of the invention is to make it appear that one application program instance of application code 50 is executing simultaneously across all of the plurality of machines M1, M2, . . . , Mn. As will be described in considerable detail hereinafter, the instant invention achieves this by running the same application program code (for example, Microsoft Word or Adobe Photoshop CS2) on each machine, but modifying the executable code of that application program on each machine such that each executing occurrence (or ‘local instance’) on each one of the machines M1 . . . Mn coordinates its local operations with the operations of the respective occurrences on each one of the other machines such that each occurrence on each one of the plurality of machines function together in a consistent, coherent and coordinated manner so as to give the appearance of being one global instance (or occurrence) of the application program and program code (i.e., a “meta-application”).
  • As a consequence of the above described arrangement, if each of the machines M1, M2, . . . , Mn has, say, an internal memory capability of 10 MB, then the total memory available to each application code 50 is not necessarily, as one might expect the number of machines (n) times 10 MB, or alternatively the additive combination of the internal memory capability of all n machines, but rather or still may only be 10 MB. In the situation where the internal memory capacities of the machines are different, which is permissible, then in the case where the internal memory in one machine is smaller than the internal memory capability of at least one other of the machines, then the size of the smallest memory of any of the machines may be used as the maximum memory capacity of the machines when such memory (or a portion thereof) is to be treated as a ‘common’ memory (i.e. similar equivalent memory on each of the machines M1 . . . Mn) or otherwise used to execute the common application code.
  • However, even though the manner that the internal memory of each machine is treated may initially appear to be a possible constraint on performance, how this results in improved operation and performance will become apparent hereafter. Naturally, each machine M1, M2 . . . Mn has an private (i.e. ‘non-common’) internal memory capability. The private internal memory capability of the machines M1, M2, . . . , Mn are normally approximately equal but need not be. It may also be advantageous to select the amounts of internal memory in each machine to achieve a desired performance level in each machine and across a constellation or network of connected or coupled plurality of machines, computers, or information appliances M1, M2, . . . , Mn. Having described these internal and common memory considerations, it will be apparent in light of the description provided herein that the amount of memory that can be common between machines is not a limitation of the invention.
  • It is known from the prior art to operate a single computer or machine (produced by one of various manufacturers and having an operating system operating in one of various different languages) in a particular language of the application, by creating a virtual machine as schematically illustrated in FIG. 6. The code and data and virtual machine configuration or arrangement of FIG. 6 takes the form of the application code 50 written in the Java language and executing within a Java Virtual Machine 61. Thus, where the intended language of the application is the language JAVA, a JAVA virtual machine is used which is able to operate code in JAVA irrespective of the machine manufacturer and internal details of the machine. For further details see “The JAVA Virtual Machine Specification” 2nd Edition by T. Lindholm & F. Yellin of Sun Microsystems Inc. of the USA, which is incorporated by reference herein.
  • This conventional art arrangement of FIG. 6 is modified in accordance with embodiments of the present invention by the provision of an additional facility which is conveniently termed “distributed run time” or “distributed run time system” DRT 71 and as seen in FIG. 7. An alternative showing the modifier 171 and loading 175 code 50 is in FIG. 17.
  • In FIG. 7, the application code 50 is loaded onto the Java Virtual Machine 72 in cooperation with the distributed runtime system 71, through the loading procedure indicated by arrow 75. As used herein the terms distributed runtime and the distributed run time system are essentially synonymous, and by means of illustration but not limitation are generally understood to include library code and processes which support software written in a particular language running on a particular platform. Additionally, a distributed runtime system may also include library code and processes which support software written in a particular language running within a particular distributed computing environment. The runtime system typically deals with the details of the interface between the program and the operation system such as system calls, program start-up and termination, and memory management. For purposes of background, a conventional Distributed Computing Environment (DCE) that does not provide the capabilities of the inventive distributed run time or distributed run time system 71 required in the invention is available from the Open Software Foundation. This Distributed Computing Environment (DCE) performs a form of computer-to-computer communication for software running on the machines, but among its many limitations, it is not able to implement the modification or communication operations of this invention. Among its functions and operations, the inventive DRT 71 coordinates the particular communications between the plurality of machines M1, M2, . . . , Mn. Moreover, the inventive distributed runtime 71 comes into operation during the loading procedure indicated by arrow 75 of the JAVA application 50 on each JAVA virtual machine 72 of machines JVM# 1, JVM# 2, . . . JVM#n. The sequence of operations during loading will be described hereafter in relation to FIG. 9. It will be appreciated in light of the description provided herein that although many examples and descriptions are provided relative to the JAVA language and JAVA virtual machines so that the reader may get the benefit of specific examples, the invention is not restricted to either the JAVA language or JAVA virtual machines, or to any other language, virtual machine, machine, or operating environment.
  • FIG. 8 shows in modified form the arrangement of FIG. 5 utilising JAVA virtual machines, each as illustrated in FIG. 7. It will be apparent that again the same application code 50 is loaded onto each machine M1, M2 . . . Mn. However, the communications between each machine M1, M2, . . . , Mn, and indicated by arrows 83, although physically routed through the machine hardware, are advantageously controlled by the individual DRT's 71/1 . . . 71/n within each machine. Thus, in practice this may be conceptionalised as the DRT's 71/1, . . . , 71/n communicating with each other via the network or other communications link 73 rather than the machines M1, M2, . . . , Mn communicating directly with themselves or each other. Actually, the invention contemplates and included either this direct communication between machines M1, M2, . . . , Mn or DRTs 71/1, 71/2, . . . , 71/n or a combination of such communications. The inventive DRT 71 provides communication that is transport, protocol, and link independent.
  • It will be appreciated in light of the description provided herein that there are alternative implementations of the modifier 51 and the distributed run time 71. For example, the modifier 51 may be implemented as a component of or within the distributed run time 71, and therefore the DRT 71 may implement the functions and operations of the modifier 51. Alternatively, the function and operation of the modifier 51 may be implemented outside of the structure, software, firmware, or other means used to implement the DRT 71. In one embodiment, the modifier 51 and DRT 71 are implemented or written in a single piece of computer program code that provides the functions of the DRT and modifier. The modifier function and structure therefore maybe subsumed into the DRT and considered to be an optional component. Independent of how implemented, the modifier function and structure is responsible for modifying the executable code of the application code program, and the distributed run time function and structure is responsible for implementing communications between and among the computers or machines. The communications functionality in one embodiment is implemented via an intermediary protocol layer within the computer program code of the DRT on each machine. The DRT may for example implement a communications stack in the JAVA language and use the Transmission Control Protocol/Internet Protocol (TCP/IP) to provide for communications or talking between the machines. Exactly how these functions or operations are implemented or divided between structural and/or procedural elements, or between computer program code or data structures within the invention are less important than that they are provided.
  • However, in the arrangement illustrated in FIG. 8, (and also in FIGS. 31-32), a plurality of individual computers or machines M1, M2, . . . , Mn are provided, each of which are interconnected via a communications network 53 or other communications link and each of which individual computers or machines provided with a modifier 51 (See in FIG. 5) and realised by or in for example the distributed run time (DRT) 71 (See FIG. 8) and loaded with a common application code 50. The term common application program is to be understood to mean an application program or application program code written to operate on a single machine, and loaded and/or executed in whole or in part on each one of the plurality of computers or machines M1, M2 . . . Mn, or optionally on each one of some subset of the plurality of computers or machines M1, M2 . . . Mn. Put somewhat differently, there is a common application program represented in application code 50, and this single copy or perhaps a plurality of identical copies are modified to generate a modified copy or version of the application program or program code, each copy or instance prepared for execution on the plurality of machines. At the point after they are modified they are common in the sense that they perform similar operations and operate consistently and coherently with each other. It will be appreciated that a plurality of computers, machines, information appliances, or the like implementing the features of the invention may optionally be connected to or coupled with other computers, machines, information appliances, or the like that do not implement the features of the invention.
  • Essentially in at least one embodiment the modifier 51 or DRT 71 or other code modifying means is responsible for modifying the application code 50 so that it may execute memory manipulation operations, such as memory putstatic and puffield instructions in the JAVA language and virtual machine environment, in a coordinated, consistent, and coherent manner across and between the plurality of individual machines M1 . . . Mn. It follows therefore that in such a computing environment it is necessary to ensure that each of memory location is manipulated in a consistent fashion (with respect to the others).
  • In some embodiments, some or all of the plurality of individual computers or machines may be contained within a single housing or chassis (such as so-called “blade servers” manufactured by Hewlett-Packard Development Company, Intel Corporation, IBM Corporation and others) or implemented on a single printed circuit board or even within a single chip or chip set.
  • A machine (produced by any one of various manufacturers and having an operating system operating in any one of various different languages) can operate in the particular language of the application program code 50, in this instance the JAVA language. That is, a JAVA virtual machine 72 is able to operate application code 50 in the JAVA language, and utilize the JAVA architecture irrespective of the machine manufacturer and the internal details of the machine.
  • When implemented in a non-JAVA language or application code environment, the generalized platform, and/or virtual machine and/or machine and/or runtime system is able to operate application code 50 in the language(s) (possibly including for example, but not limited to any one or more of source-code languages, intermediate-code languages, object-code languages, machine-code languages, and any other code languages) of that platform, and/or virtual machine and/or machine and/or runtime system environment, and utilize the platform, and/or virtual machine and/or machine and/or runtime system and/or language architecture irrespective of the machine manufacturer and the internal details of the machine. It will also be appreciated in light of the description provided herein that platform and/or runtime system may include virtual machine and non-virtual machine software and/or firmware architectures, as well as hardware and direct hardware coded applications and implementations.
  • For a more general set of virtual machine or abstract machine environments, and for current and future computers and/or computing machines and/or information appliances or processing systems, and that may not utilize or require utilization of either classes and/or objects, the inventive structure, method, and computer program and computer program product are still applicable. Examples of computers and/or computing machines that do not utilize either classes and/or objects include for example, the x86 computer architecture manufactured by Intel Corporation and others, the SPARC computer architecture manufactured by Sun Microsystems, Inc and others, the PowerPC computer architecture manufactured by International Business Machines Corporation and others, and the personal computer products made by Apple Computer, Inc., and others. For these types of computers, computing machines, information appliances, and the virtual machine or virtual computing environments implemented thereon that do not utilize the idea of classes or objects, may be generalized for example to include primitive data types (such as integer data types, floating point data types, long data types, double data types, string data types, character data types and Boolean data types), structured data types (such as arrays and records) derived types, or other code or data structures of procedural languages or other languages and environments such as functions, pointers, components, modules, structures, references and unions.
  • Turning now to FIGS. 7 and 9, during the loading procedure 75, the application code 50 being loaded onto or into each JAVA virtual machine 72 is modified by DRT 71. This modification commences at Step 90 in FIG. 9 and involves the initial step 91 of preferably scrutinizing or analysing the code and detecting all memory locations addressable by the application code 50, or optionally some subset of all memory locations addressable by the application code 50; such as for example named and unnamed memory locations, variables (such as local variables, global variables, and formal arguments to subroutines or functions), fields, registers, or any other address space or range of addresses which application code 50 may access. Such memory locations in some instances need to be identified for subsequent processing at steps 92 and 93. In some embodiments, where a list of detected memory locations is required for further processing, the DRT 71 during the loading procedure 75 creates a list of all the memory locations thus identified. In one embodiment, the memory locations in the form of JAVA fields are listed by object and class, however, the memory locations, fields, or the like may be listed or organized in any manner so long as they comport with the architectural and programming requirements of the system on which the program is to be used and the principles of the invention described herein. This detection is optional and not required in all embodiments of the invention. It may be noted that the DRT is at least in part fulfilling the roll of the modifier 51.
  • The next phase (designated Step 92 in FIG. 9) [Step 92] of the modification procedure is to search through the application code 50 in order to locate processing activity or activities that manipulate or change values or contents of any listed memory location (for example, but not limited to JAVA fields) corresponding to the list generated at step 91 when required. Preferably, all processing activities that manipulate or change any one or more values or contents of any one or more listed memory locations, are located.
  • When such a processing activity or operation (typically “putstatic” or “puffield” in the JAVA language, or for example, a memory assignment operation, or a memory write operation, or a memory manipulation operation, or more generally operations that otherwise manipulate or change value(s) or content(s) of memory or other addressable areas), is detected which changes the value or content of a listed or detected memory location, then an “updating propagation routine” is inserted by step 93 in the application code 50 corresponding to the detected memory manipulation operation, to communicate with all other machines in order to notify all other machines of the identity of the manipulated memory location, and the updated, manipulated or changed value(s) or content(s) of the manipulated memory location. The inserted “updating propagation routine” preferably takes the form of a method, function, procedure, or similar subroutine call or operation to a network communications library of DRT 71. Alternatively, the “updating propagation routine” may take the optional form of a code-block (or other inline code form) inserted into the application code instruction stream at, after, before, or otherwise corresponding to the detected manipulation instruction or operation. And preferably, in a multi-tasking or parallel processing machine environment (and in some embodiments inclusive or exclusive of operating system), such as a machine environment capable of potentially simultaneous or concurrent execution of multiple or different threads or processes, the “updating propagation routine” may execute on the same thread or process or processor as the detected memory manipulation operation of step 92. Thereafter, the loading procedure continues, by loading the modified application code 50 on the machine 72 in place of the unmodified application code 50, as indicated by step 94 in FIG. 9.
  • An alternative form of modification during loading is illustrated in the illustration of FIG. 10. Here the start and listing steps 90 and 91 and the searching step 92 are the same as in FIG. 9. However, rather than insert the “updating propagation routine” into the application code 50 corresponding to the detected memory manipulation operation identified in step 92, as is indicated in step 93, in which the application code 50, or network communications library code 71 of the DRT executing on the same thread or process or processor as the detected memory manipulation operation, carries out the updating, instead an “alert routine” is inserted corresponding to the detected memory manipulation operation, at step 103. The “alert routine” instructs, notifies or otherwise requests a different and potentially simultaneously or concurrently executing thread or process or processor not used to perform the memory manipulation operation (that is, a different thread or process or processor than the thread or process or processor which manipulated the memory location), such as a different thread or process allocated to the DRT 71, to carry out the notification, propagation, or communication of all other machines of the identity of the manipulated memory location, and the updated, manipulated or changed value(s) or content(s) of the manipulated memory location.
  • Once this modification during the loading procedure has taken place and execution begins of the modified application code 50, then either the steps of FIG. 11 or FIG. 12 take place. FIG. 11 (and the steps 112, 113, 114, and 115 therein) correspond to the execution and operation of the modified application code 50 when modified in accordance with the procedures set forth in and described relative to FIG. 9. FIG. 12 on the other hand (and the steps 112, 113, 125, 127, and 115 therein) set forth therein correspond to the execution and operation of the modified application code 50 when modified in accordance with FIG. 10.
  • This analysis or scrutiny of the application code 50 can may take place either prior to loading the application program code 50, or during the application program code 50 loading procedure, or even after the application program code 50 loading procedure. It may be likened to an instrumentation, program transformation, translation, or compilation procedure in that the application code may be instrumented with additional instructions, and/or otherwise modified by meaning-preserving program manipulations, and/or optionally translated from an input code language to a different code language (such as for example from source-code language or intermediate-code language to object-code language or machine-code language), and with the understanding that the term compilation normally or conventionally involves a change in code or language, for example, from source code to object code or from one language to another language. However, in the present instance the term “compilation” (and its grammatical equivalents) is not so restricted and can also include or embrace modifications within the same code or language. For example, the compilation and its equivalents are understood to encompass both ordinary compilation (such as for example by way of illustration but not limitation, from source-code to object-code), and compilation from source-code to source-code, as well as compilation from object-code to object-code, and any altered combinations therein. It is also inclusive of so-called “intermediary-code languages” which are a form of “pseudo object-code”.
  • By way of illustration and not limitation, in one embodiment, the analysis or scrutiny of the application code 50 may take place during the loading of the application program code such as by the operating system reading the application code from the hard disk or other storage device or source and copying it into memory and preparing to begin execution of the application program code. In another embodiment, in a JAVA virtual machine, the analysis or scrutiny may take place during the class loading procedure of the java.lang.ClassLoader loadClass method (e.g., “java.lang.ClassLoader.loadClass( )”).
  • Alternatively, the analysis or scrutiny of the application code 50 may take place even after the application program code loading procedure, such as after the operating system has loaded the application code into memory, or optionally even after execution of the relevant corresponding portion of the application program code has started, such as for example after the JAVA virtual machine has loaded the application code into the virtual machine via the “java.lang.ClassLoader.loadClass( )” method and optionally commenced execution.
  • As seen in FIG. 11, a multiple thread processing machine environment 110, on each one of the machines M1, . . . , Mn and consisting of threads 111/1 . . . 111/4 exists. The processing and execution of the second thread 111/2 (in this example) results in that thread 111/2 manipulating a memory location at step 113, by writing to a listed memory location. In accordance with the modifications made to the application code 50 in the steps 90-94 of FIG. 9, the application code 50 is modified at a point corresponding to the write to the memory location of step 113, so that it propagates, notifies, or communicates the identity and changed value of the manipulated memory location of step 113 to the other machines M2, . . . , Mn via network 53 or other communication link or path, as indicated at step 114. At this stage the processing of the application code 50 of that thread 111/2 is or may be altered and in some instances interrupted at step 114 by the executing of the inserted “updating propagation routine”, and the same thread 111/2 notifies, or propagates, or communicates to all other machines M2, . . . , Mn via the network 53 or other communications link or path of the identity and changed value of the manipulated memory location of step 113. At the end of that notification, or propagation, or communication procedure 114, the thread 111/2 then resumes or continues the processing or the execution of the modified application code 50 at step 115.
  • In the alternative arrangement illustrated in FIG. 12, a multiple thread processing machine environment 110 comprising or consisting of threads 111/1, . . . , 111/3, and a simultaneously or concurrently executing DRT processing environment 120 consisting of the thread 121/1 as illustrated, or optionally a plurality of threads, is executing on each one of the machines M1, . . . Mn. The processing and execution of the modified application code 50 on thread 111/2 results in a memory manipulation operation of step 113, which in this instance is a write to a listed memory location. In accordance with the modifications made to the application code 50 in the steps 90, 91, 92, 103, and 94 of FIG. 9, the application code 50 is modified at a point corresponding to the write to the memory location of step 113, so that it requests or otherwise notifies the threads of the DRT processing environment 120 to notify, or propagate, or communicate to the other machines M2, . . . , Mn of the identity and changed value of the manipulated memory location of step 113, as indicated at steps 125 and 128 and arrow 127. In accordance with this modification, the thread 111/2 processing and executing the modified application code 50 requests a different and potentially simultaneously or concurrently executing thread or process (such as thread 121/1) of the DRT processing environment 120 to notify the machines M2, . . . , Mn via network 53 or other communications link or path of the identity and changed value of the manipulated memory location of step 113, as indicated in step 125 and arrow 127. In response to this request of step 125 and arrow 127, a different and potentially simultaneously or concurrently executing thread or process 121/1 of the DRT processing environment 120 notifies the machines M2, . . . , Mn via network 53 or other communications link or path of the identity and changed value of the manipulated memory location of step 113, as requested of it by the modified application code 50 executing on thread 111/2 of step 125 and arrow 127.
  • When compared to the earlier described step 114 of thread 111/2 of FIG. 11, step 125 of thread 111/2 of FIG. 12 can be carried out quickly, because step 114 of thread 111/2 must notify and communicate with machines M2, . . . , Mn via the relatively slow network 53 (relatively slow for example when compared to the internal memory bus 4 of FIG. 1 or the global memory 13 of FIG. 2) of the identity and changed value of the manipulated memory location of step 113, whereas step 125 of thread 111/2 does not communicate with machines M2, . . . , Mn via the relatively slow network 53. Instead, step 125 of thread 111/2 requests or otherwise notifies a different and potentially simultaneously or concurrently executing thread 121/1 of the DRT processing environment 120 to perform the notification and communication with machines M2, . . . , Mn via the relatively slow network 53 of the identify and changed value of the manipulated memory location of step 113, as indicated by arrow 127. Thus thread 111/2 carrying out step 125 is only interrupted momentarily before the thread 111/2 resumes or continues processing or execution of modified application code in step 115. The other thread 121/1 of the DRT processing environment 120 then communicates the identity and changed value of the manipulated memory location of step 113 to machines M2, . . . , Mn via the relatively slow network 53 or other relatively slow communications link or path.
  • This second arrangement of FIG. 12 makes better utilisation of the processing power of the various threads 111/1 . . . 111/3 and 121/1 (which are not, in general, subject to equal demands). Irrespective of which arrangement is used, the identity and change value of the manipulated memory location(s) of step 113 is (are) propagated to all the other machines M2 . . . Mn on the network 53 or other communications link or path.
  • This is illustrated in FIG. 13 where step 114 of FIG. 11, or the DRT 71/1 (corresponding to the DRT processing environment 120 of FIG. 12) and its thread 121/1 of FIG. 12 (represented by step 128 in FIG. 13), send, via the network 53 or other communications link or path, the identity and changed value of the manipulated memory location of step 113 of FIGS. 11 and 12, to each of the other machines M2, . . . , Mn.
  • With reference to FIG. 13, each of the other machines M2, . . . , Mn carries out the action of receiving from the network 53 the identity and changed value of, for example, the manipulated memory location of step 113 from machine M1, indicated by step 135, and writes the value received at step 135 to the local memory location corresponding to the identified memory location received at step 135, indicated by step 136.
  • In the conventional arrangement in FIG. 3 utilising distributed software, memory access from one machine's software to memory physically located on another machine is permitted by the network interconnecting the machines. However, because the read and/or write memory access to memory physically located on another computer require the use of the slow network 14, in these configurations such memory accesses can result in substantial delays in memory read/write processing operation, potentially of the order of 106-107 cycles of the central processing unit of the machine, but ultimately being dependant upon numerous factors, such as for example, the speed, bandwidth, and/or latency of the network 14. This in large part accounts for the diminished performance of the multiple interconnected machines in the prior art arrangement of FIG. 3.
  • However, in the present arrangement as described above in connection with FIG. 8, it will be appreciated that all reading of memory locations or data is satisfied locally because a current value of all (or some subset of all) memory locations is stored on the machine carrying out the processing which generates the demand to read memory.
  • Similarly, in the present arrangement as described above in connection with FIG. 8, it will be appreciated that all writing of memory locations or data may be satisfied locally because a current value of all (or some subset of all) memory locations is stored on the machine carrying out the processing which generates the demand to write to memory.
  • Such local memory read and write processing operation as performed according to the invention can typically be satisfied within 102-103 cycles of the central processing unit. Thus, in practice, there is substantially less waiting for memory accesses which involves reads than the arrangement shown and described relative to FIG. 3. Additionally, in practice, there may be less waiting for memory accesses which involve writes than the arrangement shown and described relative to FIG. 3
  • It may be appreciated that most application software reads memory frequently but writes to memory relatively infrequently. As a consequence, the rate at which memory is being written or re-written is relatively slow compared to the rate at which memory is being read. Because of this slow demand for writing or re-writing of memory, the memory locations or fields can be continually updated at a relatively low speed via the possibly relatively slow and inexpensive commodity network 53, yet this possibly relatively slow speed is sufficient to meet the application program's demand for writing to memory. The result is that the performance of the FIG. 8 arrangement is superior to that of FIG. 3. It may be appreciated in light of the description provided herein that while a relatively slow network communication link or path 53 may advantageously be used because it provides the desired performance and low cost, the invention is not limited to a relatively low speed network connection and may be used with any communication link or path. The invention is transport, network, and communications path independent, and does not depend on how the communication between machines or DRTs takes place. In one embodiment, even electronic mail (email) exchanges between machines or DRTs may suffice for the communications.
  • In a further optional modification in relation to the above, the identity and changed value pair of a manipulated memory location sent over network 53, each pair typically sent as the sole contents of a single packet, frame or cell for example, can be grouped into batches of multiple pairs of identities and changed values corresponding to multiple manipulated memory locations, and sent together over network 53 or other communications link or path in a single packet, frame, or cell. This further modification further reduces the demands on the communication speed of the network 53 or other communications link or path interconnecting the various machines, as each packet, cell or frame may contain multiple identity and changed value pairs, and therefore fewer packets, frames, or cells require to be sent.
  • It may be apparent that in an environment where the application program code writes repeatedly to a single memory location, the embodiment illustrated of FIG. 11 of step 114 sends an updating and propagation message to all machines corresponding to every performed memory manipulation operation. In a still further optimal modification in relation to the above, the DRT thread 121/1 of FIG. 12 does not need to perform an updating and propagation operation corresponding to every local memory manipulation operation, but instead may send fewer updating and propagation messages than memory manipulation operations, each message containing the last or latest changed value or content of the manipulated memory location, or optionally may only send a single updating and propagation message corresponding to the last memory manipulation operation. This further improvement reduces the demands on the network 53 or other communications link or path, as fewer packets, frames, or cells require to be sent.
  • It will also be apparent to those skilled in the art in light of the detailed description provided herein that in a table or list or other data structure created by each DRT 71 when initially recording or creating the list of all, or some subset of all, memory locations (or fields), for each such recorded memory location on each machine M1, . . . , Mn there is a name or identity which is common or similar on each of the machines M1, . . . , Mn. However, in the individual machines the local memory location corresponding to a given name or identity (listed for example, during step 91 of FIG. 9) will or may vary over time since each machine may and generally will store changed memory values or contents at different memory locations according to its own internal processes. Thus the table, or list, or other data structure in each of the DRTs will have, in general, different local memory locations corresponding to a single memory name or identity, but each global “memory name” or identity will have the same “memory value” stored in the different local memory locations.
  • It will also be apparent to those skilled in the art in light of the description provided herein that the abovementioned modification of the application program code 50 during loading can be accomplished in many ways or by a variety of means. These ways or means include, but are not limited to at least the following five ways and variations or combinations of these five, including by:
  • (i) re-compilation at loading,
  • (ii) by a pre-compilation procedure prior to loading,
  • (iii) compilation prior to loading,
  • (iv) a “just-in-time” compilation, or
  • (v) re-compilation after loading (but, or for example, before execution of the relevant or corresponding application code in a distributed environment).
  • Traditionally the term “compilation” implies a change in code or language, for example, from source to object code or one language to another. Clearly the use of the term “compilation” (and its grammatical equivalents) in the present specification is not so restricted and can also include or embrace modifications within the same code or language
  • Given the fundamental concept of modifying memory manipulation operations to coordinate operation between and amongst a plurality of machines M1 . . . Mn, there are several different ways or embodiments in which this coordinated, coherent and consistent memory state and manipulation operation concept, method, and procedure may be carried out or implemented.
  • In the first embodiment, a particular machine, say machine M2, loads the asset (such as class or object) inclusive of memory manipulation operation(s), modifies it, and then loads each of the other machines M1, M3, . . . , Mn (either sequentially or simultaneously or according to any other order, routine or procedure) with the modified object (or class or other asset or resource) inclusive of the new modified memory manipulation operation. Note that there may be one or a plurality of memory manipulation operations corresponding to only one object in the application code, or there may be a plurality of memory manipulation operations corresponding to a plurality of objects in the application code. Note that in one embodiment, the memory manipulation operation(s) that is (are) loaded is binary executable object code. Alternatively, the memory manipulation operation(s) that is (are) loaded is executable intermediary code.
  • In this arrangement, which may be termed “master/slave” each of the slave (or secondary) machines M1, M3, . . . , Mn loads the modified object (or class), and inclusive of the new modified memory manipulation operation(s), that was sent to it over the computer communications network or other communications link or path by the master (or primary) machine, such as machine M2, or some other machine such as a machine X of FIG. 15. In a slight variation of this “master/slave” or “primary/secondary” arrangement, the computer communications network can be replaced by a shared storage device such as a shared file system, or a shared document/file repository such as a shared database.
  • Note that the modification performed on each machine or computer need not and frequently will not be the same or identical. What is required is that they are modified in a similar enough way that in accordance with the inventive principles described herein, each of the plurality of machines behaves consistently and coherently relative to the other machines to accomplish the operations and objectives described herein. Furthermore, it will be appreciated in light of the description provided herein that there are a myriad of ways to implement the modifications that may for example depend on the particular hardware, architecture, operating system, application program code, or the like or different factors. It will also be appreciated that embodiments of the invention may be implemented within an operating system, outside of or without the benefit of any operating system, inside the virtual machine, in an EPROM, in software, in firmware, or in any combination of these.
  • In a still further embodiment, each machine M1, . . . , Mn receives the unmodified asset (such as class or object) inclusive of one or more memory manipulation operation(s), but modifies the operations and then loads the asset (such as class or object) consisting of the now modified operations. Although one machine, such as the master or primary machine may customize or perform a different modification to the memory manipulation operation(s) sent to each machine, this embodiment more readily enables the modification carried out by each machine to be slightly different and to be enhanced, customized, and/or optimized based upon its particular machine architecture, hardware, processor, memory, configuration, operating system, or other factors, yet still similar, coherent and consistent with other machines with all other similar modifications and characteristics that may not need to be similar or identical.
  • In all of the described instances or embodiments, the supply or the communication of the asset code (such as class code or object code) to the machines M1, . . . , Mn, and optionally inclusive of a machine X of FIG. 15, can be branched, distributed or communicated among and between the different machines in any combination or permutation; such as by providing direct machine to machine communication (for example, M2 supplies each of M1, M3, M4, etc. directly), or by providing or using cascaded or sequential communication (for example, M2 supplies M1 which then supplies M3 which then supplies M4, and so on), or a combination of the direct and cascaded and/or sequential.
  • Reference is made to the accompanying Annexure A in which: Annexure A5 is a typical code fragment from a memory manipulation operation prior to modification (e.g., an exemplary unmodified routine with a memory manipulation operation), and Annexure A6 is the same routine with a memory manipulation operation after modification (e.g., an exemplary modified routine with a memory manipulation operation). These code fragments are exemplary only and identify one software code means for performing the modification in an exemplary language. It will be appreciated that other software/firmware or computer program code may be used to accomplish the same or analogous function or operation without departing from the invention.
  • Annexures A5 and A6 (also reproduced in part in Table VI and Table VII below) are exemplary code listings that set forth the conventional or unmodified computer program software code (such as may be used in a single machine or computer environment) of a routine with a memory manipulation operation of application program code 50 and a post-modification excerpt of the same routine such as may be used in embodiments of the present invention having multiple machines. The modified code that is added to the routine is highlighted in bold text.
    TABLE I
    Summary Listing of Contents of Annexure A
    Annexure A includes exemplary program listings in the JAVA language to further
    illustrate features, aspects, methods, and procedures of described in the detailed
    description
    A1.   This first excerpt is part of an illustration of the modification code of the modifier 51 in
    accordance with steps 92 and 103 of FIG. 10.  It searches through the code array of the
    application program code 50, and when it detects a memory manipulation instruction (i.e. a
    putstatic instruction (opcode 178) in the JAVA language and virtual machine environment) it
    modifies the application program code by the insertion of an “alert” routine.
    A2.   This second excerpt is part of the DRT.alert( ) method and implements the step of 125
    and arrow of 127 of FIG. 12. This DRT.alert( ) method requests one or more threads of the
    DRT processing environment of FIG. 12 to update and propagate the value and identity of the
    changed memory location corresponding to the operation of Annexure A1.
    A3.   This third excerpt is part of the DRT 71, and corresponds to step 128 of FIG. 12.
    This code fragment shows the DRT in a separate thread, such as thread 121/1 of FIG. 12,
    after being notified or requested by step 125 and array 127, and sending the changed value
    and changed value location/identity across the network 53 to the other of the plurality of
    machines M1...Mn.
    A4.   The fourth excerpt is part of the DRT 71, and corresponds to steps 135 and 136 of
    FIG. 13. This is a fragment of code to receive a propagated identity and value pair sent by
    another DRT 71 over the network, and write the changed value to the identified memory
    location.
    A5.   The fifth excerpt is an disassembled compiled form of the example.java application of
    Annexure A7, which performs a memory manipulation operation (putstatic and putfield).
    A6.   The sixth excerpt is the disassembled compiled form of the same example application
    in Annexure A5 after modification has been performed by FieldLoader.java of Annexure A11,
    in accordance with FIG. 9 of this invention. The modifications are highlighted in bold.
    A7.   The seventh excerpt is the source-code of the example.java application used in
    excerpt A5 and A6. This example application has two memory locations (staticValue and
    instanceValue) and performs two memory manipulation operations.
    A8.   The eighth excerpt is the source-code of FieldAlert.java which corresponds to step
    125 and arrow 127 of FIG. 12, and which requests a thread 121/1 executing FieldSend.java of
    the “distributed run-time” 71 to propagate a changed value and identity pair to the other
    machines M1...Mn.
    A9.   The ninth excerpt is the source-code of FieldSend.java which corresponds to step
    128 of FIG. 12, and waits for a request/notification generated by FieldAlert.java of A8
    corresponding to step 125 and arrow 127, and which propagates a changed value/identity
    pair requested of it by FieldAlert.java, via network 53.
    A10.  The tenth excerpt is the source-code of FieldReceive.java, which corresponds to
    steps 135 and 136 of FIG. 13, and which receives a propagated changed value and identity
    pair sent to it over the network 53 via FieldSend.java of annexure A9.
    A11. FieldLoader.java. This excerpt is the source-code of FieldLoader.java, which modifies
    an application program code, such as the example.java application code of Annexure A7, as it
    is being loaded into a JAVA virtual machine in accordance with steps 90, 91, 92, 103, and 94
    of FIG. 10.  FieldLoader.java makes use of the convenience classes of Annexures A12
    through to A36 during the modification of a compiled JAVA
    A12. Attribute_info.java
    Convience class for representing attribute_info structures within ClassFiles.
    A13. ClassFile.java
    Convience class for representing ClassFile structures.
    A14. Code_attribute.java
    Convience class for representing Code_attribute structures within ClassFiles.
    A15. CONSTANT_Class_info.java
    Convience class for representing CONSTANT_Class_info structures within ClassFiles.
    A16. CONSTANT_Double_info.java
    Convience class for representing CONSTANT_Double_info structures within ClassFiles.
    A17. CONSTANT_Fieldref_info.java
    Convience class for representing CONSTANT_Fieldref_info structures within ClassFiles.
    A18. CONSTANT_Float_info.java
    Convience class for representing CONSTANT_Float_info structures within ClassFiles.
    A19.                   CONSTANT_Integer_info.java
    Convience class for representing CONSTANT_Integer_info structures within ClassFiles.
    A20. CONSTANT_InterfaceMethodref_info.java
    Convience class for representing CONSTANT_InterfaceMethodref_info structures within
    ClassFiles.
    A21. CONSTANT_Long_info.java
    Convience class for representing CONSTANT_Long_info structures within ClassFiles.
    A22. CONSTANT_Methodref_info.java
    Convience class for representing CONSTANT_Methodref_info structures within ClassFiles.
    A23. CONSTANT_NameAndType_info.java
    Convience class for representing CONSTANT_NameAndType_info structures within
    ClassFiles.
    A24. CONSTANT_String_info.java
    Convience class for representing CONSTANT_String_info structures within ClassFiles.
    A25. CONSTANT_Utf8_info.java
    Convience class for representing CONSTANT_Utf8_info structures within ClassFiles.
    A26. ConstantValue_attribute.java
    Convience class for representing ConstantValue_attribute structures within ClassFiles.
    A27. cp_info.java
    Convience class for representing cp_info structures within ClassFiles.
    A28. Deprecated_attribute.java
    Convience class for representing Deprecated_attribute structures within ClassFiles.
    A29. Exceptions_attribute.java
    Convience class for representing Exceptions_attribute structures within ClassFiles.
    A30. field_info.java
    Convience class for representing field_info structures within ClassFiles.
    A31. InnerClasses_attribute.java
    Convience class for representing InnerClasses_attribute structures within ClassFiles.
    A32. LineNumberTable_attribute.java
    Convience class for representing LineNumberTable_attribute structures within ClassFiles.
    A33. LocalVariableTable_attribute.java
    Convience class for representing LocalVariableTable_attribute structures within ClassFiles.
    A34. method_info.java
    Convience class for representing method_info structures within ClassFiles.
    A35. SourceFile_attribute.java
    Convience class for representing SourceFile_attribute structures within ClassFiles.
    A36. Synthetic_attribute.java
    Convience class for representing Synthetic_attribute structures within ClassFiles.
  • TABLE II
    Exemplary code listing showing embodiment of modified code.
    A1.  This first excerpt is part of an illustration of the modification code
    of the modifier 51 in accordance with steps 92 and 103 of FIG. 10.
    It searches through the code array of the application program code 50,
    and when it detects a memory manipulation instruction (i.e. a putstatic
    instruction (opcode 178) in the JAVA language and virtual machine
    environment) it modifies the application program code by the insertion
    of an “alert” routine.
    // START
    byte[ ] code = Code_attribute.code; // Bytecode of a given method in a
    // given classfile.
    int code_length = Code_attribute.code_length;
    int DRT = 99; // Location of the CONSTANT_Methodref_info for the
    // DRT.alert( ) method.
    for (int i=0; i<code_length; i++){
     if ((code[i] & 0xff) == 179){ // Putstatic instruction.
      System.arraycopy(code, i+3, code, i+6, code_length−(i+3));
      code[i+3] = (byte) 184; // Invokestatic instruction for the
    // DRT.alert( ) method.
      code[i+4] = (byte) ((DRT >>> 8) & 0xff);
      code[i+5] = (byte) (DRT & 0xff);
     }
    }
    // END
  • TABLE III
    Exemplary code listing showing embodiment of code for alert method
    A2.  This second excerpt is part of the DRT.alert( ) method and
    implements the step of 125 and arrow of 127 of FIG. 12. This
    DRT.alert( ) method requests one or more threads of the DRT
    processing environment of FIG. 12 to update and propagate the
    value and identity of the changed memory location corresponding to
    the operation of Annexure A1.
    // START
    public static void alert( ){
     synchronized (ALERT_LOCK){
      ALERT_LOCK.notify( ); // Alerts a waiting DRT thread in the
      background.
     }
    }
    // END
  • TABLE IV
    Exemplary code listing showing embodiment of code for DRT
    A3.  This third excerpt is part of the DRT 71, and corresponds to step
    128 of FIG. 12. This code fragment shows the DRT in a separate thread,
    such as thread 121/1 of FIG. 12, after being notified or requested by
    step 125 and array 127, and sending the changed value and changed
    value location/identity across the network 53 to the other of the
    plurality of machines M1...Mn.
    // START
    MulticastSocket ms = DRT.getMulticastSocket( ); // The multicast socket
    // used by the DRT for
    // communication.
    byte nameTag = 33; // This is the “name tag” on the network for this
    // field.
    Field field = modifiedClass.getDeclaredField(“myField1”); // Stores
    // the field
    // from the
    // modified
    // class.
    // In this example, the field is a byte field.
    while (DRT.isRunning( )){
     synchronized (ALERT_LOCK){
      ALERT_LOCK.wait( ); // The DRT thread is waiting for the alert
    // method to be called.
      byte[ ] b = new byte[ ]{nameTag, field.getByte(null)}; // Stores
    // the
    // nameTag
    // and the
    // value
    // of the
    // field from
    // the
    // modified
    // class in a
    buffer.
      DatagramPacket dp = new DatagramPacket(b, 0, b.length);
      ms.send(dp); // Send the buffer out across the network.
     }
    }
    // END
  • TABLE V
    Exemplary code listing showing embodiment of code for DRT receiving.
    A4.  The fourth excerpt is part of the DRT 71, and corresponds to steps
    135 and 136 of FIG. 13.  This is a fragment of code to receive a
    propagated identity and value pair sent by another DRT 71 over the
    network, and write the changed value to the identified memory location.
    // START
    MulticastSocket ms = DRT.getMulticastSocket( ); // The multicast socket
    // used by the DRT for
    // communication.
    DatagramPacket dp = new DatagramPacket(new byte[2], 0, 2);
    byte nameTag = 33; // This is the “name tag” on the network for this
    // field.
    Field field = modifiedClass.getDeclaredField(“myField1”); // Stores the
    // field from
    // the
    // modified
    class.
    // In this example, the field is a byte field.
    while (DRT.isRunning){
     ms.receive(dp); // Receive the previously sent buffer from the network.
     byte[ ] b = dp.getData( );
     if (b[0] == nameTag){ // Check the nametags match.
      field.setByte(null, b[1]); // Write the value from the network packet
    // into the field location in memory.
     }
    }
    // END
  • TABLE VI
    Exemplary code listing showing embodiment of application before
    modification is made.
    A5.  The fifth excerpt is an disassembled compiled form of the
    example.java application of Annexure A7, which performs a memory
    manipulation operation (putstatic and putfield).
    Method void setValues(int, int)
     0 iload_1
     1 putstatic #3 <Field int staticValue>
     4 aload_0
     5 iload_2
     6 putfield #2 <Field int instanceValue>
     9 return
  • TABLE VII
    Exemplary code listing showing embodiment of application
    after modification is made.
    A6.  The sixth excerpt is the disassembled compiled form of the same
    example application in Annexure A5 after modification has been
    performed by FieldLoader.java of Annexure A11, in accordance with
    FIG. 9 of this invention. The modifications are highlighted in bold.
    Method void setValues(int, int)
      0 iload_1
      1 putstatic #3 <Field int staticValue>
      4 ldc #4 <String “example”>
      6 iconst 0
      7 invokestatic #5 <Method void alert(java.lang.Object, int)>
     10 aload_0
     11 iload_2
     12 putfield #2 <Field int instanceValue>
    15 aload 0
    16 iconst 1
    17 invokestatic #5 <Method void alert(java.lang.Object, int)>
     20 return
  • TABLE VIII
    Exemplary code listing showing embodiment of source-code
    of the example application.
    A7.  The seventh excerpt is the source-code of the example.java
    application used in excerpt A5 and A6.  This example application
    has two memory locations (staticValue and instanceValue) and
    performs two memory manipulation operations.
    import java.lang.*;
    public class example{
     /** Shared static field. */
     public static int staticValue = 0;
     /** Shared instance field. */
     public int instanceValue = 0;
     /** Example method that writes to memory (instance field). */
     public void setValues(int a, int b){
      staticValue = a;
      instanceValue = b;
     }
    }
  • TABLE IX
    Exemplary code listing showing embodiment
    of the source-code of FieldAlert.
    A8.  The eighth excerpt is the source-code of FieldAlert.java which
    corresponds to step 125 and arrow 127 of FIG. 12, and which requests
    a thread 121/1 executing FieldSend.java of the “distributed run-time”
    71 to propagate a changed value and identity pair to the other
    machines M1...Mn.
    import java.lang.*;
    import java.util.*;
    import java.net.*;
    import java.io.*;
    public class FieldAlert{
     /** Table of alerts. */
     public final static Hashtable alerts = new Hashtable( );
     /** Object handle. */
     public Object reference = null;
     /** Table of field alerts for this object. */
     public boolean[ ] fieldAlerts = null;
     /** Constructor. */
     public FieldAlert(Object o, int initialFieldCount){
      reference = o;
      fieldAlerts = new boolean[initialFieldCount];
     }
     /** Called when an application modifies a value. (Both objects and
      classes) */
     public static void alert(Object o, int fieldID){
      // Lock the alerts table.
      synchronized (alerts){
       FieldAlert alert = (FieldAlert) alerts.get(o);
       if (alert == null){ // This object hasn't been alerted already,
    // so add to alerts table.
        alert = new FieldAlert(o, fieldID + 1);
        alerts.put(o, alert);
       }
       if (fieldID >= alert.fieldAlerts.length){
        // Ok, enlarge fieldAlerts array.
        boolean[ ] b = new boolean[fieldID+1];
        System.arraycopy(alert.fieldAlerts, 0, b, 0,
         alert.fieldAlerts.length);
        alert.fieldAlerts = b;
       }
       // Record the alert.
       alert.fieldAlerts[fieldID] = true;
       // Mark as pending.
       FieldSend.pending = true; // Signal that there is one or more
    // propagations waiting.
       // Finally, notify the waiting FieldSend thread(s)
       if (FieldSend.waiting){
        FieldSend.waiting = false;
       alerts.notify( );
       }
      }
     }
    }
  • It is noted that the compiled code in the annexure and portion repeated in the table is taken from the source-code of the file “example.java” which is included in the Annexure A7 (Table VIII). In the procedure of Annexure A5 and Table VI, the procedure name “Method void setValues(int, int)” of Step 001 is the name of the displayed disassembled output of the setValues method of the compiled application code of “example.java”. The name “Method void setvalues(int, int)” is arbitrary and selected for this example to indicate a typical JAVA method inclusive of a memory manipulation operation. Overall the method is responsible for writing two values to two different memory locations through the use of an memory manipulation assignment statement (being “putstatic” and “putfield” in this example) and the steps to accomplish this are described in turn.
  • First (Step 002), the Java Virtual Machine instruction “iload 1” causes the Java Virtual Machine to load the integer value in the local variable array at index 1 of the current method frame and store this item on the top of the stack of the current method frame and results in the integer value passed to this method as the first argument and stored in the local variable array at index 1 being pushed onto the stack.
  • The Java Virtual Machine instruction “putstatic #3<Field int staticValue>” (Step 003) causes the Java Virtual Machine to pop the topmost value off the stack of the current method frame and store the value in the static field indicated by the CONSTANT_Fieldref_info constant-pool item stored in the 3rd index of the classfile structure of the application program containing this example setValues( ) method and results in the topmost integer value of the stack of the current method frame being stored in the integer field named “staticValue”.
  • The Java Virtual Machine instruction “aload0” (Step 004) causes the Java Virtual Machine to load the item in the local variable array at index 0 of the current method frame and store this item on the top of the stack of the current method frame and results in the ‘this’ object reference stored in the local variable array at index 0 being pushed onto the stack.
  • First (Step 005), the Java Virtual Machine instruction “iload 2” causes the Java Virtual Machine to load the integer value in the local variable array at index 2 of the current method frame and store this item on the top of the stack of the current method frame and results in the integer value passed to this method as the first argument and stored in the local variable array at index 2 being pushed onto the stack.
  • The Java Virtual Machine instruction “puffield #2<Field int instanceValue>” (Step 006) causes the Java Virtual Machine to pop the two topmost values off the stack of the current method frame and store the topmost value in the object instance field of the second popped value, indicated by the CONSTANT_Fieldref_info constant-pool item stored in the 2nd index of the classfile structure of the application program containing this example setValues method and results in the integer value on the top of the stack of the current method frame being stored in the instance field named “instanceValue” of the object reference below the integer value on the stack.
  • Finally, the JAVA virtual machine instruction “return” (Step 007) causes the JAVA virtual machine to cease executing this setValues( ) method by returning control to the previous method frame and results in termination of execution of this setValues( ) method.
  • As a result of these steps operating on a single machine of the conventional configurations in FIG. 1 and FIG. 2, the JAVA virtual machine manipulates (i.e. writes to) the staticValue and instanceValue memory locations, and in executing the setValues( ) method containing the memory manipulation operation(s) is able to ensure that memory is and remains consistent between multiple threads of a single application instance, and therefore ensure that unwanted behaviour, such as for example inconsistent or incoherent memory between multiple threads of a single application instance (such inconsistent or incoherent memory being for example incorrect or different values or contents with respect to a single memory location) does not occur. Were these steps to be carried out on the plurality of machines of the configurations of FIG. 5 and FIG. 8 by concurrently executing the application program code 50 on each one of the plurality of machines M1 . . . Mn, the memory manipulation operations of each concurrently executing application program occurrence on each one of the machines would be performed without coordination between any other machine(s), such coordination being for example updating of corresponding memory locations on each machine such that they each report a same content or value. Given the goal of consistent, coordinated and coherent memory state and manipulation and updating operation across a plurality of a machines, this prior art arrangement would fail to perform such consistent, coherent, and coordinated memory state and manipulation and updating operation across the plurality of machines, as each machine performs memory manipulation only locally and without any attempt to coordinate or update their local memory state and manipulation operation with any other similar memory state on any one or more other machines. Such an arrangement would therefore be susceptible to inconsistent and incoherent memory state amongst machines M1 . . . Mn due to uncoordinated, inconsistent and/or incoherent memory manipulation and updating operation. Therefore it is the goal of the present invention to overcome this limitation of the prior art arrangement.
  • In the exemplary code in Table VII (Annexure A6), the code has been modified so that it solves the problem of consistent, coordinated memory manipulation and updating operation for a plurality of machines M1 . . . Mn, that was not solved in the code example from Table VI (Annexure A5). In this modified setValues( ) method code, an “ldc #4<String “example”>” instruction is inserted after the “putstatic #3” instruction in order to be the first instruction following the execution of the “putstatic #3” instruction. This causes the JAVA virtual machine to load the String value “example” onto the stack of the current method frame and results in the String value of “example” loaded onto the top of the stack of the current method frame. This change is significant because it modifies the setValues( ) method to load a String identifier corresponding to the classname of the class containing the static field location written to by the “putstatic #3” instruction onto the stack.
  • Furthermore, the JAVA virtual machine instruction “iconst0” is inserted after the “ldc #4” instruction so that the JAVA virtual machine loads an integer value of “0” onto the stack of the current method frame and results in the integer value of “0” loaded onto the top of the stack of the current method frame. This change is significant because it modifies the setValues( ) method to load an integer value, which in this example is “0”, which represents the identity of the memory location (field) manipulated by the preceding “putstatic #3” operation. It is to be noted that the choice or particular form of the memory identifier used for the implementation of this invention is for illustration purposes only. In this example, the integer value of “0” is the identifier used of the manipulated memory location, and corresponds to the “staticValue” field as the first field of the “example.java” application, as shown in Annexure A7. Therefore, corresponding to the “putstatic #3” instruction, the “iconst0” instruction loads the integer value “0” corresponding to the index of the manipulated field of the “putstatic #3” instruction, and which in this case is the first field of “example.java” hence the “0” integer index value, onto the stack.
  • Additionally, the JAVA virtual machine instruction “invokestatic #5<Method boolean alert(java.lang.Object, int)>” is inserted after the “iconst0” instruction so that the JAVA virtual machine pops the two topmost items off the stack of the current method frame (which in accordance with the preceding “ldc #4” instruction is a reference to the String object with the value “example” corresponding to the name of the class to which manipulated field belongs, and the integer “0” corresponding to the index of the manipulated field in the example.java application) and invokes the “alert” method, passing the two topmost items popped off the stack to the new method frame as its first two arguments. This change is significant because it modifies the setValues( ) method to execute the “alert” method and associated operations, corresponding to the preceding memory manipulation operation (that is, the “putstatic #3” instruction) of the setValues( ) method.
  • Likewise, in this modified setValues( ) method code, an “aload0” instruction is inserted after the “puffield #2” instruction in order to be the first instruction following the execution of the “puffield #2” instruction. This causes the JAVA virtual machine to load the instance object of the example class to which the manipulated field of the preceding “puffield #2” instruction belongs, onto the stack of the current method frame and results in the object reference corresponding to the instance field written to by the “puffield #2” instruction, loaded onto the top of the stack of the current method frame. This change is significant because it modifies the setValues( ) method to load a reference to the object corresponding to the manipulated field onto the stack.
  • Furthermore, the JAVA virtual machine instruction “iconst 1” is inserted after the “aload0” instruction so that the JAVA virtual machine loads an integer value of “1” onto the stack of the current method frame and results in the integer value of “1” loaded onto the top of the stack of the current method frame. This change is significant because it modifies the setValues( ) method to load an integer value, which in this example is “1”, which represents the identity of the memory location (field) manipulated by the preceding “puffield #2” operation. It is to be noted that the choice or particular form of the identifier used for the implementation of this invention is for illustration purposes only. In this example, the integer value of “1” corresponds to the “instanceValue” field as the second field of the “example.java” application, as shown in Annexure A7. Therefore, corresponding to the “puffield #2” instruction, the “iconst 1” instruction loads the integer value “1” corresponding to the index of the manipulated field of the “puffield #2” instruction, and which in this case is the second field of “example.java” hence the “1” integer index value, onto the stack.
  • Additionally, the JAVA virtual machine instruction “invokestatic #5<Method boolean alert(java.lang.Object, int)>” is inserted after the “iconst 1” instruction so that the JAVA virtual machine pops the two topmost item off the stack of the current method frame (which in accordance with the preceding “aload0” instruction is a reference to the object corresponding to the object to which the manipulated instance field belongs, and the integer “1” corresponding to the index of the manipulated field in the example.java application) and invokes the “alert” method, passing the two topmost items popped off the stack to the new method frame as its first two arguments. This change is significant because it modifies the setValues( ) method to execute the “alert” method and associated operations, corresponding to the preceding memory manipulation operation (that is, the “puffield #2” instruction) of the setValues( ) method.
  • The method void alert(java.lang.Object, int), part of the FieldAlert code of Annexure A8 and part of the distributed runtime system (DRT) 71, requests or otherwise notifies a DRT thread 121/1 executing the FieldSend.java code of Annexure A9 to update and propagate the changed identity and value of the manipulated memory location to the plurality of machines M1 . . . Mn.
  • It will be appreciated that the modified code permits, in a distributed computing environment having a plurality of computers or computing machines, the coordinated operation of memory manipulation operations so that the problems associated with the operation of the unmodified code or procedure on a plurality of machines M1 . . . Mn (such as for example inconsistent and incoherent memory state and manipulation and updating operation) does not occur when applying the modified code or procedure.
  • Having now described aspects of the memory management and replication, attention is now directed to an exemplary operational scenario illustrating the manner in which application programs on two computers may simultaneously execute the same application program in a consistent, coherent manner.
  • In this regard, attention is directed to FIGS. 14-16, two laptop computers 101 and 102 are illustrated. The computers 101 and 102 are not necessarily identical and indeed, one can be an IBM or IBM-clone and the other can be an APPLE computer. The computers 101 and 102 have two screens 105, 115 two keyboards 106, 116 but a single mouse 107. The two machines 101, 102 are interconnected by a means of a single coaxial cable or twisted pair cable 314.
  • Two simple application programs are downloaded onto each of the machines 101, 102, the programs being modified as they are being loaded as described above. In this embodiment the first application is a simple calculator program and results in the image of a calculator 108 being displayed on the screen 105. The second program is a graphics program which displays four coloured blocks 109 which are of different colours and which move about at random within a rectangular box 310. Again, after loading, the box 310 is displayed on the screen 105. Each application operates independently so that the blocks 109 are in random motion on the screen 105 whilst numerals within the calculator 108 can be selected (with the mouse 107) together with a mathematical operator (such as addition or multiplication) so that the calculator 108 displays the result.
  • The mouse 107 can be used to “grab” the box 310 and move same to the right across the screen 105 and onto the screen 115 so as to arrive at the situation illustrated in FIG. 15. In this arrangement, the calculator application is being conducted on machine 101 whilst the graphics application resulting in display of box 310 is being conducted on machine 102.
  • However, as illustrated in FIG. 16, it is possible by means of the mouse 107 to drag the calculator 108 to the right as seen in FIG. 15 so as to have a part of the calculator 108 displayed by each of the screens 105, 115. Similarly, the box 310 can be dragged by means of the mouse 107 to the left as seen in FIG. 15 so that the box 310 is partially displayed by each of the screens 105, 115 as indicated FIG. 16. In this configuration, part of the calculator operation is being performed on machine 101 and part on machine 102 whilst part of the graphics application is being carried out the machine 101 and the remainder is carried out on machine 102.
  • FURTHER DESCRIPTION
  • The foregoing describes only some embodiments of the present invention and modifications, obvious to those skilled in the art, can be made thereto without departing from the scope of the present invention. For example, reference to JAVA includes both the JAVA language and also JAVA platform and architecture.
  • In all described instances of modification, where the application code 50 is modified before, or during loading, or even after loading but before execution of the unmodified application code has commenced, it is to be understood that the modified application code is loaded in place of, and executed in place of, the unmodified application code subsequently to the modifications being performed.
  • Alternatively, in the instances where modification takes place after loading and after execution of the unmodified application code has commenced, it is to be understood that the unmodified application code may either be replaced with the modified application code in whole, corresponding to the modifications being performed, or alternatively, the unmodified application code may be replaced in part or incrementally as the modifications are performed incrementally on the executing unmodified application code. Regardless of which such modification routes are used, the modifications subsequent to being performed execute in place of the unmodified application code.
  • An advantage of using a global identifier in the invention described is as a form of ‘meta-name’ or ‘meta-identity’ for all the similar equivalent local objects (or classes, or assets or resources or the like) on each one of the plurality of machines M1, . . . , Mn. For example, rather than having to keep track of each unique local name or identity of each similar equivalent local object on each machine of the plurality of similar equivalent objects, one may instead define or use a global name corresponding to the plurality of similar equivalent objects on each machine (e.g., “globalname7787”), and with the understanding that each machine relates the global name to a specific local name or object (e.g., “globalname7787” corresponds to object “localobject456” on machine M1, and “globalname7787” corresponds to object “localobject885” on machine M2, and “globalname7787” corresponds to object “localobject111” on machine M3, and so forth).
  • Those skilled in the programming arts will be aware that when additional code or instructions is/are inserted into an existing code or instruction set to modify same, the existing code or instruction set may well require further modification (such as for example, by re-numbering of sequential instructions) so that offsets, branching, attributes, mark up and the like are catered for.
  • Similarly, in the JAVA language memory locations include, for example, both fields and array types. The above description deals with fields and the changes required for array types are essentially the same mutatis mutandis. Also the present invention is equally applicable to similar programming languages (including procedural, declarative and object orientated) to JAVA including Micrsoft.NET platform and architecture (Visual Basic, Visual C/C++, and C#) FORTRAN, C/C++, COBOL, BASIC etc.
  • The abovementioned arrangement, in which the JAVA code which updates memory locations or field values is modified, is based on the assumption that either the runtime system (say, JAVA HOTSPOT VIRTUAL MACHINE written in C and Java) or the operating system (LINUX written in C and Assembler, for example) of each machine M1 . . . Mn will ordinarily update memory on the local machine (say M2) but not on any corresponding other machines (M1, M3 . . . Mn). It is possible to leave the JAVA code which updates memory locations or field values unamended and instead amend the LINUX or HOTSPOT routine which updates memory locally, so that it correspondingly updates memory on all other machines as well. In order to embrace such an arrangement the term “updating propagation routine” used herein in conjunction with maintaining the memory of all machines M1 . . . Mn essentially the same, is to be understood to include within its scope both the JAVA putfield and putstatic instructions and related operations and the “combination” of the JAVA putfield and putstatic operations and the LINUX or HOTSPOT code fragments which perform memory updating.
  • The terms object and class used herein are derived from the JAVA environment and are intended to embrace similar terms derived from different environments such as dynamically linked libraries (DLL), or object code packages, or function unit or memory locations.
  • Various means are described relative to embodiments of the invention, including for example but not limited to lock means, distributed run time means, modifier or modifying means, propagation means, distribution update means, counter means, synchronization means, and the like. In at least one embodiment of the invention, any one or each of these various means may be implemented by computer program code statements or instructions (possibly including by a plurality of computer program code statements or instructions) that execute within computer logic circuits, processors, ASICs, microprocessors, microcontrollers, or other logic to modify the operation of such logic or circuits to accomplish the recited operation or function. In another embodiment, any one or each of these various means may be implemented in firmware and in other embodiments such may be implemented in hardware. Furthermore, in at least one embodiment of the invention, any one or each of these various means may be implemented by an combination of computer program software, firmware, and/or hardware.
  • Any and each of the aforedescribed methods, procedures, and/or routines may advantageously be implemented as a computer program and/or computer program product stored on any tangible media or existing in electronic, signal, or digital form. Such computer program or computer program products comprising instructions separately and/or organized as modules, programs, subroutines, or in any other way for execution in processing logic such as in a processor or microprocessor of a computer, computing machine, or information appliance; the computer program or computer program products modifying the operation of the computer on which it executes or on a computer coupled with, connected to, or otherwise in signal communications with the computer on which the computer program or computer program product is present or executing. Such computer program or computer program product modifying the operation and architectural structure of the computer, computing machine, and/or information appliance to alter the technical operation of the computer and realize the technical effects described herein.
  • The invention may therefore includes a computer program product comprising a set of program instructions stored in a storage medium or exiting electronically in any form and operable to permit a plurality of computers to carry out any of the methods, procedures, routines, or the like as described herein including in any of the claims.
  • Furthermore, the invention may include a plurality of computers interconnected via a communication network or other communications ink or path and each operable to substantially simultaneously or concurrently execute the same or a different portion of an application program code written to operate on only a single computer on a corresponding different one of computers, wherein the computers being programmed to carry out any of the methods, procedures, or routines described in the specification or set forth in any of the claims, or being loaded with a computer program product.
  • The term “comprising” (and its grammatical variations) as used herein is used in the inclusive sense of “having” or “including” and not in the exclusive sense of “consisting only of”.
  • Copyright Notice
  • This patent specification and the Annexures which form a part thereof contains material which is subject to copyright protection. The copyright owner (which is the applicant) has no objection to the reproduction of this patent specification or related materials from publicly available associated Patent Office files for the purposes of review, but otherwise reserves all copyright whatsoever. In particular, the various instructions are not to be entered into a computer without the specific written approval of the copyright owner.
  • Computer Architecture and Method of Operation for Multi-Computer Distributed Processing with Replicated Memory Annexure A Annexure A
  • The following are program listings in the JAVA language:
  • A1. This first excerpt is part of an illustration of the modification code of the modifier 51 in accordance with steps 92 and 103 of FIG. 10. It searches through the code array of the application program code 50, and when it detects a memory manipulation instruction (i.e. a putstatic instruction (opcode 178) in the JAVA language and virtual machine environment) it modifies the application program code by the insertion of an “alert” routine.
    // START
    byte[ ] code = Code_attribute.code; // Bytecode of a given method in a
    // given classfile.
    int code_length = Code_attribute.code_length;
    int DRT = 99; // Location of the CONSTANT_Methodref_info for the
    // DRT.alert( ) method.
    for (int i=0; i<code_length; i++){
     if ((code[i] & 0xff) == 179){ // Putstatic instruction.
      System.arraycopy(code, i+3, code, i+6, code_length−(i+3));
      code[i+3] = (byte) 184; // Invokestatic instruction for the
    // DRT.alert( ) method.
      code[i+4] = (byte) ((DRT >>> 8) & 0xff);
      code[i+5] = (byte) (DRT & 0xff);
     }
    }
    // END
  • A2. This second excerpt is part of the DRT.alert( ) method and implements the step of 125 and arrow of 127 of FIG. 12. This DRT.alert( ) method requests one or more threads of the DRT processing environment of FIG. 12 to update and propagate the value and identity of the changed memory location corresponding to the operation of Annexure A1.
    // START
    public static void alert( ){
     synchronized (ALERT_LOCK){
      ALERT_LOCK.notify( ); // Alerts a waiting DRT thread
      in the background.
     }
    }
    // END
  • A3. This third excerpt is part of the DRT 71, and corresponds to step 128 of FIG. 12. This code fragment shows the DRT in a separate thread, such as thread 121/1 of FIG. 12, after being notified or requested by step 125 and array 127, and sending the changed value and changed value location/identity across the network 53 to the other of the plulality of machines M1 . . . Mn.
    // START
    MulticastSocket ms = DRT.getMulticastSocket( ); // The multicast socket
    // used by the DRT for
    // communication.
    byte nameTag = 33; // This is the “name tag” on the network for this
    // field.
    Field field = modifiedClass.getDeclaredField(“myField1”); // Stores
    // the field
    // from the
    // modified
    // class.
    // In this example, the field is a byte field.
    while (DRT.isRunning( )){
     synchronized (ALERT_LOCK){
      ALERT_LOCK.wait( ); // The DRT thread is waiting for the alert
    // method to be called.
      byte[ ] b = new byte[ ]{nameTag, field.getByte(null)}; // Stores
    // the
    // nameTag
    // and the
    // value
    // of the
    // field from
    // the
    // modified
    // class in a
    buffer.
      DatagramPacket dp = new DatagramPacket(b, 0, b.length);
      ms.send(dp); // Send the buffer out across the network.
     }
    }
    // END
  • A4. The fourth excerpt is part of the DRT 71, and corresponds to steps 135 and 136 of FIG. 13. This is a fragment of code to receive a propagated identity and value pair sent by another DRT 71 over the network, and write the changed value to the idenitified memory location.
    // START
    MulticastSocket ms = DRT.getMulticastSocket( ); // The multicast socket
    // used by the DRT for
    // communication.
    DatagramPacket dp = new DatagramPacket(new byte[2], 0, 2);
    byte nameTag = 33; // This is the “name tag” on the network for this
    // field.
    Field field = modifiedClass.getDeclaredField(“myField1”); // Stores the
    // field from
    // the
    // modified
    class.
    // In this example, the field is a byte field.
    while (DRT.isRunning){
     ms.receive(dp); // Receive the previously sent buffer from the network.
     byte[ ] b = dp.getData( );
     if (b[0] == nameTag){ // Check the nametags match.
      field.setByte(null, b[1]); // Write the value from the network packet
    // into the field location in memory.
     }
    }
    // END
  • A5. The fifth excerpt is an dissassembled compiled form of the example.java application of Annexure A7, which performs a memory manipulation operation (putstatic and putfield).
    Method void setValues(int, int)
     0 iload_1
     1 putstatic #3 <Field int staticValue>
     4 aload_0
     5 iload_2
     6 putfield #2 <Field int instanceValue>
     9 return
  • A6. The sixth excerpt is the dissassembled compiled form of the same example application in Annexure A5 after modification has been performed by FieldLoader.java of Annexure A11, in accordance with FIG. 9 of this invention. The modifications are highlighted in bold.
    Method void setValues(int, int)
     0 iload_1
     1 putstatic #3 <Field int staticValue>
    4 ldc #4 <String “example”>
    6 iconst _0
    7 invokestatic #5 <Method void alert(java.lang.Object, int)>
     10 aload_0
     11 iload_2
     12 putfield #2 <Field int instanceValue>
    15 aload_0
    16 iconst_1
    17 invokestatic #5 <Method void alert(java.lang.Object, int)>
     20 return
  • A7. The seventh excerpt is the source-code of the example.java application used in excerpt A5 and A6. This example application has two memory locations (staticValue and instanceValue) and performs two memory manipulation operations.
    import java.lang.*;
    public class example{
     /** Shared static field. */
     public static int staticValue = 0;
     /** Shared instance field. */
     public int instanceValue = 0;
     /** Example method that writes to memory (instance field). */
     public void setValues(int a, int b){
      staticValue = a;
      instanceValue = b;
     }
    }
  • A8. The eighth excerpt is the source-code of FieldAlert.java which corresponds to step 125 and arrow 127 of FIG. 12, and which requests a thread 121/1 executing FieldSend.java of the “distributed run-time” 71 to propagate a changed value and identity pair to the other machines M1 . . . Mn.
    import java.lang.*;
    import java.util.*;
    import java.net.*;
    import java.io.*;
    public class FieldAlert{
     /** Table of alerts. */
     public final static Hashtable alerts = new Hashtable( );
     /** Object handle. */
     public Object reference = null;
     /** Table of field alerts for this object. */
     public boolean[ ] fieldAlerts = null;
     /** Constructor. */
     public FieldAlert(Object o, int initialFieldCount){
      reference = o;
      fieldAlerts = new boolean[initialFieldCount];
     }
     /** Called when an application modifies a value. (Both objects and
      classes) */
     public static void alert(Object o, int fieldID){
      // Lock the alerts table.
      synchronized (alerts){
       FieldAlert alert = (FieldAlert) alerts.get(o);
       if (alert == null){ // This object hasn't been alerted already,
    // so add to alerts table.
        alert = new FieldAlert(o, fieldID + 1);
        alerts.put(o, alert);
       }
       if (fieldID >= alert.fieldAlerts.length){
        // Ok, enlarge fieldAlerts array.
        boolean[ ] b = new boolean[fieldID+1];
        System.arraycopy(alert.fieldAlerts, 0, b, 0,
         alert.fieldAlerts.length);
        alert.fieldAlerts = b;
       }
       // Record the alert.
       alert.fieldAlerts[fieldID] = true;
       // Mark as pending.
       FieldSend.pending = true; // Signal that there is one or more
    // propagations waiting.
       // Finally, notify the waiting FieldSend thread(s)
       if (FieldSend.waiting){
        FieldSend.waiting = false;
        alerts.notify( );
       }
      }
     }
    }
  • A9. The ninth excerpt is the source-code of FieldSend.java which corresponds to step 128 of FIG. 12, and waits for a request/notification generated by FieldAlert.java of A8 corresponding to step 125 and arrow 127, and which propagates a changed value/identity pair requested of it by FieldAlert.java, via network 53.
    import java.lang.*;
    import java.lang. reflect.*;
    import java.util.*;
    import java.net.*;
    import java.io.*;
    public class FieldSend implements Runnable{
     /** Protocol specific values. */
     public final static int CLOSE = −1;
     public final static int NACK = 0;
     public final static int ACK = 1;
     public final static int PROPAGATE_OBJECT = 10;
     public final static int PROPAGATE_CLASS = 20;
     /** FieldAlert network values. */
     public final static String group =
      System.getProperty(“FieldAlert_network_group”);
     public final static int port =
      Integer.parseInt(System.getProperty(“FieldAlert_network_port”));
     /** Table of global ID's for local objects. (hashcode-to-globalID
      mappings) */
     public final static Hashtable objectToGlobalID = new Hashtable( );
     /** Table of global ID's for local classnames. (classname-to-globalID
      mappings) */
     public final static Hashtable classNameToGlobalID = new Hashtable( );
     /** Pending. True if a propagation is pending. */
     public static boolean pending = false;
     /** Waiting. True if the FieldSend thread(s) are waiting. */
     public static boolean waiting = false;
     /** Background send thread. Propagates values as this thread is alerted
      to their alteration. */
     public void run( ){
      System.out.println(“FieldAlert_network_group=” + group);
      System.out.println(“FieldAlert_network_port=” + port);
      try{
       // Create a DatagramSocket to send propagated field values.
       DatagramSocket datagramSocket =
        new DatagramSocket(port, InetAddress.getByName(group));
       // Next, create the buffer and packet for all transmissions.
       byte[ ] buffer = new byte[512]; // Working limit of 512 bytes
    // per packet.
       DatagramPacket datagramPacket =
        new DatagramPacket(buffer, 0, buffer.length);
       while (!Thread.interrupted( )){
        Object[ ] entries = null;
        // Lock the alerts table.
        synchronized (FieldAlert.alerts){
         // Await for an alert to propagate something.
         while (!pending){
          waiting = true;
          FieldAlert.alerts.wait( );
          waiting = false;
         }
         pending = false;
         entries = FieldAlert.alerts.entrySet( ).toArray( );
         // Clear alerts once we have copied them.
         FieldAlert.alerts.clear( );
        }
        // Process each object alert in turn.
        for (int i=0; i<entries.length; i++){
         FieldAlert alert = (FieldAlert) entries[i];
         int index = 0;
         datagramPacket.setLength(buffer.length);
         Object reference = null;
         if (alert.reference instanceof String){
          // PROPAGATE_CLASS field operation.
          buffer[index++] = (byte)
          ((PROPAGATE_CLASS >> 24) & 0xff);
          buffer[index++] = (byte)
          ((PROPAGATE_CLASS >> 16) & 0xff);
          buffer[index++] = (byte)
          ((PROPAGATE_CLASS >> 8) & 0xff);
          buffer[index++] = (byte)
          ((PROPAGATE_CLASS >> 0) & 0xff);
          String name = (String) alert.reference;
          int length = name.length( );
          buffer[index++] = (byte) ((length >> 24) & 0xff);
          buffer[index++] = (byte) ((length >> 16) & 0xff);
          buffer[index++] = (byte) ((length >> 8) & 0xff);
          buffer[index++] = (byte) ((length >> 0) & 0xff);
          byte[ ] bytes = name.getBytes( );
          System.arraycopy(bytes, 0, buffer, index, length);
          index += length;
         }else{      // PROPAGATE_OBJECT
         field operation.
          buffer[index++] =
           (byte) ((PROPAGATE_OBJECT >> 24) & 0xff);
          buffer[index++] =
           (byte) ((PROPAGATE_OBJECT >> 16) & 0xff);
          buffer[index++] = (byte)
          ((PROPAGATE_OBJECT >> 8) & 0xff);
          buffer[index++] = (byte)
          ((PROPAGATE_OBJECT >> 0) & 0xff);
          int globalID = ((Integer)
           objectToGlobalID.get(alert.reference)).intValue( );
          buffer[index++] = (byte) ((globalID >> 24) & 0xff);
          buffer[index++] = (byte) ((globalID >> 16) & 0xff);
          buffer[index++] = (byte) ((globalID >> 8) & 0xff);
          buffer[index++] = (byte) ((globalID >> 0) & 0xff);
          reference = alert.reference;
         }
         // Use reflection to get a table of fields that correspond to
         // the field indexes used internally.
         Field[ ] fields = null;
         if (reference == null){
          fields = FieldLoader.loadClass((String)
           alert.reference).getDeclaredFields( );
         }else{
          fields = alert.reference.getClass( ).getDeclaredFields( );
         }
         // Now encode in batch mode the fieldID/value pairs.
         for (int j=0; j<alert.fieldAlerts.length; j++){
          if (alert.fieldAlerts[j] == false)
           continue;
          buffer[index++] = (byte) ((j >> 24) & 0xff);
          buffer[index++] = (byte) ((j >> 16) & 0xff);
          buffer[index++] = (byte) ((j >> 8) & 0xff);
          buffer[index++] = (byte) ((j >> 0) & 0xff);
          // Encode value.
          Class type = fields[j].getType( );
          if (type == Boolean.TYPE){
           buffer[index++] =(byte)
            (fields[j].getBoolean(reference)? 1 : 0);
          }else if (type == Byte.TYPE){
           buffer[index++] = fields[j].getByte(reference);
          }else if (type == Short.TYPE){
           short v = fields[j].getShort(reference);
           buffer[index++] = (byte) ((v >> 8) & 0xff);
           buffer[index++] = (byte) ((v >> 0) & 0xff);
          }else if (type == Character.TYPE){
           char v = fields[j].getChar(reference);
           buffer[index++] = (byte) ((v >> 8) & 0xff);
           buffer[index++] = (byte) ((v >> 0) & 0xff);
          }else if (type == Integer.TYPE){
           int v = fields[j].getInt(reference);
           buffer[index++] = (byte) ((v >> 24) & 0xff);
           buffer[index++] = (byte) ((v >> 16) & 0xff);
           buffer[index++] = (byte) ((v >> 8) & 0xff);
           buffer[index++] = (byte) ((v >> 0) & 0xff);
          }else if (type == Float.TYPE){
           int v = Float.floatToIntBits(
            fields[j].getFloat(reference));
           buffer[index++] = (byte) ((v >> 24) & 0xff);
           buffer[index++] = (byte) ((v >> 16) & 0xff);
           buffer[index++] = (byte) ((v >> 8) & 0xff);
           buffer[index++] = (byte) ((v >> 0) & 0xff);
          }else if (type == Long.TYPE){
           long v = fields[j].getLong(reference);
           buffer[index++] = (byte) ((v >> 56) & 0xff);
           buffer[index++] = (byte) ((v >> 48) & 0xff);
           buffer[index++] = (byte) ((v >> 40) & 0xff);
           buffer[index++] = (byte) ((v >> 32) & 0xff);
           buffer[index++] = (byte) ((v >> 24) & 0xff);
           buffer[index++] = (byte) ((v >> 16) & 0xff);
           buffer[index++] = (byte) ((v >> 8) & 0xff);
           buffer[index++] = (byte) ((v >> 0) & 0xff);
          }else if (type == Double.TYPE){
           long v = Double.doubleToLongBits(
            fields[j].getDouble(reference));
           buffer[index++] = (byte) ((v >> 56) & 0xff);
           buffer[index++] = (byte) ((v >> 48) & 0xff);
           buffer[index++] = (byte) ((v >> 40) & 0xff);
           buffer[index++] = (byte) ((v >> 32) & 0xff);
           buffer[index++] = (byte) ((v >> 24) & 0xff);
           buffer[index++] = (byte) ((v >> 16) & 0xff);
           buffer[index++] = (byte) ((v >> 8) & 0xff);
           buffer[index++] = (byte) ((v >> 0) & 0xff);
          }else{
           throw new AssertionError(“Unsupported type.”);
          }
         }
         // Now set the length of the datagrampacket.
         datagramPacket.setLength(index);
         // Now send the packet.
         datagramSocket.send(datagramPacket);
        }
       }
      }catch (Exception e){
       throw new AssertionError(“Exception: ” + e.toString( ));
      }
     }
    }
  • A10. The tenth excerpt is the source-code of FieldReceive.java, which corresponds to steps 135 and 136 of FIG. 13, and which receives a propagated changed value and identity pair sent to it over the network 53 via FieldSend.java of annexure A9.
    import java.lang.*;
    import java.lang.reflect.*;
    import java.util.*;
    import java.net.*;
    import java.io.*;
    public class FieldReceive implements Runnable{
     /** Protocol specific values. */
     public final static int CLOSE = −1;
     public final static int NACK = 0;
     public final static int ACK = 1;
     public final static int PROPAGATE_OBJECT = 10;
     public final static int PROPAGATE_CLASS = 20;
     /** FieldAlert network values. */
     public final static String group =
      System.getProperty(“FieldAlert_network_group”);
     public final static int port =
      Integer.parseInt(System.getProperty(“FieldAlert_network_port”));
     /** Table of global ID's for local objects. (globalID-to-hashcode
      mappings) */
     public final static Hashtable globalIDToObject = new Hashtable( );
     /** Table of global ID's for local classnames. (globalID-to-classname
      mappings) */
     public final static Hashtable globalIDToClassName = new Hashtable( );
     /** Called when an application is to acquire a lock. */
     public void run( ){
      System.out.printIn(“FieldAlert_network_group=” + group);
      System.out.printIn(“FieldAlert_network_port=” + port);
      try{
       // Create a DatagramSocket to send propagated field values from
       MulticastSocket multicastSocket = new MulticastSocket(port);
       multicastSocket.joinGroup(InetAddress.getByName(group));
       // Next, create the buffer and packet for all transmissions.
       byte[ ] buffer = new byte[512]; // Working limit of 512
    // bytes per packet.
       DatagramPacket datagramPacket =
        new DatagramPacket(buffer, 0, buffer.length);
       while (!Thread.interrupted( )){
        // Make sure to reset length.
        datagramPacket.setLength(buffer.length);
        // Receive the next available packet.
        multicastSocket.receive(datagramPacket);
        int index = 0, length = datagramPacket.getLength( );
        // Decode the command.
        int command = (int) (((buffer[index++] & 0xff) << 24)
         | ((buffer[index++] & 0xff) << 16)
         | ((buffer[index++] & 0xff) << 8)
         | (buffer[index++] & 0xff));
        if (command == PROPAGATE_OBJECT){ // Propagate
    // operation
    for object fields.
         // Decode global id.
         int globalID = (int) (((buffer[index++] & 0xff) << 24)
          | ((buffer[index++] & 0xff) << 16)
          | ((buffer[index++] & 0xff) << 8)
          | (buffer[index++] & 0xff));
         // Now, need to resolve the object in question.
         Object reference = globalIDToObject.get(
          new Integer(globalID));
         // Next, get the array of fields for this object.
         Field[ ] fields = reference.getClass( ).getDeclaredFields( );
         while (index < length){
          // Decode the field id.
          int fieldID = (int) (((buffer[index++] & 0xff) << 24)
           | ((buffer[index++] & 0xff) << 16)
           | ((buffer[index++] & 0xff) << 8)
           | (buffer[index++] & 0xff));
          // Determine value length based on corresponding field
          // type.
          Field field = fields[fieldID];
          Class type = field.getType( );
          if (type == Boolean.TYPE){
           boolean v = (buffer[index++] == 1 ? true : false);
           field.setBoolean(reference, v);
          }else if (type == Byte.TYPE){
           byte v = buffer[index++];
           field.setByte(reference, v);
          }else if (type == Short.TYPE){
           short v = (short) (((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setShort(reference, v);
          }else if (type == Character.TYPE){
           char v = (char) (((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setChar(reference, v);
          }else if (type == Integer.TYPE){
           int v = (int) (((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setInt(reference, v);
          }else if (type == Float.TYPE){
           int v = (int) (((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setFloat(reference, Float.intBitsToFloat(v));
          }else if (type == Long.TYPE){
           long v = (long) (((buffer[index++] & 0xff) << 56)
            | ((buffer[index++] & 0xff) << 48)
            | ((buffer[index++] & 0xff) << 40)
            | ((buffer[index++] & 0xff) << 32)
            | ((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setLong(reference, v);
          }else if (type == Double.TYPE){
           long v = (long) (((buffer[index++] & 0xff) << 56)
            | ((buffer[index++] & 0xff) << 48)
            | ((buffer[index++] & 0xff) << 40)
            | ((buffer[index++] & 0xff) << 32)
            | ((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setDouble(reference, Double.longBitsToDouble(v));
          }else{
           throw new AssertionError(“Unsupported type.”);
          }
         }
        }else if (command == PROPAGATE_CLASS){ // Propagate
    // an update
    to class
    fields.
         // Decode the classname.
         int nameLength = (int) (((buffer[index++] & 0xff) << 24)
          | ((buffer[index++] & 0xff) << 16)
          | ((buffer[index++] & 0xff) << 8)
          | (buffer[index++] & 0xff));
         String name = new String(buffer, index, nameLength);
         index += nameLength;
         // Next, get the array of fields for this class.
         Field[ ] fields =
          FieldLoader.loadClass(name).getDeclaredFields( );
         // Decode all batched fields included in this propagation
         // packet.
         while (index < length){
          // Decode the field id.
          int fieldID = (int) (((buffer[index++] & 0xff) << 24)
           | ((buffer[index++] & 0xff) << 16)
           | ((buffer[index++] & 0xff) << 8)
           | (buffer[index++] & 0xff));
          // Determine field type to determine value length.
          Field field = fields[fieldID];
          Class type = field.getType( );
          if (type == Boolean.TYPE){
           boolean v = (buffer[index++] == 1 ? true : false);
           field.setBoolean(null, v);
          }else if (type == Byte.TYPE){
           byte v = buffer[index++];
           field.setByte(null, v);
          }else if (type == Short.TYPE){
           short v = (short) (((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setShort(null, v);
          }else if (type == Character.TYPE){
           char v = (char) (((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setChar(null, v);
          }else if (type == Integer.TYPE){
           int v = (int) (((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setInt(null, v);
          }else if (type == Float.TYPE){
           int v = (int) (((buffer[index++ ] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setFloat(null, Float.intBitsToFloat(v));
          }else if (type == Long.TYPE){
           long v = (long) (((buffer[index++] & 0xff) << 56)
            | ((buffer[index++] & 0xff) << 48)
            | ((buffer[index++] & 0xff) << 40)
            | ((buffer[index++] & 0xff) << 32)
            | ((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setLong(null, v);
          }else if (type == Double.TYPE){
           long v = (long) (((buffer[index++] & 0xff) << 56)
            | ((buffer[index++] & 0xff) << 48)
            | ((buffer[index++] & 0xff) << 40)
            | ((buffer[index++] & 0xff) << 32)
            | ((buffer[index++] & 0xff) << 24)
            | ((buffer[index++] & 0xff) << 16)
            | ((buffer[index++] & 0xff) << 8)
            | (buffer[index++] & 0xff));
           field.setDouble(null, Double.longBitsToDouble(v));
          }else{    // Unsupported field type.
           throw new AssertionError(“Unsupported type.”);
          }
         }
        }
       }
      }catch (Exception e){
       throw new AssertionError(“Exception: ” + e.toString( ));
      }
     }
    }

    A11. FieldLoader.java
  • This excerpt is the source-code of FieldLoader.java, which modifies an application program code, such as the example.java application code of Annexure A7, as it is being loaded into a JAVA virtual machine in accordance with steps 90, 91, 92, 103, and 94 of FIG. 10. FieldLoader.java makes use of the convenience classes of Annexures A12 through to A36 during the modification of a compiled JAVA classfile.
    import java.lang.*;
    import java.io.*;
    import java.net.*;
    public class FieldLoader extends URLClassLoader{
     public FieldLoader(URL[ ] urls){
      super(urls);
     }
     protected Class findClass(String name)
     throws ClassNotFoundException{
      ClassFile cf = null;
      try{
       BufferedInputStream in =
       new BufferedInputStream(findResource(
        name.replace(‘.’, ‘/’).concat(“.class”)).openStream( ));
       cf = new ClassFile(in);
      }catch (Exception e){throw new
      ClassNotFoundException(e.toString( ));}
      // Class-wide pointers to the ldc and alert index.
      int ldcindex = −1;
      int alertindex = −1;
      for (int i=0; j<cf.methods_count; i++){
       for (int j=0; j<cf.methods[i].attributes_count; j++){
        if (!(cf.methods[i].attributes[j] instanceof Code_attribute))
         continue;
        Code_attribute ca = (Code_attribute)
    cf.methods[i].attributes[j];
        boolean changed = false;
        for (int z=0; z<ca.code.length; z++){
         if ((ca.code[z][0] & 0xff) == 179){ // Opcode for a
    PUTSTATIC
    // instruction.
          changed = true;
          // The code below only supports fields in this class.
          // Thus, first off, check that this field is local to this
          // class.
          CONSTANT_Fieldref_info
          fi = (CONSTANT_Fieldref_info)
           cf.constant_pool[(int) (((ca.code[z][1] & 0xff) << 8) |
           (ca.code[z][2] & 0xff))];
          CONSTANT_Class_info ci = (CONSTANT_Class_info)
           cf.constant_pool[fi.class_index];
          String className =
           cf.constant_pool[ci.name_index].toString( );
          if (!name.equals(className)){
           throw new AssertionError(“This code only supports
    fields “
            “local to this class”);
          }
          // Ok, now search for the fields name and index.
          int index = 0;
          CONSTANT_NameAndType_info
          ni = (CONSTANT_NameAndType_info)
           cf.constant_pool[fi.name_and_type_index];
          String fieldName =
           cf.constant_pool[ni.name_index].toString( );
          for (int a=0; a<cf.fields_count; a++){
           String fn = cf.constant_pool[
            cf.fields[a].name_index].toString( );
           if (fieldName.equals(fn)){
            index = a;
            break;
           }
          }
          // Next, realign the code array, making room for the
          // insertions.
          byte[ ][ ] code2 = new byte[ca.code.length+3][ ];
          System.arraycopy(ca.code, 0, code2, 0, z+1);
          System.arraycopy(ca.code, z+1, code2, z+4,
           ca.code.length−(z+1));
          ca.code = code2;
          // Next, insert the LDC_W instruction.
          if (ldcindex == −1){
           CONSTANT_String_info csi =
            new CONSTANT_String_info(ci.name_index);
           cp_info[ ] cpi = new
           cp_info[cf.constant_pool.length+1];
           System.arraycopy(cf.constant_pool, 0, cpi, 0,
            cf.constant_pool.length);
           cpi[cpi.length − 1] = csi;
           ldcindex = cpi.length−1;
           cf.constant_pool = cpi;
           cf.constant_pool_count++;
          }
          ca.code[z+1] = new byte[3];
          ca.code[z+1][0] = (byte) 19;
          ca.code[z+1][1] = (byte) ((ldcindex >> 8) & 0xff);
          ca.code[z+1][2] = (byte) (ldcindex & 0xff);
          // Next, insert the SIPUSH instruction.
          ca.code[z+2] = new byte[3];
          ca.code[z+2][0] = (byte) 17;
          ca.code[z+2][1] = (byte) ((index >> 8) & 0xff);
          ca.code[z+2][2] = (byte) (index & 0xff);
          // Finally, insert the INVOKESTATIC instruction.
          if (alertindex == −1){
           // This is the first time this class is encourtering
    the
           // alert instruction, so have to add it to the constant
           // pool.
           cp_info[ ] cpi = new
           cp_info[cf.constant_pool.length+6];
           System.arraycopy(cf.constant_pool, 0, cpi, 0,
            cf.constant_pool.length);
           cf.constant_pool = cpi;
           cf.constant_pool_count += 6;
           CONSTANT_Utf8_info u1 =
            new CONSTANT_Utf8_info(“FieldAlert”);
           cf.constant_pool[cf.constant_pool.length−6] = u1;
           CONSTANT_Class_info c1 = new
           CONSTANT_Class_info(
            cf.constant_pool_count−6);
           cf.constant_pool[cf.constant_pool.length−5] = c1;
           u1 = new CONSTANT_Utf8_info(“alert”);
           cf.constant_pool[cf.constant_pool.length−4] = u1;
           u1 = new
           CONSTANT_Utf8_info(“(Ljava/lang/Object;I)V”);
           cf.constant_pool[cf.constant_pool.length−3] = u1;
           CONSTANT_NameAndType_info n1 =
            new CONSTANT_NameAndType_info(
            cf.constant_pool.length−4, cf.constant_pool.length−
    3);
           cf.constant_pool[cf.constant_pool.length−2] = n1;
           CONSTANT_Methodref_info m1 = new
    CONSTANT_Methodref_info(
            cf.constant_pool.length−5, cf.constant_pool.length−
    2);
           cf.constant_pool[cf.constant_pool.length−1] = m1;
           alertindex = cf.constant_pool.length−1;
          }
          ca.code[z+3] = new byte[3];
          ca.code[z+3][0] = (byte) 184;
          ca.code[z+3][1] = (byte) ((alertindex >> 8) & 0xff);
          ca.code[z+3][2] = (byte) (alertindex & 0xff);
          // And lastly, increase the CODE_LENGTH and
    ATTRIBUTE_LENGTH
          // values.
          ca.code_length += 9;
          ca.attribute_length += 9;
         }
        }
        // If we changed this method, then increase the stack size by
    one.
        if (changed){
         ca.max_stack++;    // Just to make sure.
        }
       }
      }
      try{
       ByteArrayOutputStream out = new ByteArrayOutputStream( );
       cf.serialize(out);
       byte[ ] b = out.toByteArray( );
       return defineClass(name, b, 0, b.length);
      }catch (Exception e){
       throw new ClassNotFoundException(name);
      }
     }
    }

    A12. Attribute_info.java
  • Convience class for representing attribute_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** This abstract class represents all types of attribute_info
     * that are used in the JVM specifications.
     *
     * All new attribute_info subclasses are to always inherit from this
     * class.
     */
    public abstract class attribute_info{
      public int attribute_name_index;
      public int attribute_length;
      /** This is used by subclasses to register themselves
       * to their parent classFile.
       */
      attribute_info(ClassFile cf){ }
      /** Used during input serialization by ClassFile only. */
      attribute_info(ClassFile cf, DataInputStream in)
        throws IOException{
        attribute_name_index = in.readChar( );
        attribute_length = in.readInt( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        out.writeChar(attribute_name_index);
        out.writeInt(attribute_length);
      }
      /** This class represents an unknown attribute_info that
       * this current version of classfile specification does
       * not understand.
       */
      public final static class Unknown extends attribute_info{
        byte[ ] info;
        /** Used during input serialization by ClassFile only. */
        Unknown(ClassFile cf, DataInputStream in)
          throws IOException{
          super(cf, in);
          info = new byte[attribute_length];
          in.read(info, 0, attribute_length);
        }
        /** Used during output serialization by ClassFile only. */
        void serialize(DataOutputStream out)
          throws IOException{
          ByteArrayOutputStream baos =
          new ByteArrayOutputStream( );
          super.serialize(out);
          out.write(info, 0, attribute_length);
        }
      }
    }

    A13. ClassFile.java
  • Convience class for representing ClassFile structures.
    import java.lang.*;
    import java.io.*;
    import java.util.*;
    /** The ClassFile follows verbatim from the JVM specification. */
    public final class ClassFile {
      public int magic;
      public int minor_version;
      public int major_version;
      public int constant_pool_count;
      public cp_info[ ] constant_pool;
      public int access_flags;
      public int this_class;
      public int super_class;
      public int interfaces_count;
      public int[ ] interfaces;
      public int fields_count;
      public field_info[ ] fields;
      public int methods_count;
      public method_info[ ] methods;
      public int attributes_count;
      public attribute_info[ ] attributes;
      /** Constructor. Takes in a byte stream representation and transforms
       * each of the attributes in the ClassFile into objects to allow for
       * easier manipulation.
       */
      public ClassFile(InputStream ins)
        throws IOException{
        DataInputStream in = (ins instanceof DataInputStream ?
          (DataInputStream) ins : new DataInputStream(ins));
        magic = in.readInt( );
        minor_version = in.readChar( );
        major_version = in.readChar( );
        constant_pool_count = in.readChar( );
        constant_pool = new cp_info[constant_pool_count];
        for (int i=1; i<constant_pool_count; i++){
          in.mark(1);
          int s = in.read( );
          in.reset( );
          switch (s){
            case 1:
              constant_pool[i] = new CONSTANT_Utf8_info(this, in);
              break;
            case 3:
              constant_pool[i] = new CONSTANT_Integer_info(this, in);
              break;
            case 4:
              constant_pool[i] = new CONSTANT_Float_info(this, in);
              break;
            case 5:
              constant_pool[i] = new CONSTANT_Long_info(this, in);
              i++;
              break;
            case 6:
              constant_pool[i] = new CONSTANT_Double_info(this, in);
              i++;
              break;
            case 7:
              constant_pool[i] = new CONSTANT_Class_info(this, in);
              break;
            case 8:
              constant_pool[i] = new CONSTANT_String_info(this, in);
              break;
            case 9:
              constant_pool[i] = new CONSTANT_Fieldref_info(this, in);
              break;
            case 10:
              constant_pool[i] = new CONSTANT_Methodref_info(this,
    in);
              break;
            case 11:
              constant_pool[i] =
                new CONSTANT_InterfaceMethodref_info(this, in);
              break;
            case 12:
              constant_pool[i] = new
    CONSTANT_NameAndType_info(this, in);
              break;
            default:
              throw new ClassFormatError(“Invalid ConstantPoolTag”);
          }
        }
        access_flags = in.readChar( );
        this_class = in.readChar( );
        super_class = in.readChar( );
        interfaces_count = in.readChar( );
        interfaces = new int[interfaces_count];
        for (int i=0; i<interfaces_count; i++)
          interfaces[i] = in.readChar( );
        fields_count = in.readChar( );
        fields = new field_info[fields_count];
        for (int i=0; i<fields_count; i++) {
          fields[i] = new field_info(this, in);
        }
        methods_count = in.readChar( );
        methods = new method_info[methods_count];
        for (int i=0; i<methods_count; i++) {
          methods[i] = new method_info(this, in);
        }
        attributes_count = in.readChar( );
        attributes = new attribute_info[attributes_count];
        for (int i=0; i<attributes_count; i++){
          in.mark(2);
          String s = constant_pool[in.readChar( )].toString( );
          in.reset( );
          if (s.equals(“SourceFile”))
            attributes[i] = new SourceFile_attribute(this, in);
          else if (s.equals(“Deprecated”))
            attributes[i] = new Deprecated_attribute(this, in);
          else if (s.equals(“InnerClasses”))
            attributes[i] = new InnerClasses_attribute(this, in);
          else
            attributes[i] = new attribute_info.Unknown(this, in);
        }
      }
      /** Serializes the ClassFile object into a byte stream. */
      public void serialize(OutputStream o)
        throws IOException{
        DataOutputStream out = (o instanceof DataOutputStream ?
          (DataOutputStream) o : new DataOutputStream(o));
        out.writeInt(magic);
        out.writeChar(minor_version);
        out.writeChar(major_version);
        out.writeChar(constant_pool_count);
        for (int i=1; i<constant_pool_count; i++){
          constant_pool[i].serialize(out);
          if (constant_pool[i] instanceof CONSTANT_Long_info ||
              constant_pool[i] instanceof CONSTANT_Double_info)
            i++;
        }
        out.writeChar(access_flags);
        out.writeChar(this_class);
        out.writeChar(super_class);
        out.writeChar(interfaces_count);
        for (int i=0; i<interfaces_count; i++)
          out.writeChar(interfaces[i]);
        out.writeChar(fields_count);
        for (int i=0; i<fields_count; i++)
          fields[i].serialize(out);
        out.writeChar(methods_count);
        for (int i=0; i<methods_count; i++)
          methods[i].serialize(out);
        out.writeChar(attributes_count);
        for (int i=0; i<attributes_count; i++)
          attributes[i].serialize(out);
        // Flush the outputstream just to make sure.
        out.flush( );
      }
    }

    A14. Code_attribute.java
  • Convience class for representing Code_attribute structures within ClassFiles.
    import java.util.*;
    import java.lang.*;
    import java.io.*;
    /**
     * The code[ ] is stored as a 2D array. */
    public final class Code_attribute extends attribute_info{
      public int max_stack;
      public int max_locals;
      public int code_length;
      public byte[ ][ ] code;
      public int exception_table_length;
      public exception_table[ ] exception_table;
      public int attributes_count;
      public attribute_info[ ] attributes;
      /** Internal class that handles the exception table. */
      public final static class exception_table{
        public int start_pc;
        public int end_pc;
        public int handler_pc;
        public int catch_type;
      }
      /** Constructor called only by method_info. */
      Code_attribute(ClassFile cf, int ani, int al, int ms, int ml, int cl,
          byte[ ][ ] cd, int etl, exception_table[ ] et, int ac,
          attribute_info[ ] a){
        super(cf);
        attribute_name_index = ani;
        attribute_length = al;
        max_stack = ms;
        max_locals = ml;
        code_length = cl;
        code = cd;
        exception_table_length = etl;
        exception_table = et;
        attributes_count = ac;
        attributes = a;
      }
      /** Used during input serialization by ClassFile only. */
      Code_attribute(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        max_stack = in.readChar( );
        max_locals = in.readChar( );
        code_length = in.readInt( );
        code = new byte[code_length][ ];
        int i = 0;
      for (int pos=0; pos<code_length; i++){
        in.mark(1);
        int s = in.read( );
        in.reset( );
        switch (s){
          case 16:
          case 18:
          case 21:
          case 22:
          case 23:
          case 24:
          case 25:
          case 54:
          case 55:
          case 56:
          case 57:
          case 58:
          case 169:
          case 188:
          case 196:
            code[i] = new byte[2];
            break;
          case 17:
          case 19:
          case 20:
          case 132:
          case 153:
          case 154:
          case 155:
          case 156:
          case 157:
          case 158:
          case 159:
          case 160:
          case 161:
          case 162:
          case 163:
          case 164:
          case 165:
          case 166:
          case 167:
          case 168:
          case 178:
          case 179:
          case 180:
          case 181:
          case 182:
          case 183:
          case 184:
          case 187:
          case 189:
          case 192:
          case 193:
          case 198:
          case 199:
          case 209:
            code[i] = new byte[3];
            break;
          case 197:
            code[i] = new byte[4];
            break;
          case 185:
          case 200:
          case 201:
            code[i] = new byte[5];
            break;
          case 170:{
            int pad = 3 − (pos % 4);
            in.mark(pad+13); // highbyte
            in.skipBytes(pad+5); // lowbyte
            int low = in.readInt( );
            code[i] =
              new byte[pad + 13 + ((in.readInt( ) − low +
              1) * 4)];
            in.reset( );
            break;
          }case 171:{
            int pad = 3 − (pos % 4);
            in.mark(pad+9);
            in.skipBytes(pad+5);
            code[i] = new byte[pad + 9 + (in.readInt( ) * 8)];
            in.reset( );
            break;
          }default:
            code[i] = new byte[1];
        }
        in.read(code[i], 0, code[i].length);
        pos += code[i].length;
      }
      // adjust the array to the new size and store the size
      byte[ ][ ] temp = new byte[i][ ];
      System.arraycopy(code, 0, temp, 0, i);
      code = temp;
      exception_table_length = in.readChar( );
      exception_table =
        new Code_attribute.exception_table[exception_table_length];
      for (i=0; i<exception_table_length; i++){
        exception_table[i] = new exception_table( );
        exception_table[i].start_pc = in.readChar( );
        exception_table[i].end_pc = in.readChar( );
        exception_table[i].handler_pc = in.readChar( );
        exception_table[i].catch_type = in.readChar( );
      }
      attributes_count = in.readChar( );
      attributes = new attribute_info[attributes_count];
      for (i=0; i<attributes_count; i++){
        in.mark(2);
        String s = cf.constant_pool[in.readChar( )].toString( );
        in.reset( );
        if (s.equals(“LineNumberTable”))
          attributes[i] = new LineNumberTable_attribute(cf, in);
        else if (s.equals(“LocalVariableTable”))
          attributes[i] = new LocalVariableTable_attribute(cf, in);
        else
          attributes[i] = new attribute_info.Unknown(cf, in);
      }
    }
    /** Used during output serialization by ClassFile only.
    */
    void serialize(DataOutputStream out)
      throws IOException{
        attribute_length = 12 + code_length +
          (exception_table_length * 8);
        for (int i=0; i<attributes_count; i++)
          attribute_length += attributes[i].attribute_length + 6;
        super.serialize(out);
        out.writeChar(max_stack);
        out.writeChar(max_locals);
        out.writeInt(code_length);
        for (int i=0, pos=0; pos<code_length; i++){
          out.write(code[i], 0, code[i].length);
          pos += code[i].length;
        }
        out.writeChar(exception_table_length);
        for (int i=0; i<exception_table_length; i++){
          out.writeChar(exception_table[i].start_pc);
          out.writeChar(exception_table[i].end_pc);
          out.writeChar(exception_table[i].handler_pc);
          out.writeChar(exception_table[i].catch_type);
        }
        out.writeChar(attributes_count);
        for (int i=0; i<attributes_count; i++)
          attributes[i].serialize(out);
      }
    }

    A15. CONSTANT_Class_info.java
  • Convience class for representing CONSTANT_Class_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Class subtype of a constant pool entry. */
    public final class CONSTANT_Class_info extends cp_info{
      /** The index to the name of this class. */
      public int name_index = 0;
      /** Convenience constructor.
       */
      public CONSTANT_Class_info(int index) {
        tag = 7;
        name_index = index;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_Class_info(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 7)
          throw new ClassFormatError( );
        name_index = in.readChar( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(tag);
        out.writeChar(name_index);
      }
    }

    A16. CONSTANT_Double_info.java
  • Convience class for representing CONSTANT_Double_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Double subtype of a constant pool entry. */
    public final class CONSTANT_Double_info extends cp_info{
      /** The actual value. */
      public double bytes;
      public CONSTANT_Double_info(double d){
        tag = 6;
        bytes = d;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_Double_info(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 6)
          throw new ClassFormatError( );
        bytes = in.readDouble( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(tag);
        out.writeDouble(bytes);
        long l = Double.doubleToLongBits(bytes);
      }
    }

    A17. CONSTANT_Fieldref info.java
  • Convience class for representing CONSTANT_Fieldref_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Fieldref subtype of a constant pool entry. */
    public final class CONSTANT_Fieldref_info extends cp_info{
      /** The index to the class that this field is referencing to. */
      public int class_index;
      /** The name and type index this field if referencing to. */
      public int name_and_type_index;
      /** Convenience constructor. */
      public CONSTANT_Fieldref_info(int class_index,
      int name_and_type_index)
    {
        tag = 9;
        this.class_index = class_index;
        this.name_and_type_index = name_and_type_index;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_Fieldref_info(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 9)
          throw new ClassFormatError( );
        class_index = in.readChar( );
        name_and_type_index = in.readChar( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(tag);
        out.writeChar(class_index);
        out.writeChar(name_and_type_index);
      }
    }

    A18. CONSTANT_Float_info.java
  • Convience class for representing CONSTANT_Float_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Float subtype of a constant pool entry. */
    public final class CONSTANT_Float_info extends cp_info{
      /** The actual value. */
      public float bytes;
      public CONSTANT_Float_info(float f){
        tag = 4;
        bytes = f;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_Float_info(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 4)
          throw new ClassFormatError( );
        bytes = in.readFloat( );
      }
      /** Used during output serialization by ClassFile only. */
      public void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(4);
        out.writeFloat(bytes);
      }
    }

    A19. CONSTANT_Integer info.java
  • Convience class for representing CONSTANT_Integer_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Integer subtype of a constant pool entry. */
    public final class CONSTANT_Integer_info extends cp_info{
      /** The actual value. */
      public int bytes;
      public CONSTANT_Integer_info(int b) {
        tag = 3;
        bytes = b;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_Integer_info(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 3)
          throw new ClassFormatError( );
        bytes = in.readInt( );
      }
      /** Used during output serialization by ClassFile only. */
      public void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(tag);
        out.writeInt(bytes);
      }
    }

    A20. CONSTANT_InterfaceMethodref_info.java
  • Convience class for representing CONSTANT_InterfaceMethodref_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** InterfaceMethodref subtype of a constant pool entry.
     */
    public final class CONSTANT_InterfaceMethodref_info extends
    cp_info{
      /** The index to the class that this field is referencing to. */
      public int class_index;
      /** The name and type index this field if referencing to. */
      public int name_and_type_index;
      public CONSTANT_InterfaceMethodref_info(int class_index,
                  int name_and_type_index) {
        tag = 11;
        this.class_index = class_index;
        this.name_and_type_index = name_and_type_index;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_InterfaceMethodref_info(ClassFile cf,
      DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 11)
          throw new ClassFormatError( );
        class_index = in.readChar( );
        name_and_type_index = in.readChar( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(tag);
        out.writeChar(class_index);
        out.writeChar(name_and_type_index);
      }
    }

    A21. CONSTANT_Long_info.java
  • Convience class for representing CONSTANT_Long_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Long subtype of a constant pool entry. */
    public final class CONSTANT_Long_info extends cp_info{
      /** The actual value. */
      public long bytes;
      public CONSTANT_Long_info(long b){
        tag = 5;
        bytes = b;
      }
      /** Used during input serialization by ClassFile only. */
      CONSTANT_Long_info(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        if (tag != 5)
          throw new ClassFormatError( );
        bytes = in.readLong( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        out.writeByte(tag);
        out.writeLong(bytes);
      }
    }

    A22. CONSTANT_Methodref_info.java
  • Convience class for representing CONSTANT_Methodref_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Methodref subtype of a constant pool entry.
     */
    public final class CONSTANT_Methodref_info extends cp_info{
     /** The index to the class that this field is referencing to. */
     public int class_index;
     /** The name and type index this field if referencing to. */
     public int name_and_type_index;
     public CONSTANT_Methodref_info(int class_index,
     int name_and_type_index)
    {
      tag = 10;
      this.class_index = class_index;
      this.name_and_type_index = name_and_type_index;
     }
     /** Used during input serialization by ClassFile only. */
     CONSTANT_Methodref_info(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      if (tag != 10)
       throw new ClassFormatError( );
      class_index = in.readChar( );
      name_and_type_index = in.readChar( );
     }
     /** Used during output serialization by ClassFile only. */
     void serialize (DataOutputStream out)
      throws IOException{
      out.writeByte(tag);
      out.writeChar(class_index);
      out.writeChar(name_and_type_index);
     }
    }

    A23. CONSTANT_NameAndType_info.java
  • Convience class for representing CONSTANT_NameAndType_info structures within ClassFiles.
    import java.io.*;
    import java.lang.*;
    /** NameAndType subtype of a constant pool entry.
     */
    public final class CONSTANT_NameAndType_info extends cp_info{
     /** The index to the Utf8 that contains the name. */
     public int name_index;
     /** The index fo the Utf8 that constains the signature. */
     public int descriptor_index;
     public CONSTANT_NameAndType_info(int name_index,
     int descriptor_index) {
      tag = 12;
      this.name_index = name_index;
      this.descriptor_index = descriptor_index;
     }
     /** Used during input serialization by ClassFile only. */
     CONSTANT_NameAndType_info(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      if (tag != 12)
       throw new ClassFormatError( );
      name_index = in.readChar( );
      descriptor_index = in.readChar( );
     }
     /** Used during output serialization by ClassFile only. */
     void serialize(DataOutputStream out)
      throws IOException{
      out.writeByte(tag);
      out.writeChar(name_index);
      out.writeChar(descriptor_index);
     }
    }

    A24. CONSTANT_String_info.java
  • Convience class for representing CONSTANT_String_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** String subtype of a constant pool entry.
     */
    public final class CONSTANT_String_info extends cp_info{
     /** The index to the actual value of the string. */
     public int string_index;
     public CONSTANT_String_info(int value) {
      tag = 8;
      string_index = value;
     }
     /** ONLY TO BE USED BY CLASSFILE! */
     public CONSTANT_String_info(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      if (tag != 8)
       throw new ClassFormatError( );
      string_index = in.readChar( );
     }
     /** Output serialization, ONLY TO BE USED BY CLASSFILE! */
     public void serialize(DataOutputStream out)
      throws IOException{
      out.writeByte(tag);
      out.writeChar(string_index);
     }
    }

    A25. CONSTANT_Utf8_info.java
  • Convience class for representing CONSTANT_Utf8_info structures within ClassFiles.
    import java.io.*;
    import java.lang.*;
    /** Utf8 subtype of a constant pool entry.
     *  We internally represent the Utf8 info byte array
     *  as a String.
     */
    public final class CONSTANT_Utf8_info extends cp_info{
     /** Length of the byte array. */
     public int length;
     /** The actual bytes, represented by a String. */
     public String bytes;
     /** This constructor should be used for the purpose
      *  of part creation. It does not set the parent
      *  ClassFile reference.
      */
     public CONSTANT_Utf8_info(String s) {
      tag = 1;
      length = s.length( );
      bytes = s;
     }
     /** Used during input serialization by ClassFile only. */
     public CONSTANT_Utf8_info(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      if (tag != 1)
       throw new ClassFormatError( );
      length = in.readChar( );
      byte[ ] b = new byte[length];
      in.read(b, 0, length);
      // WARNING: String constructor is deprecated.
      bytes = new String(b, 0, length);
     }
     /** Used during output serialization by ClassFile only. */
     public void serialize(DataOutputStream out)
      throws IOException{
      out.writeByte(tag);
      out.writeChar(length);
      // WARNING: Handling of String coversion here might be
      problematic.out.writeBytes(bytes);
     }
     public String toString( ){
      return bytes;
     }
    }

    A26. ConstantValue_attribute.java
  • Convience class for representing ConstantValue_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Attribute that allows for initialization of static variables in
     *  classes. This attribute will only reside in a field_info struct.
     */
    public final class ConstantValue_attribute extends attribute_info{
     public int constantvalue_index;
     public ConstantValue_attribute(ClassFile cf, int ani, int al, int cvi){
      super(cf);
      attribute_name_index = ani;
      attribute_length = al;
      constantvalue_index = cvi;
     }
     public ConstantValue_attribute(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      constantvalue_index = in.readChar( );
     }
     public void serialize(DataOutputStream out)
      throws IOException{
      attribute_length = 2;
      super.serialize(out);
      out.writeChar(constantvalue_index);
     }
    }

    A27. cp_info.java
  • Convience class for representing cp_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Represents the common interface of all constant pool parts
     *  that all specific constant pool items must inherit from.
     *
     */
    public abstract class cp_info{
     /** The type tag that signifies what kind of constant pool
      *  item it is */
     public int tag;
     /** Used for serialization of the object back into a bytestream. */
     abstract void serialize(DataOutputStream out) throws IOException;
     /** Default constructor. Simply does nothing. */
     public cp_info( ) { }
     /** Constructor simply takes in the ClassFile as a reference to
      *  it's parent
      */
     public cp_info(ClassFile cf) { }
     /** Used during input serialization by ClassFile only. */
     cp_info(ClassFile cf, DataInputStream in)
      throws IOException{
      tag = in.readUnsignedByte( );
     }
    }

    A28. Deprecated_attribute.java
  • Convience class for representing Deprecated_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** A fix attributed that can be located either in the ClassFile,
     *  field_info or the method_info attribute. Mark deprecated to
     *  indicate that the method, class or field has been superceded.
     */
    public final class Deprecated_attribute extends attribute_info{
     public Deprecated_attribute(ClassFile cf, int ani, int al){
      super(cf);
      attribute_name_index = ani;
      attribute_length = al;
     }
     /** Used during input serialization by ClassFile only. */
     Deprecated_attribute(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      }
     }

    A29. Exceptions_attribute.java
  • Convience class for representing Exceptions_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** This is the struct where the exceptions table are located.
     *  <br><br>
     *  This attribute can only appear once in a method_info struct.
     */
    public final class Exceptions_attribute extends attribute_info{
     public int number_of_exceptions;
     public int[ ] exception_index_table;
     public Exceptions_attribute(ClassFile cf, int ani, int al, int noe,
               int[ ] eit){
      super(cf);
      attribute_name_index = ani;
      attribute_length = al;
      number_of_exceptions = noe;
      exception_index_table = eit;
     }
     /** Used during input serialization by ClassFile only. */
     Exceptions_attribute(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      number_of_exceptions = in.readChar( );
      exception_index_table = new int [number_of_exceptions];
      for (int i=0; i<number_of_exceptions; i++)
       exception_index_table[i] = in.readChar( );
     }
     /** Used during output serialization by ClassFile only. */
     public void serialize(DataOutputStream out)
      throws IOException{
      attribute_length = 2 + (number_of_exceptions*2);
      super.serialize(out);
      out.writeChar(number_of_exceptions);
      for (int i=0; i<number_of_exceptions; i++)
       out.writeChar(exception_index_table[i]);
     }
    }

    A30. field_info.java
  • Convience class for representing field_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /**  Represents the field_info structure as specified in the JVM
    specification.
     */
    public final class field_info{
     public int access_flags;
     public int name_index;
     public int descriptor_index;
     public int attributes_count;
     public attribute_info[ ] attributes;
     /** Convenience constructor. */
     public field_info(ClassFile cf, int flags, int ni, int di){
      access_flags = flags;
      name_index = ni;
      descriptor_index = di;
      attributes_count = 0;
      attributes = new attribute_info[0];
     }
     /** Constructor called only during the serialization process.
      *  <br><br>
      *  This is intentionally left as package protected as we
      *  should not normally call this constructor directly.
      *  <br><br>
      *  Warning: the handling of len is not correct (after String s =...)
      */
     field_info(ClassFile cf, DataInputStream in)
      throws IOException{
      access_flags = in.readChar( ) ;
      name_index = in.readChar( ) ;
      descriptor_index = in.readChar( ) ;
      attributes_count = in.readChar( ) ;
      attributes = new attribute_info[attributes_count];
      for (int i=0; i<attributes_count; i++){
       in.mark(2);
       String s = cf.constant_pool[in.readChar( ) ].toString( ) ;
       in.reset( ) ;
       if (s.equals(“ConstantValue”))
        attributes[i] = new ConstantValue_attribute(cf, in);
       else if (s.equals(“Synthetic”))
        attributes[i] = new Synthetic_attribute(cf, in);
       else if (s.equals(“Deprecated”))
        attributes[i] = new Deprecated_attribute(cf, in);
       else
        attributes[i] = new attribute_info.Unknown(cf, in);
      }
     }
     /** To serialize the contents into the output format.
      */
     public void serialize(DataOutputStream out)
      throws IOException{
      out.writeChar(access_flags);
      out.writeChar(name_index);
      out.writeChar(descriptor_index);
      out.writeChar(attributes_count);
      for (int i=0; i<attributes_count; i++)
       attributes[i].serialize(out);
     }
    }

    A31. InnerClasses_attribute.java
  • Convience class for representing InnerClasses_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** A variable length structure that contains information about an
     *  inner class of this class.
     */
    public final class InnerClasses_attribute extends attribute_info{
     public int number_of_classes;
     public classes[ ] classes;
     public final static class classes{
      int inner_class_info_index;
      int outer_class_info_index;
      int inner_name_index;
      int inner_class_access_flags;
     }
     public InnerClasses_attribute(ClassFile cf, int ani, int al,
               int noc, classes[ ] c){
      super(cf);
      attribute_name_index = ani;
      attribute_length = al;
      number_of_classes = noc;
      classes = c;
     }
     /** Used during input serialization by ClassFile only. */
     InnerClasses_attribute(ClassFile cf, DataInputStream in)
      throws IOException{
      super(cf, in);
      number_of_classes = in.readChar( ) ;
      classes = new InnerClasses_attribute.classes[number_of_classes];
      for (int i=0; i<number_of_classes; i++){
       classes[i] = new classes( ) ;
       classes[i].inner_class_info_index = in.readChar( ) ;
       classes[i].outer_class_info_index = in.readChar( ) ;
       classes[i].inner_name_index = in.readChar( ) ;
       classes[i].inner_class_access_flags = in.readChar( ) ;
      }
     }
     /** Used during output serialization by ClassFile only. */
     public void serialize(DataOutputStream out)
      throws IOException{
      attribute_length = 2 + (number_of_classes * 8);
      super.serialize(out);
      out.writeChar(number_of_classes);
      for (int i=0; i<number_of_classes; i++){
       out.writeChar(classes[i].inner_class_info_index);
       out.writeChar(classes[i].outer_class_info_index);
       out.writeChar(classes[i].inner_name_index);
       out.writeChar(classes[i].inner_class_access_flags);
      }
     }
    }

    A32. LineNumberTable_attribute.java
  • Convience class for representing LineNumberTable_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Determines which line of the binary code relates to the
     * corresponding source code.
     */
    public final class LineNumberTable_attribute extends attribute_info{
      public int line_number_table_length;
      public line_number_table[ ] line_number_table;
      public final static class line_number_table{
        int start_pc;
        int line_number;
      }
      public LineNumberTable_attribute(ClassFile cf, int ani, int al, int
    lntl,
                  line_number_table[ ] lnt){
        super(cf);
        attribute_name_index = ani;
        attribute_length = al;
        line_number_table_length = lntl;
        line_number_table = lnt;
      }
      /** Used during input serialization by ClassFile only. */
      LineNumberTable_attribute(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        line_number_table_length = in.readChar( );
        line_number_table = new
    LineNumberTable_attribute.line_number_table-
    [line_number_table_length];
        for (int i=0; i<line_number_table_length; i++){
          line_number_table[i] = new line_number_table( );
          line_number_table[i].start_pc = in.readChar( );
          line_number_table[i].line_number = in.readChar( );
        }
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        attribute_length = 2 + (line_number_table_length * 4);
        super.serialize(out);
        out.writeChar(line_number_table_length);
        for (int i=0; i<line_number_table_length; i++){
          out.writeChar(line_number_table[i].start_pc);
          out.writeChar(line_number_table[i].line_number);
        }
      }
    }

    A33. LocalVariableTable_attribute.java
  • Convience class for representing LocalVariableTable_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** Used by debugger to find out how the source file line number is linked
     * to the binary code. It has many to one correspondence and is found in
     * the Code_attribute.
     */
    public final class LocalVariableTable_attribute extends attribute_info{
      public int local_variable_table_length;
      public local_variable_table[ ] local_variable_table;
      public final static class local_variable_table{
        int start_pc;
        int length;
        int name_index;
        int descriptor_index;
        int index;
      }
      public LocalVariableTable_attribute(ClassFile cf, int ani, int al,
                  int lvtl, local_variable_table[ ]
    lvt){
        super(cf);
        attribute_name_index = ani;
        attribute_length = al;
        local_variable_table_length = lvtl;
        local_variable_table = lvt;
      }
      /** Used during input serialization by ClassFile only. */
      LocalVariableTable_attribute(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        local_variable_table_length = in.readChar( );
        local_variable_table = new
    LocalVariableTable_attribute.local_variable_table-
    [local_variable_table_length];
        for (int i=0; i<local_variable_table_length; i++){
          local_variable_table[i] = new local_variable_table( );
          local_variable_table[i].start_pc = in.readChar( );
          local_variable_table[i].length = in.readChar( );
          local_variable_table[i].name_index = in.readChar( );
          local_variable_table[i].descriptor_index = in.readChar( );
          local_variable_table[i].index = in.readChar( );
        }
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream out)
        throws IOException{
        attribute_length = 2 + (local_variable_table_length * 10);
        super.serialize(out);
        out.writeChar(local_variable_table_length);
        for (int i=0; i<local_variable_table_length; i++){
          out.writeChar(local_variable_table[i].start_pc);
          out.writeChar(local_variable_table[i].length);
          out.writeChar(local_variable_table[i].name_index);
          out.writeChar(local_variable_table[i].descriptor_index);
          out.writeChar(local_variable_table[i].index);
        }
      }
    }

    A34. method_info.java
  • Convience class for representing method_info structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** This follows the method_info in the JVM specification.
     */
    public final class method_info {
      public int access_flags;
      public int name_index;
      public int descriptor_index;
      public int attributes_count;
      public attribute_info[ ] attributes;
      /** Constructor. Creates a method_info, initializes it with
       * the flags set, and the name and descriptor indexes given.
       * A new uninitialized code attribute is also created, and stored
       * in the <i>code</i> variable.*/
      public method_info(ClassFile cf, int flags, int ni, int di,
                int ac, attribute_info[ ] a) {
        access_flags = flags;
        name_index = ni;
        descriptor_index = di;
        attributes_count = ac;
      attributes = a;
      }
      /** This method creates a method_info from the current pointer in
       * the data stream. Only called by during the serialization of a
       * complete ClassFile from a bytestream, not normally
       invoked directly.
       */
      method_info(ClassFile cf, DataInputStream in)
        throws IOException{
        access_flags = in.readChar( );
        name_index = in.readChar( );
        descriptor_index = in.readChar( );
        attributes_count = in.readChar( );
        attributes = new attribute_info[attributes_count];
        for (int i=0; i<attributes_count; i++){
          in.mark(2);
          String s = cf.constant_pool[in.readChar( )].toString( );
          in.reset( );
          if (s.equals(“Code”))
            attributes[i] = new Code_attribute(cf, in);
          else if (s.equals(“Exceptions”))
            attributes[i] = new Exceptions_attribute(cf, in);
          else if (s.equals(“Synthetic”))
            attributes[i] = new Synthetic_attribute(cf, in);
          else if (s.equals(“Deprecated”))
            attributes[i] = new Deprecated_attribute(cf, in);
          else
            attributes[i] = new attribute_info.Unknown(cf, in);
        }
      }
      /** Output serialization of the method_info to a byte array.
       * Not normally invoked directly.
      */
      public void serialize(DataOutputStream Out)
        throws IOException{
        out.writeChar(access_flags);
        out.writeChar(name_index);
        out.writeChar(descriptor_index);
        out.writeChar(attributes_count);
        for (int i=0; i<attributes_count; i++)
          attributes[i].serialize(out);
      }
    }

    A35. SourceFile_attribute.java
  • Convience class for representing SourceFile_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** A SourceFile attribute is an optional fixed_length attribute in
     * the attributes table. Only located in the ClassFile struct only
     * once.
     */
    public final class SourceFile_attribute extends attribute_info{
      public int sourcefile_index;
      public SourceFile_attribute(ClassFile cf, int ani, int al, int sfi){
        super(cf);
        attribute_name_index = ani;
        attribute_length = al;
        sourcefile_index = sfi;
      }
      /** Used during input serialization by ClassFile only. */
      SourceFile_attribute(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
        sourcefile_index = in.readChar( );
      }
      /** Used during output serialization by ClassFile only. */
      void serialize(DataOutputStream Out)
        throws IOException{
        attribute_length = 2;
        super.serialize(out);
        out.writeChar(sourcefile_index);
      }
    }

    A36. Synthetic_attribute.java
  • Convience class for representing Synthetic_attribute structures within ClassFiles.
    import java.lang.*;
    import java.io.*;
    /** A synthetic attribute indicates that this class does not have
     * a generated code source. It is likely to imply that the code
     * is generated by machine means rather than coded directly. This
     * attribute can appear in the classfile, method_info or field_info.
     * It is fixed length.
     */
    public final class Synthetic_attribute extends attribute_info{
      public Synthetic_attribute(ClassFile cf, int ani, int al){
        super(cf);
        attribute_name_index = ani;
        attribute_length = al;
      }
      /** Used during output serialization by ClassFile only. */
      Synthetic_attribute(ClassFile cf, DataInputStream in)
        throws IOException{
        super(cf, in);
      }
    }

    End of Annexure

Claims (39)

1. A multiple computer system having at least one application program each written to operate on only a single computer but running simultaneously on a plurality of computers interconnected by a communications network, wherein different portions of said application program(s) execute substantially simultaneously on different ones of said computers, wherein each computer has an independent local memory accessible only by the corresponding portion of said application program(s) and wherein for each said portion a like plurality of substantially identical objects are created, each in the corresponding computer.
2. The system as claimed in claim 1 wherein each of said plurality of substantially identical objects has a substantially identical name.
3. The system as claimed in claim 2 wherein each said computer includes a distributed run time means with the distributed run time means of each said computer able to communicate with all other computers whereby if a portion of said application program(s) running on one of said computers changes the contents of an object in that computer then the change in content for said object is propagated by the distributed run time means of said one computer to all other computers to change the content of the corresponding object in each of said other computers.
4. The system as claimed in claim 3 wherein each said application program is modified before, during, or after loading by inserting an updating propagation routine to modify each instance at which said application program writes to memory, said updating propagation routine propagating every memory write by one computer to all said other computers.
5. The system as claimed in claim 4 wherein the application program is modified in accordance with a procedure selected from the group of procedures consisting of re-compilation at loading, pre-compilation prior to loading, compilation prior to loading, just-in-time compilation, and re-compilation after loading and before execution of the relevant portion of application program.
6. The system as claimed in claim 3 wherein said modified application program is transferred to all said computers in accordance with a procedure selected from the group consisting of master/slave transfer, branched transfer and cascaded transfer.
7. A plurality of computers interconnected via a communications link and each having an independent local memory and substantially simultaneously operating a different portion at least one application program each written to operate on only a single computer, each local memory being accessible only by the corresponding portion of said application program
8. The plurality of computers as claimed in claim 7 wherein each said computer in operating said at least one application program reads and writes only to local memory physically located in each said computer, the contents of the local memory utilized by each said computer is fundamentally similar but not, at each instant, identical, and every one of said computers has distribution update means to distribute to all other said computers the content of any memory location updated by said one computer.
9. The plurality of computers as claimed in claim 8 wherein the local memory capacity allocated to the or each said application program is substantially identical and the total memory capacity available to the or each said application program is said allocated memory capacity.
10. The plurality of computers as claimed in claim 8 wherein all said distribution update means communicate via said communications link at a data transfer rate which is substantially less than the local memory read rate.
11. The plurality of computers as claimed in claim 7 wherein at least some of said computers are manufactured by different manufacturers and/or have different operating systems.
12. A method of running simultaneously on a plurality of computers at least one application program each written to operate on only a single computer, said computers being interconnected by means of a communications network and each having an independent local memory, said method comprising the step of:
(i) executing different portions of said application program(s) on different ones of said computers and for each said portion creating a like plurality of substantially identical objects each in the corresponding computer and each accessible only by the corresponding portion of said application program.
13. The method as claimed in claim 12 comprising the further step of:
(ii) naming each of said plurality of substantially identical objects with a substantially identical name.
14. The method as claimed in claim 13 comprising the further step of:
(iii) if a portion of said application program running on one of said computers changes the contents of an object in that computer, then the change in content of said object is propagated to all of the other computers via said communications network to change the content of the corresponding object in each of said other computers.
15. The method as claimed in claim 14 including the further step of:
(iv) modifying said application program before, during or after loading by inserting an updating propagation routine to modify each instance at which said application program writes to memory, said updating propagation routine propagating every memory write by one computer to all said other computers.
16. The method as claimed in claim 15 including the further step of:
(v) modifying said application program utilizing a procedure selected from the group of procedures consisting of re-compilation at loading, pre-compilation prior to loading, compilation prior to loading, just-in-time compilation, and re-compilation after loading and before execution of the relevant portion of application program.
17. The method as claimed in claim 14 including the further step of:
(vi) transferring the modified application program to all said computers utilizing a procedure selected from the group consisting of master/slave transfer, branched transfer and cascaded transfer.
18. A method of loading an application program written to operate only on a single computer onto each of a plurality of computers, the computers being interconnected via a communications link, and different portions of said application program(s) being substantially simultaneously executable on different computers with each computer having an independent local memory accessible only by the corresponding portion of said application program(s), the method comprising the step of modifying the application before, during, or after loading and before execution of the relevant portion of the application program.
19. The method as claimed in claim 18 wherein the modification of the application is different for different computers.
20. The method as claimed in claim 18 wherein said modifying step comprises:—
(i) detecting instructions which share memory records utilizing one of said computers,
(ii) listing all such shared memory records and providing a naming tag for each listed memory record,
(iii) detecting those instructions which write to, or manipulate the contents of, any of said listed memory records, and
(iv) generating an updating propagation routine corresponding to each said detected write or manipulate instruction, said updating propagation routine forwarding the re-written or manipulated contents and name tag of each said re-written or manipulated listed memory record to all of the others of said computers.
21. A method of operating simultaneously on a plurality of computers all interconnected via a communications link at least one application program each written to operate on only a single computer, each of said computers having at least a minimum predetermined local memory capacity, different portions of said application program(s) being substantially simultaneously executed on different ones of said computers with the local memory of each computer being only accessible by the corresponding portion of said application program(s), said method comprising the steps of:
(i) initially providing each local memory in substantially identical condition,
(ii) satisfying all memory reads and writes generated by each said application program portion from said corresponding local memory, and
(iii) communicating via said communications link all said memory writes at each said computer which take place locally to all the remainder of said plurality of computers whereby the contents of the local memory utilised by each said computer, subject to an updating data transmission delay, remains substantially identical.
22. The method as claimed in claim 21 including the further step of:
(iv) communicating said local memory writes constituting an updating data transmission at a data transfer rate which is substantially less than the local memory read rate.
23. A method of compiling or modifying an application program written to operate on only a single computer but to run simultaneously on a plurality of computers interconnected via a communications link, with different portions of said application program(s) executing substantially simultaneously on different ones of said computers each of which has an independent local memory accessible only by the corresponding portion of said application program, said method comprising the steps of:
(i) detecting instructions which share memory records utilizing one of said computers,
(ii) listing all such shared memory records and providing a naming tag for each listed memory record,
(iii) detecting those instructions which write to, or manipulate the contents of, any of said listed memory records, and
(iv) activating an updating propagation routine following each said detected write or manipulate instruction, said updating propagation routine forwarding the re-written or manipulated contents and name tag of each said re-written or manipulated listed memory record to the remainder of said computers.
24. The method as claimed in claim 23 and carried out prior to loading the application program onto each said computer, or during loading of the application program onto each said computer, or after loading of the application program onto each said computer and before execution of the relevant portion of the application program.
25. In a multiple thread processing computer operation in which individual threads of a single application program written to operate on only a single computer are simultaneously being processed each on a different corresponding one of a plurality of computers each having an independent local memory accessible only by the corresponding thread and each being interconnected via a communications link, the improvement comprising communicating changes in the contents of local memory physically associated with the computer processing each thread to the local memory of each other said computer via said communications link.
26. The improvement as claimed in claim 25 wherein changes to the memory associated with one said thread are communicated by the computer of said one thread to all other said computers.
27. The improvement as claimed in claim 25 wherein changes to the memory associated with one said thread are transmitted to the computer associated with another said thread and are transmitted thereby to all said other computers.
28. A computer program product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the method as claimed in claim 12.
29. A computer program product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the method as claimed in claim 18.
30. A computer program product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the method as claimed in claim 21.
31. A computer program product comprising a set of program instructions stored in a storage medium and operable to permit a plurality of computers to carry out the method as claimed in claim 23.
32. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being programmed to carry out the method as claimed in claim 12.
33. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being programmed to carry out the method as claimed in claim 18.
34. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being programmed to carry out the method as claimed in claim 21.
35. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being programmed to carry out the method as claimed in claim 23.
36. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being loaded with the computer program product as claimed in claim 28.
37. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being loaded with the computer program product as claimed in claim 29.
38. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being loaded with the computer program product as claimed in claim 30.
39. A plurality of computers interconnected via a communication network and each operable to substantially simultaneously execute a different portion of an application program written to operate on only a single computer on a corresponding different one of computers, said computers being loaded with the computer program product as claimed in claim 31.
US11/259,634 2004-04-23 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with replicated memory Abandoned US20060265703A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/396,446 US7860829B2 (en) 2004-04-23 2009-03-02 Computer architecture and method of operation for multi-computer distributed processing with replicated memory
US12/820,758 US20100262590A1 (en) 2004-04-23 2010-06-22 Multi-computer distributed processing with replicated local memory exclusive read and write and network value update propagation

Applications Claiming Priority (10)

Application Number Priority Date Filing Date Title
AU2005902027A AU2005902027A0 (en) 2005-04-21 Modified Computer Architecture with Coordinated Objects
AU2005902024 2005-04-21
AU2005902023 2005-04-21
AU2005902025A AU2005902025A0 (en) 2005-04-21 Modified Computer Architecture with Finalization of Objects
AU2005902026A AU2005902026A0 (en) 2005-04-21 Multiple Computer Architecture with Synchronization
AU2005902027 2005-04-21
AU2005902026 2005-04-21
AU2005902025 2005-04-21
AU2005902024A AU2005902024A0 (en) 2005-04-21 Modified Computer Architecture with Initialization of Objects
AU2005902023A AU2005902023A0 (en) 2005-04-21 Multiple Computer Architecture with Replicated Memory Fields

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US11/111,757 Continuation-In-Part US20050257219A1 (en) 2004-04-23 2005-04-22 Multiple computer architecture with replicated memory fields
US11/259,744 Continuation-In-Part US20060253844A1 (en) 2004-04-23 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with initialization of objects

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US12/051,701 Continuation-In-Part US8316190B2 (en) 2007-04-06 2008-03-19 Computer architecture and method of operation for multi-computer distributed processing having redundant array of independent systems with replicated memory and code striping
US12/396,446 Continuation US7860829B2 (en) 2004-04-23 2009-03-02 Computer architecture and method of operation for multi-computer distributed processing with replicated memory

Publications (1)

Publication Number Publication Date
US20060265703A1 true US20060265703A1 (en) 2006-11-23

Family

ID=37114615

Family Applications (10)

Application Number Title Priority Date Filing Date
US11/259,744 Abandoned US20060253844A1 (en) 2004-04-23 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US11/259,762 Active 2027-01-24 US8028299B2 (en) 2005-04-21 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with finalization of objects
US11/259,761 Abandoned US20060265704A1 (en) 2005-04-21 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with synchronization
US11/259,885 Active 2026-12-11 US7788314B2 (en) 2004-04-23 2005-10-25 Multi-computer distributed processing with replicated local memory exclusive read and write and network value update propagation
US11/259,634 Abandoned US20060265703A1 (en) 2004-04-23 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with replicated memory
US11/912,141 Abandoned US20090055603A1 (en) 2005-04-21 2006-04-20 Modified computer architecture for a computer to operate in a multiple computer system
US12/340,303 Abandoned US20090198776A1 (en) 2004-04-23 2008-12-19 Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US12/343,419 Expired - Lifetime US7818296B2 (en) 2005-04-21 2008-12-23 Computer architecture and method of operation for multi-computer distributed processing with synchronization
US12/396,446 Expired - Lifetime US7860829B2 (en) 2004-04-23 2009-03-02 Computer architecture and method of operation for multi-computer distributed processing with replicated memory
US12/820,758 Abandoned US20100262590A1 (en) 2004-04-23 2010-06-22 Multi-computer distributed processing with replicated local memory exclusive read and write and network value update propagation

Family Applications Before (4)

Application Number Title Priority Date Filing Date
US11/259,744 Abandoned US20060253844A1 (en) 2004-04-23 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US11/259,762 Active 2027-01-24 US8028299B2 (en) 2005-04-21 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with finalization of objects
US11/259,761 Abandoned US20060265704A1 (en) 2005-04-21 2005-10-25 Computer architecture and method of operation for multi-computer distributed processing with synchronization
US11/259,885 Active 2026-12-11 US7788314B2 (en) 2004-04-23 2005-10-25 Multi-computer distributed processing with replicated local memory exclusive read and write and network value update propagation

Family Applications After (5)

Application Number Title Priority Date Filing Date
US11/912,141 Abandoned US20090055603A1 (en) 2005-04-21 2006-04-20 Modified computer architecture for a computer to operate in a multiple computer system
US12/340,303 Abandoned US20090198776A1 (en) 2004-04-23 2008-12-19 Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US12/343,419 Expired - Lifetime US7818296B2 (en) 2005-04-21 2008-12-23 Computer architecture and method of operation for multi-computer distributed processing with synchronization
US12/396,446 Expired - Lifetime US7860829B2 (en) 2004-04-23 2009-03-02 Computer architecture and method of operation for multi-computer distributed processing with replicated memory
US12/820,758 Abandoned US20100262590A1 (en) 2004-04-23 2010-06-22 Multi-computer distributed processing with replicated local memory exclusive read and write and network value update propagation

Country Status (3)

Country Link
US (10) US20060253844A1 (en)
EP (1) EP1880303A4 (en)
WO (2) WO2006110937A1 (en)

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050240737A1 (en) * 2004-04-23 2005-10-27 Waratek (Australia) Pty Limited Modified computer architecture
US20050257219A1 (en) * 2004-04-23 2005-11-17 Holt John M Multiple computer architecture with replicated memory fields
US20050262313A1 (en) * 2004-04-23 2005-11-24 Waratek Pty Limited Modified computer architecture with coordinated objects
US20050262513A1 (en) * 2004-04-23 2005-11-24 Waratek Pty Limited Modified computer architecture with initialization of objects
US20060020913A1 (en) * 2004-04-23 2006-01-26 Waratek Pty Limited Multiple computer architecture with synchronization
US20060253844A1 (en) * 2005-04-21 2006-11-09 Holt John M Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US20070100828A1 (en) * 2005-10-25 2007-05-03 Holt John M Modified machine architecture with machine redundancy
US20070100954A1 (en) * 2005-10-25 2007-05-03 Holt John M Modified machine architecture with partial memory updating
US20070100918A1 (en) * 2005-10-25 2007-05-03 Holt John M Multiple computer system with enhanced memory clean up
US20070101080A1 (en) * 2005-10-25 2007-05-03 Holt John M Multiple machine architecture with overhead reduction
US20070101057A1 (en) * 2005-10-25 2007-05-03 Holt John M Modified machine architecture with advanced synchronization
US20070126750A1 (en) * 2005-10-25 2007-06-07 Holt John M Replication of object graphs
US20070174734A1 (en) * 2005-10-25 2007-07-26 Holt John M Failure resistant multiple computer system and method
US20080114853A1 (en) * 2006-10-05 2008-05-15 Holt John M Network protocol for network communications
US20080114943A1 (en) * 2006-10-05 2008-05-15 Holt John M Adding one or more computers to a multiple computer system
US20080114896A1 (en) * 2006-10-05 2008-05-15 Holt John M Asynchronous data transmission
US20080114945A1 (en) * 2006-10-05 2008-05-15 Holt John M Contention detection
US20080120477A1 (en) * 2006-10-05 2008-05-22 Holt John M Contention detection with modified message format
US20080120478A1 (en) * 2006-10-05 2008-05-22 Holt John M Advanced synchronization and contention resolution
US20080126505A1 (en) * 2006-10-05 2008-05-29 Holt John M Multiple computer system with redundancy architecture
US20080126516A1 (en) * 2006-10-05 2008-05-29 Holt John M Advanced contention detection
US20080123642A1 (en) * 2006-10-05 2008-05-29 Holt John M Switch protocol for network communications
US20080126703A1 (en) * 2006-10-05 2008-05-29 Holt John M Cyclic redundant multiple computer architecture
US20080126322A1 (en) * 2006-10-05 2008-05-29 Holt John M Synchronization with partial memory replication
US20080133689A1 (en) * 2006-10-05 2008-06-05 Holt John M Silent memory reclamation
US20080133690A1 (en) * 2006-10-05 2008-06-05 Holt John M Contention detection and resolution
US20080133869A1 (en) * 2006-10-05 2008-06-05 Holt John M Redundant multiple computer architecture
US20080133870A1 (en) * 2006-10-05 2008-06-05 Holt John M Hybrid replicated shared memory
US20080134189A1 (en) * 2006-10-05 2008-06-05 Holt John M Job scheduling amongst multiple computers
US20080133884A1 (en) * 2006-10-05 2008-06-05 Holt John M Multiple network connections for multiple computers
US20080133859A1 (en) * 2006-10-05 2008-06-05 Holt John M Advanced synchronization and contention resolution
US20080130652A1 (en) * 2006-10-05 2008-06-05 Holt John M Multiple communication networks for multiple computers
US20080140801A1 (en) * 2006-10-05 2008-06-12 Holt John M Multiple computer system with dual mode redundancy architecture
US20080140975A1 (en) * 2006-10-05 2008-06-12 Holt John M Contention detection with data consolidation
US20080155127A1 (en) * 2006-10-05 2008-06-26 Holt John M Multi-path switching networks
US8316190B2 (en) 2007-04-06 2012-11-20 Waratek Pty. Ltd. Computer architecture and method of operation for multi-computer distributed processing having redundant array of independent systems with replicated memory and code striping
US20150019733A1 (en) * 2013-06-26 2015-01-15 Amazon Technologies, Inc. Management of computing sessions
US10313345B2 (en) 2013-03-11 2019-06-04 Amazon Technologies, Inc. Application marketplace for virtual desktops
US10616129B2 (en) 2013-03-11 2020-04-07 Amazon Technologies, Inc. Automated desktop placement

Families Citing this family (84)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060095483A1 (en) * 2004-04-23 2006-05-04 Waratek Pty Limited Modified computer architecture with finalization of objects
US7357332B2 (en) * 2005-04-29 2008-04-15 International Business Machines Corporation Providing diagnostic information on the processing of variables in source code
US20070157212A1 (en) * 2006-01-04 2007-07-05 Berg Douglas C Context key routing for parallel processing in an application serving environment
US7873953B1 (en) * 2006-01-20 2011-01-18 Altera Corporation High-level language code sequence optimization for implementing programmable chip designs
US20120124550A1 (en) * 2006-02-22 2012-05-17 Robert Nocera Facilitating database application code translation from a first application language to a second application language
US8156493B2 (en) * 2006-04-12 2012-04-10 The Mathworks, Inc. Exception handling in a concurrent computing process
US7747996B1 (en) * 2006-05-25 2010-06-29 Oracle America, Inc. Method of mixed lock-free and locking synchronization
US8082289B2 (en) 2006-06-13 2011-12-20 Advanced Cluster Systems, Inc. Cluster computing support for application programs
CA2557343C (en) * 2006-08-28 2015-09-22 Ibm Canada Limited-Ibm Canada Limitee Runtime code modification in a multi-threaded environment
WO2008040076A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Contention resolution with echo cancellation
US20080250221A1 (en) * 2006-10-09 2008-10-09 Holt John M Contention detection with data consolidation
US7934121B2 (en) 2006-11-21 2011-04-26 Microsoft Corporation Transparent replacement of a system processor
US8473460B2 (en) * 2006-11-21 2013-06-25 Microsoft Corporation Driver model for replacing core system hardware
US8332866B2 (en) * 2006-11-29 2012-12-11 Qualcomm Incorporated Methods, systems, and apparatus for object invocation across protection domain boundaries
US8341609B2 (en) * 2007-01-26 2012-12-25 Oracle International Corporation Code generation in the presence of paged memory
US8037460B2 (en) * 2007-01-26 2011-10-11 Oracle International Corporation Code persistence and dependency management for dynamic compilation in a database management system
US8413125B2 (en) * 2007-01-26 2013-04-02 Oracle International Corporation Asynchronous dynamic compilation based on multi-session profiling to produce shared native code
US8086906B2 (en) * 2007-02-15 2011-12-27 Microsoft Corporation Correlating hardware devices between local operating system and global management entity
US7685381B2 (en) * 2007-03-01 2010-03-23 International Business Machines Corporation Employing a data structure of readily accessible units of memory to facilitate memory access
US7899663B2 (en) * 2007-03-30 2011-03-01 International Business Machines Corporation Providing memory consistency in an emulated processing environment
US8458724B2 (en) 2007-06-15 2013-06-04 Microsoft Corporation Automatic mutual exclusion
US20090044186A1 (en) * 2007-08-07 2009-02-12 Nokia Corporation System and method for implementation of java ais api
US8291393B2 (en) * 2007-08-20 2012-10-16 International Business Machines Corporation Just-in-time compiler support for interruptible code
WO2009027138A1 (en) 2007-08-30 2009-03-05 International Business Machines Corporation Accessing data entities
US8181180B1 (en) * 2007-09-14 2012-05-15 Hewlett-Packard Development Company, L.P. Managing jobs in shared file systems
US8381174B2 (en) * 2007-10-31 2013-02-19 National Instruments Corporation Global variable structure in a graphical program
US8347266B2 (en) * 2007-12-10 2013-01-01 Microsoft Corporation Declarative object identity
US8250524B2 (en) * 2007-12-21 2012-08-21 Microsoft Corporation Contract programming for code error reduction
US10552391B2 (en) * 2008-04-04 2020-02-04 Landmark Graphics Corporation Systems and methods for real time data management in a collaborative environment
US8245222B2 (en) * 2008-07-09 2012-08-14 Aspect Software, Inc. Image installer
US8630976B2 (en) * 2008-08-20 2014-01-14 Sap Ag Fast search replication synchronization processes
US9542222B2 (en) * 2008-11-14 2017-01-10 Oracle International Corporation Resource broker system for dynamically deploying and managing software services in a virtual environment based on resource usage and service level agreement
US8645922B2 (en) * 2008-11-25 2014-02-04 Sap Ag System and method of implementing a concurrency profiler
US8307350B2 (en) * 2009-01-14 2012-11-06 Microsoft Corporation Multi level virtual function tables
US8782607B2 (en) 2009-02-20 2014-07-15 Microsoft Corporation Contract failure behavior with escalation policy
US9268608B2 (en) 2009-02-26 2016-02-23 Oracle International Corporation Automatic administration of UNIX commands
US8667483B2 (en) * 2009-03-25 2014-03-04 Microsoft Corporation Device dependent on-demand compiling and deployment of mobile applications
US10534644B2 (en) * 2009-06-25 2020-01-14 Wind River Systems, Inc. Method and system for a CPU-local storage mechanism
US8725402B2 (en) 2009-11-13 2014-05-13 The Boeing Company Loss of separation avoidance maneuvering
US9262933B2 (en) * 2009-11-13 2016-02-16 The Boeing Company Lateral avoidance maneuver solver
US8290991B2 (en) * 2009-12-15 2012-10-16 Juniper Networks, Inc. Atomic deletion of database data categories
US8214411B2 (en) * 2009-12-15 2012-07-03 Juniper Networks, Inc. Atomic deletion of database data categories
US8407444B2 (en) * 2009-12-23 2013-03-26 International Business Machines Corporation Hardware off-load garbage collection acceleration for languages with finalizers
US8943108B2 (en) * 2009-12-23 2015-01-27 International Business Machines Corporation Hardware off-load memory garbage collection acceleration
US8407531B2 (en) * 2010-02-26 2013-03-26 Bmc Software, Inc. Method of collecting and correlating locking data to determine ultimate holders in real time
WO2011130869A1 (en) * 2010-04-19 2011-10-27 Hewlett-Packard Development Company, L.P. Object linking based on determined linker order
RU2554509C2 (en) * 2010-10-06 2015-06-27 Александр Яковлевич Богданов System and method of distributed computations
US8453130B2 (en) * 2011-02-09 2013-05-28 International Business Machines Corporation Memory management for object oriented applications during runtime
US8615544B2 (en) * 2011-02-25 2013-12-24 Wyse Technology Inc. System and method for unlocking a device remotely from a server
US8572754B2 (en) 2011-02-25 2013-10-29 Wyse Technology Inc. System and method for facilitating unlocking a device connected locally to a client
US8671204B2 (en) 2011-06-29 2014-03-11 Qualcomm Incorporated Cooperative sharing of subscriptions to a subscriber-based network among M2M devices
US9275369B2 (en) 2011-08-24 2016-03-01 Oracle International Corporation Demystifying obfuscated information transfer for performing automated system administration
US10803028B2 (en) * 2011-12-21 2020-10-13 Sybase, Inc. Multiphase approach to data availability
US9015702B2 (en) * 2012-01-13 2015-04-21 Vasanth Bhat Determining compatibility of an application with different versions of an operating system
US9588756B2 (en) * 2012-03-22 2017-03-07 Irdeto B.V. Updating software components in a program
US9037558B2 (en) 2012-05-25 2015-05-19 International Business Machines Corporation Management of long-running locks and transactions on database tables
RU2012127578A (en) * 2012-07-02 2014-01-10 ЭлЭсАй Корпорейшн ANALYZER OF APPLICABILITY OF THE SOFTWARE MODULE FOR THE DEVELOPMENT AND TESTING OF THE SOFTWARE FOR MULTIPROCESSOR MEDIA
US8612407B1 (en) 2012-07-18 2013-12-17 International Business Machines Corporation Source control inheritance locking
US20140040218A1 (en) * 2012-07-31 2014-02-06 Hideaki Kimura Methods and systems for an intent lock engine
US9092281B2 (en) 2012-10-02 2015-07-28 Qualcomm Incorporated Fast remote procedure call
US9804945B1 (en) * 2013-01-03 2017-10-31 Amazon Technologies, Inc. Determinism for distributed applications
CN103067797B (en) * 2013-01-30 2015-04-08 烽火通信科技股份有限公司 Maintenance method of intelligent ODN (Optical Distribution Network) managing system
US9317472B2 (en) * 2013-06-07 2016-04-19 International Business Machines Corporation Processing element data sharing
US10540148B2 (en) * 2014-06-12 2020-01-21 Oracle International Corporation Complex constants
US20160070759A1 (en) * 2014-09-04 2016-03-10 Palo Alto Research Center Incorporated System And Method For Integrating Real-Time Query Engine And Database Platform
EP3189436B1 (en) 2014-09-05 2018-07-18 FTS Computertechnik GmbH Computer system and method for security sensible applications
US9853908B2 (en) 2014-11-25 2017-12-26 Red Hat Inc. Utilizing access control data structures for sharing computing resources
US10318271B2 (en) 2015-01-05 2019-06-11 Irdeto Canada Corporation Updating software components in a program
EP3256916B1 (en) 2015-02-12 2023-10-11 Glowforge Inc. Moving material during laser fabrication
US10509390B2 (en) 2015-02-12 2019-12-17 Glowforge Inc. Safety and reliability guarantees for laser fabrication
DE102015112143B4 (en) * 2015-07-24 2017-04-06 Infineon Technologies Ag A method of determining an integrity of an execution of a code fragment and a method of providing an abstract representation of a program code
US9977786B2 (en) * 2015-12-23 2018-05-22 Github, Inc. Distributed code repository with limited synchronization locking
SG10201600224SA (en) * 2016-01-12 2017-08-30 Huawei Int Pte Ltd Dedicated ssr pipeline stage of router for express traversal (extra) noc
CN107526742B (en) 2016-06-21 2021-10-08 伊姆西Ip控股有限责任公司 Method and apparatus for processing multilingual text
WO2018098396A1 (en) 2016-11-25 2018-05-31 Glowforge Inc. Multi-user computer-numerically-controlled machine
WO2018098397A1 (en) 2016-11-25 2018-05-31 Glowforge Inc. Calibration of computer-numerically-controlled machine
US20180225110A1 (en) * 2017-02-08 2018-08-09 International Business Machines Corporation Legacy program code analysis and optimization
US11132274B2 (en) * 2018-03-01 2021-09-28 Target Brands, Inc. Establishing and monitoring programming environments
CN108415719B (en) * 2018-03-29 2019-03-19 网易(杭州)网络有限公司 The hot update method of code and device, storage medium, processor and terminal
US10649685B2 (en) 2018-07-16 2020-05-12 International Business Machines Corporation Site-centric alerting in a distributed storage system
CN110113407B (en) * 2019-04-30 2021-08-17 上海连尚网络科技有限公司 Applet state synchronization method, apparatus and computer storage medium
US11263091B2 (en) * 2020-02-19 2022-03-01 International Business Machines Corporation Using inode entries to mirror data operations across data storage sites
US11698898B2 (en) * 2020-11-04 2023-07-11 Salesforce, Inc. Lock wait tracing
US12099476B2 (en) * 2021-10-22 2024-09-24 Dell Products L.P. Distributed smart lock system

Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5214776A (en) * 1988-11-18 1993-05-25 Bull Hn Information Systems Italia S.P.A. Multiprocessor system having global data replication
US5418966A (en) * 1992-10-16 1995-05-23 International Business Machines Corporation Updating replicated objects in a plurality of memory partitions
US5434994A (en) * 1994-05-23 1995-07-18 International Business Machines Corporation System and method for maintaining replicated data coherency in a data processing system
US5488723A (en) * 1992-05-25 1996-01-30 Cegelec Software system having replicated objects and using dynamic messaging, in particular for a monitoring/control installation of redundant architecture
US5918248A (en) * 1996-12-30 1999-06-29 Northern Telecom Limited Shared memory control algorithm for mutual exclusion and rollback
US6049809A (en) * 1996-10-30 2000-04-11 Microsoft Corporation Replication optimization system and method
US6163801A (en) * 1998-10-30 2000-12-19 Advanced Micro Devices, Inc. Dynamic communication between computer processes
US6192514B1 (en) * 1997-02-19 2001-02-20 Unisys Corporation Multicomputer system
US6323630B1 (en) * 1997-07-29 2001-11-27 Hironori Banba Reference voltage generation circuit and reference current generation circuit
US6370625B1 (en) * 1999-12-29 2002-04-09 Intel Corporation Method and apparatus for lock synchronization in a microprocessor system
US6389423B1 (en) * 1999-04-13 2002-05-14 Mitsubishi Denki Kabushiki Kaisha Data synchronization method for maintaining and controlling a replicated data
US6425016B1 (en) * 1997-05-27 2002-07-23 International Business Machines Corporation System and method for providing collaborative replicated objects for synchronous distributed groupware applications
US20030004924A1 (en) * 2001-06-29 2003-01-02 International Business Machines Corporation Apparatus for database record locking and method therefor
US6571278B1 (en) * 1998-10-22 2003-05-27 International Business Machines Corporation Computer data sharing system and method for maintaining replica consistency
US6574674B1 (en) * 1996-05-24 2003-06-03 Microsoft Corporation Method and system for managing data while sharing application programs
US6625751B1 (en) * 1999-08-11 2003-09-23 Sun Microsystems, Inc. Software fault tolerant computer system
US6668260B2 (en) * 2000-08-14 2003-12-23 Divine Technology Ventures System and method of synchronizing replicated data
US20040073828A1 (en) * 2002-08-30 2004-04-15 Vladimir Bronstein Transparent variable state mirroring
US6757896B1 (en) * 1999-01-29 2004-06-29 International Business Machines Corporation Method and apparatus for enabling partial replication of object stores
US6775831B1 (en) * 2000-02-11 2004-08-10 Overture Services, Inc. System and method for rapid completion of data processing tasks distributed on a network
US20050086384A1 (en) * 2003-09-04 2005-04-21 Johannes Ernst System and method for replicating, integrating and synchronizing distributed information
US20050108481A1 (en) * 2003-11-17 2005-05-19 Iyengar Arun K. System and method for achieving strong data consistency
US6968372B1 (en) * 2001-10-17 2005-11-22 Microsoft Corporation Distributed variable synchronizer
US20060041733A1 (en) * 2004-07-31 2006-02-23 Chris Hyser Operating-system-transparent distributed memory
US7010576B2 (en) * 2002-05-30 2006-03-07 International Business Machines Corporation Efficient method of globalization and synchronization of distributed resources in distributed peer data processing environments
US7020736B1 (en) * 2000-12-18 2006-03-28 Redback Networks Inc. Method and apparatus for sharing memory space across mutliple processing units
US7031989B2 (en) * 2001-02-26 2006-04-18 International Business Machines Corporation Dynamic seamless reconfiguration of executing parallel software
US7047341B2 (en) * 2001-12-29 2006-05-16 Lg Electronics Inc. Multi-processing memory duplication system
US7082604B2 (en) * 2001-04-20 2006-07-25 Mobile Agent Technologies, Incorporated Method and apparatus for breaking down computing tasks across a network of heterogeneous computer for parallel execution by utilizing autonomous mobile agents
US20060167878A1 (en) * 2005-01-27 2006-07-27 International Business Machines Corporation Customer statistics based on database lock use
US20060242464A1 (en) * 2004-04-23 2006-10-26 Holt John M Computer architecture and method of operation for multi-computer distributed processing and coordinated memory and asset handling

Family Cites Families (147)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4068298A (en) * 1975-12-03 1978-01-10 Systems Development Corporation Information storage and retrieval system
US4300127A (en) * 1978-09-27 1981-11-10 Bernin Victor M Solid state noncontacting keyboard employing a differential transformer element
JPS60186919A (en) 1984-01-30 1985-09-24 Nec Corp Autonomous timer circuit
US4780821A (en) * 1986-07-29 1988-10-25 International Business Machines Corp. Method for multiple programs management within a network having a server computer and a plurality of remote computers
US4969092A (en) 1988-09-30 1990-11-06 Ibm Corp. Method for scheduling execution of distributed application programs at preset times in an SNA LU 6.2 network environment
US5062037A (en) * 1988-10-24 1991-10-29 Ibm Corp. Method to provide concurrent execution of distributed application programs by a host computer and an intelligent work station on an sna network
EP0367188B1 (en) * 1988-11-01 1995-05-03 Asahi Kasei Kogyo Kabushiki Kaisha Thermoplastic polymer composition
EP0457308B1 (en) 1990-05-18 1997-01-22 Fujitsu Limited Data processing system having an input/output path disconnecting mechanism and method for controlling the data processing system
US6682608B2 (en) 1990-12-18 2004-01-27 Advanced Cardiovascular Systems, Inc. Superelastic guiding member
US5754207A (en) * 1992-08-12 1998-05-19 Hewlett-Packard Company Volume indicating ink reservoir cartridge system
US5581555A (en) * 1993-09-17 1996-12-03 Scientific-Atlanta, Inc. Reverse path allocation and contention resolution scheme for a broadband communications system
AU7684094A (en) 1993-09-24 1995-04-10 Oracle Corporation Method and apparatus for data replication
US5544345A (en) 1993-11-08 1996-08-06 International Business Machines Corporation Coherence controls for store-multiple shared data coordinated by cache directory entries in a shared electronic storage
JPH086854A (en) * 1993-12-23 1996-01-12 Unisys Corp Outboard-file-cache external processing complex
US5568605A (en) 1994-01-13 1996-10-22 International Business Machines Corporation Resolving conflicting topology information
US5692193A (en) * 1994-03-31 1997-11-25 Nec Research Institute, Inc. Software architecture for control of highly parallel computer systems
US6017118A (en) * 1995-04-27 2000-01-25 Hewlett-Packard Company High performance ink container with efficient construction
US6318850B1 (en) 1995-12-04 2001-11-20 Hewlett-Packard Company Ink container refurbishment system
JP3927600B2 (en) 1995-05-30 2007-06-13 コーポレーション フォー ナショナル リサーチ イニシアチブス System for distributed task execution
US5612865A (en) 1995-06-01 1997-03-18 Ncr Corporation Dynamic hashing method for optimal distribution of locks within a clustered system
GB9600823D0 (en) 1996-01-16 1996-03-20 British Telecomm Distributed processing
US5960087A (en) 1996-07-01 1999-09-28 Sun Microsystems, Inc. Distributed garbage collection system and method
US5802585A (en) * 1996-07-17 1998-09-01 Digital Equipment Corporation Batched checking of shared memory accesses
US6327630B1 (en) 1996-07-24 2001-12-04 Hewlett-Packard Company Ordered message reception in a distributed data processing system
US6760903B1 (en) * 1996-08-27 2004-07-06 Compuware Corporation Coordinated application monitoring in a distributed computing environment
US6314558B1 (en) 1996-08-27 2001-11-06 Compuware Corporation Byte code instrumentation
FR2756070B1 (en) 1996-11-18 1999-01-22 Bull Sa SYSTEM FOR MANAGING AND PROCESSING DISTRIBUTED OBJECT TRANSACTIONS AND METHOD IMPLEMENTED THEREWITH
US6148377A (en) 1996-11-22 2000-11-14 Mangosoft Corporation Shared memory computer networks
JPH10250104A (en) 1997-03-12 1998-09-22 Seiko Epson Corp Ink cartridge for ink-jet type recording apparatus, and its manufacture
US6633577B1 (en) 1997-03-26 2003-10-14 Nec Corporation Handshaking circuit for resolving contention on a transmission medium regardless of its length
US6048809A (en) * 1997-06-03 2000-04-11 Lear Automotive Dearborn, Inc. Vehicle headliner formed of polyester fibers
US6585359B1 (en) 1997-06-04 2003-07-01 Hewlett-Packard Development Company, L.P. Ink container providing pressurized ink with ink level sensor
US6010210A (en) 1997-06-04 2000-01-04 Hewlett-Packard Company Ink container having a multiple function chassis
US5913213A (en) 1997-06-16 1999-06-15 Telefonaktiebolaget L M Ericsson Lingering locks for replicated data objects
US6072953A (en) * 1997-09-30 2000-06-06 International Business Machines Corporation Apparatus and method for dynamically modifying class files during loading for execution
US6324587B1 (en) 1997-12-23 2001-11-27 Microsoft Corporation Method, computer program product, and data structure for publishing a data object over a store and forward transport
US6473773B1 (en) 1997-12-24 2002-10-29 International Business Machines Corporation Memory management in a partially garbage-collected programming system
US6449734B1 (en) 1998-04-17 2002-09-10 Microsoft Corporation Method and system for discarding locally committed transactions to ensure consistency in a server cluster
JP3866426B2 (en) 1998-11-05 2007-01-10 日本電気株式会社 Memory fault processing method in cluster computer and cluster computer
EP0969377B1 (en) 1998-06-30 2009-01-07 International Business Machines Corporation Method of replication-based garbage collection in a multiprocessor system
US6496871B1 (en) * 1998-06-30 2002-12-17 Nec Research Institute, Inc. Distributed agent software system and method having enhanced process mobility and communication in a computer network
JP2000094710A (en) 1998-09-24 2000-04-04 Seiko Epson Corp Print head device, ink jet printer and ink cartridge
US6460051B1 (en) 1998-10-28 2002-10-01 Starfish Software, Inc. System and methods for synchronizing datasets in a communication environment having high-latency or other adverse characteristics
US6266747B1 (en) 1998-10-30 2001-07-24 Telefonaktiebolaget Lm Ericsson (Publ) Method for writing data into data storage units
US6611955B1 (en) 1999-06-03 2003-08-26 Swisscom Ag Monitoring and testing middleware based application software
US6680942B2 (en) * 1999-07-02 2004-01-20 Cisco Technology, Inc. Directory services caching for network peer to peer service locator
US6578068B1 (en) * 1999-08-31 2003-06-10 Accenture Llp Load balancer in environment services patterns
GB9921720D0 (en) * 1999-09-14 1999-11-17 Tao Group Ltd Loading object-oriented computer programs
DE19961274C1 (en) * 1999-12-18 2001-02-15 Wella Ag New 2,5-diamino-1-(N-aminophenyl)-aminomethyl-benzene derivatives are used in colorants for keratinous fibers e.g. human hair
US6823511B1 (en) * 2000-01-10 2004-11-23 International Business Machines Corporation Reader-writer lock for multiprocessor systems
US20020161848A1 (en) 2000-03-03 2002-10-31 Willman Charles A. Systems and methods for facilitating memory access in information management environments
JP3416614B2 (en) 2000-04-26 2003-06-16 キヤノン株式会社 Ink jet recording device
US6922685B2 (en) * 2000-05-22 2005-07-26 Mci, Inc. Method and system for managing partitioned data resources
US20030005407A1 (en) * 2000-06-23 2003-01-02 Hines Kenneth J. System and method for coordination-centric design of software systems
IL153454A0 (en) * 2000-06-26 2003-07-06 Ibm Data management application programming interface for a parallel file system
US6826570B1 (en) * 2000-07-18 2004-11-30 International Business Machines Corporation Dynamically switching between different types of concurrency control techniques to provide an adaptive access strategy for a parallel file system
US6662359B1 (en) 2000-07-20 2003-12-09 International Business Machines Corporation System and method for injecting hooks into Java classes to handle exception and finalization processing
US6865585B1 (en) * 2000-07-31 2005-03-08 Microsoft Corporation Method and system for multiprocessor garbage collection
US6725014B1 (en) 2000-08-17 2004-04-20 Honeywell International, Inc. Method and system for contention resolution in radio frequency identification systems
US7058826B2 (en) * 2000-09-27 2006-06-06 Amphus, Inc. System, architecture, and method for logical server and other network devices in a dynamically configurable multi-server network environment
TW486661B (en) * 2000-10-05 2002-05-11 Elan Microelectronics Corp Input device using multi-dimension electrodes to define keys and its encoding method
AU2002232411A1 (en) 2000-11-28 2002-06-11 Gregory L. Gingerich A method and system for software and hardware multiplicity
US6754859B2 (en) 2001-01-03 2004-06-22 Bull Hn Information Systems Inc. Computer processor read/alter/rewrite optimization cache invalidate signals
US7383329B2 (en) 2001-02-13 2008-06-03 Aventail, Llc Distributed cache for state transfer operations
JP4026407B2 (en) 2001-05-17 2007-12-26 セイコーエプソン株式会社 Ink cartridge and ink jet recording apparatus using the same
US20020194015A1 (en) * 2001-05-29 2002-12-19 Incepto Ltd. Distributed database clustering using asynchronous transactional replication
US7047521B2 (en) 2001-06-07 2006-05-16 Lynoxworks, Inc. Dynamic instrumentation event trace system and methods
JP2003018204A (en) * 2001-07-02 2003-01-17 Hitachi Ltd Packet transfer device provided with flow detection function and flow management method
US6862608B2 (en) * 2001-07-17 2005-03-01 Storage Technology Corporation System and method for a distributed shared memory
WO2003017114A1 (en) 2001-08-20 2003-02-27 Gausa, Llc System and method for real-time multi-directional file-based data streaming editor
DE60211443T2 (en) 2001-10-05 2006-11-16 Canon K.K. Ink tank, liquid supply device and recording device
US20030097395A1 (en) * 2001-11-16 2003-05-22 Petersen Paul M. Executing irregular parallel control structures
US6668312B2 (en) * 2001-12-21 2003-12-23 Celoxica Ltd. System, method, and article of manufacture for dynamically profiling memory transfers in a program
US6779093B1 (en) 2002-02-15 2004-08-17 Veritas Operating Corporation Control facility for processing in-band control messages during data replication
CA2475857C (en) * 2002-03-11 2008-12-23 University Of Southern California Named entity translation
EP1495414B1 (en) 2002-03-25 2019-06-12 Open Invention Network LLC Transparent consistent active replication of multithreaded application programs
US7024519B2 (en) 2002-05-06 2006-04-04 Sony Computer Entertainment Inc. Methods and apparatus for controlling hierarchical cache memory
US7206827B2 (en) 2002-07-25 2007-04-17 Sun Microsystems, Inc. Dynamic administration framework for server systems
US6954794B2 (en) 2002-10-21 2005-10-11 Tekelec Methods and systems for exchanging reachability information and for switching traffic between redundant interfaces in a network cluster
US7287247B2 (en) * 2002-11-12 2007-10-23 Hewlett-Packard Development Company, L.P. Instrumenting a software application that includes distributed object technology
US7275239B2 (en) 2003-02-10 2007-09-25 International Business Machines Corporation Run-time wait tracing using byte code insertion
US7114150B2 (en) * 2003-02-13 2006-09-26 International Business Machines Corporation Apparatus and method for dynamic instrumenting of code to minimize system perturbation
TWI242131B (en) * 2003-04-25 2005-10-21 Via Networking Technologies In Method and related circuit for increasing network transmission efficiency by speeding data updating rate of memory
US20050010683A1 (en) 2003-06-30 2005-01-13 Prabhanjan Moleyar Apparatus, system and method for performing table maintenance
CN1275480C (en) * 2003-07-31 2006-09-13 上海贝尔阿尔卡特股份有限公司 Multi standard software radio (SDR) base band treating method
US20050039171A1 (en) 2003-08-12 2005-02-17 Avakian Arra E. Using interceptors and out-of-band data to monitor the performance of Java 2 enterprise edition (J2EE) applications
US7549149B2 (en) * 2003-08-21 2009-06-16 International Business Machines Corporation Automatic software distribution and installation in a multi-tiered computer network
US7412580B1 (en) * 2003-10-06 2008-08-12 Sun Microsystems, Inc. Concurrent incremental garbage collector with a card table summarizing modified reference locations
US20050086661A1 (en) * 2003-10-21 2005-04-21 Monnie David J. Object synchronization in shared object space
US7275241B2 (en) * 2003-11-21 2007-09-25 International Business Machines Corporation Dynamic instrumentation for a mixed mode virtual machine
US7380039B2 (en) * 2003-12-30 2008-05-27 3Tera, Inc. Apparatus, method and system for aggregrating computing resources
US7565653B2 (en) * 2004-02-20 2009-07-21 Sony Computer Entertainment Inc. Methods and apparatus for processor task migration in a multi-processor system
JP2005301590A (en) 2004-04-09 2005-10-27 Hitachi Ltd Storage system and data copying method
EP1763774B1 (en) * 2004-04-22 2012-12-05 Waratek Pty Limited Multiple computer architecture with replicated memory fields
US7849452B2 (en) 2004-04-23 2010-12-07 Waratek Pty Ltd. Modification of computer applications at load time for distributed execution
US20050257219A1 (en) 2004-04-23 2005-11-17 Holt John M Multiple computer architecture with replicated memory fields
US7707179B2 (en) 2004-04-23 2010-04-27 Waratek Pty Limited Multiple computer architecture with synchronization
US7844665B2 (en) 2004-04-23 2010-11-30 Waratek Pty Ltd. Modified computer architecture having coordinated deletion of corresponding replicated memory locations among plural computers
US20060095483A1 (en) * 2004-04-23 2006-05-04 Waratek Pty Limited Modified computer architecture with finalization of objects
US20050262513A1 (en) 2004-04-23 2005-11-24 Waratek Pty Limited Modified computer architecture with initialization of objects
US7639656B2 (en) 2004-04-28 2009-12-29 Symbol Technologies, Inc. Protocol for communication between access ports and wireless switches
US7149866B2 (en) * 2004-06-04 2006-12-12 International Business Machines Corporation Free item distribution among multiple free lists during garbage collection for more efficient object allocation
US20050289558A1 (en) 2004-06-08 2005-12-29 Daniel Illowsky Device interoperability runtime establishing event serialization and synchronization amongst a plurality of separate processing units and method for coordinating control data and operations
US7225371B2 (en) 2004-08-03 2007-05-29 International Business Machines Corporation Method and apparatus for storing and retrieving multiple point-in-time consistent data sets
US7590803B2 (en) * 2004-09-23 2009-09-15 Sap Ag Cache eviction
US7614045B2 (en) * 2004-09-24 2009-11-03 Sap (Ag) Sharing classes and class loaders
US20060075408A1 (en) * 2004-10-06 2006-04-06 Digipede Technologies, Llc Distributed object execution system
US7437516B2 (en) 2004-12-28 2008-10-14 Sap Ag Programming models for eviction policies
US7712081B2 (en) * 2005-01-19 2010-05-04 International Business Machines Corporation Using code motion and write and read delays to increase the probability of bug detection in concurrent systems
US8015236B2 (en) * 2005-10-25 2011-09-06 Waratek Pty. Ltd. Replication of objects having non-primitive fields, especially addresses
US7761670B2 (en) 2005-10-25 2010-07-20 Waratek Pty Limited Modified machine architecture with advanced synchronization
US7581069B2 (en) 2005-10-25 2009-08-25 Waratek Pty Ltd. Multiple computer system with enhanced memory clean up
US7958322B2 (en) * 2005-10-25 2011-06-07 Waratek Pty Ltd Multiple machine architecture with overhead reduction
US20070100828A1 (en) 2005-10-25 2007-05-03 Holt John M Modified machine architecture with machine redundancy
US7660960B2 (en) 2005-10-25 2010-02-09 Waratek Pty, Ltd. Modified machine architecture with partial memory updating
US7849369B2 (en) 2005-10-25 2010-12-07 Waratek Pty Ltd. Failure resistant multiple computer system and method
US7245556B1 (en) 2005-12-28 2007-07-17 Sandisk Corporation Methods for writing non-volatile memories for increased endurance
US7500067B2 (en) 2006-03-29 2009-03-03 Dell Products L.P. System and method for allocating memory to input-output devices in a multiprocessor computer system
US7647454B2 (en) 2006-06-12 2010-01-12 Hewlett-Packard Development Company, L.P. Transactional shared memory system and method of control
US20100121935A1 (en) * 2006-10-05 2010-05-13 Holt John M Hybrid replicated shared memory
US20080114899A1 (en) * 2006-10-05 2008-05-15 Holt John M Switch protocol for network communications
US20080120475A1 (en) * 2006-10-05 2008-05-22 Holt John M Adding one or more computers to a multiple computer system
US7831779B2 (en) 2006-10-05 2010-11-09 Waratek Pty Ltd. Advanced contention detection
WO2008040079A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Multiple network connections for multiple computers
US20080114896A1 (en) * 2006-10-05 2008-05-15 Holt John M Asynchronous data transmission
US20080133689A1 (en) * 2006-10-05 2008-06-05 Holt John M Silent memory reclamation
WO2008040076A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Contention resolution with echo cancellation
US20080141092A1 (en) 2006-10-05 2008-06-12 Holt John M Network protocol for network communications
WO2008040069A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Hybrid replicated shared memory
US7849151B2 (en) * 2006-10-05 2010-12-07 Waratek Pty Ltd. Contention detection
WO2008040066A1 (en) 2006-10-05 2008-04-10 Waratek Pty Limited Redundant multiple computer architecture
WO2008040074A1 (en) 2006-10-05 2008-04-10 Waratek Pty Limited Contention detection with data consolidation
US20080155127A1 (en) 2006-10-05 2008-06-26 Holt John M Multi-path switching networks
US20080120478A1 (en) 2006-10-05 2008-05-22 Holt John M Advanced synchronization and contention resolution
WO2008040082A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Multiple computer system with dual mode redundancy architecture
US20080140863A1 (en) * 2006-10-05 2008-06-12 Holt John M Multiple communication networks for multiple computers
WO2008040067A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Multiple computer with redundancy architecture
US20080127213A1 (en) * 2006-10-05 2008-05-29 Holt John M Contention resolution with counter rollover
US20080126703A1 (en) * 2006-10-05 2008-05-29 Holt John M Cyclic redundant multiple computer architecture
US20080133690A1 (en) * 2006-10-05 2008-06-05 Holt John M Contention detection and resolution
US20080140970A1 (en) 2006-10-05 2008-06-12 Holt John M Advanced synchronization and contention resolution
US20080133862A1 (en) * 2006-10-05 2008-06-05 Holt John M Contention detection with modified message format
WO2008040078A1 (en) * 2006-10-05 2008-04-10 Waratek Pty Limited Synchronization with partial memory replication
US20080250221A1 (en) 2006-10-09 2008-10-09 Holt John M Contention detection with data consolidation
US8195749B2 (en) * 2006-11-13 2012-06-05 Bindu Rama Rao Questionnaire server capable of providing questionnaires based on device capabilities
US8554981B2 (en) 2007-02-02 2013-10-08 Vmware, Inc. High availability virtual machine cluster
US8316190B2 (en) 2007-04-06 2012-11-20 Waratek Pty. Ltd. Computer architecture and method of operation for multi-computer distributed processing having redundant array of independent systems with replicated memory and code striping

Patent Citations (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5214776A (en) * 1988-11-18 1993-05-25 Bull Hn Information Systems Italia S.P.A. Multiprocessor system having global data replication
US5488723A (en) * 1992-05-25 1996-01-30 Cegelec Software system having replicated objects and using dynamic messaging, in particular for a monitoring/control installation of redundant architecture
US5418966A (en) * 1992-10-16 1995-05-23 International Business Machines Corporation Updating replicated objects in a plurality of memory partitions
US5434994A (en) * 1994-05-23 1995-07-18 International Business Machines Corporation System and method for maintaining replicated data coherency in a data processing system
US6574674B1 (en) * 1996-05-24 2003-06-03 Microsoft Corporation Method and system for managing data while sharing application programs
US6049809A (en) * 1996-10-30 2000-04-11 Microsoft Corporation Replication optimization system and method
US5918248A (en) * 1996-12-30 1999-06-29 Northern Telecom Limited Shared memory control algorithm for mutual exclusion and rollback
US6192514B1 (en) * 1997-02-19 2001-02-20 Unisys Corporation Multicomputer system
US6425016B1 (en) * 1997-05-27 2002-07-23 International Business Machines Corporation System and method for providing collaborative replicated objects for synchronous distributed groupware applications
US6323630B1 (en) * 1997-07-29 2001-11-27 Hironori Banba Reference voltage generation circuit and reference current generation circuit
US6571278B1 (en) * 1998-10-22 2003-05-27 International Business Machines Corporation Computer data sharing system and method for maintaining replica consistency
US6163801A (en) * 1998-10-30 2000-12-19 Advanced Micro Devices, Inc. Dynamic communication between computer processes
US6757896B1 (en) * 1999-01-29 2004-06-29 International Business Machines Corporation Method and apparatus for enabling partial replication of object stores
US6389423B1 (en) * 1999-04-13 2002-05-14 Mitsubishi Denki Kabushiki Kaisha Data synchronization method for maintaining and controlling a replicated data
US6625751B1 (en) * 1999-08-11 2003-09-23 Sun Microsystems, Inc. Software fault tolerant computer system
US6370625B1 (en) * 1999-12-29 2002-04-09 Intel Corporation Method and apparatus for lock synchronization in a microprocessor system
US6775831B1 (en) * 2000-02-11 2004-08-10 Overture Services, Inc. System and method for rapid completion of data processing tasks distributed on a network
US6668260B2 (en) * 2000-08-14 2003-12-23 Divine Technology Ventures System and method of synchronizing replicated data
US7020736B1 (en) * 2000-12-18 2006-03-28 Redback Networks Inc. Method and apparatus for sharing memory space across mutliple processing units
US7031989B2 (en) * 2001-02-26 2006-04-18 International Business Machines Corporation Dynamic seamless reconfiguration of executing parallel software
US7082604B2 (en) * 2001-04-20 2006-07-25 Mobile Agent Technologies, Incorporated Method and apparatus for breaking down computing tasks across a network of heterogeneous computer for parallel execution by utilizing autonomous mobile agents
US20030004924A1 (en) * 2001-06-29 2003-01-02 International Business Machines Corporation Apparatus for database record locking and method therefor
US6968372B1 (en) * 2001-10-17 2005-11-22 Microsoft Corporation Distributed variable synchronizer
US7047341B2 (en) * 2001-12-29 2006-05-16 Lg Electronics Inc. Multi-processing memory duplication system
US7010576B2 (en) * 2002-05-30 2006-03-07 International Business Machines Corporation Efficient method of globalization and synchronization of distributed resources in distributed peer data processing environments
US20040073828A1 (en) * 2002-08-30 2004-04-15 Vladimir Bronstein Transparent variable state mirroring
US20050086384A1 (en) * 2003-09-04 2005-04-21 Johannes Ernst System and method for replicating, integrating and synchronizing distributed information
US20050108481A1 (en) * 2003-11-17 2005-05-19 Iyengar Arun K. System and method for achieving strong data consistency
US20060242464A1 (en) * 2004-04-23 2006-10-26 Holt John M Computer architecture and method of operation for multi-computer distributed processing and coordinated memory and asset handling
US20060041733A1 (en) * 2004-07-31 2006-02-23 Chris Hyser Operating-system-transparent distributed memory
US20060167878A1 (en) * 2005-01-27 2006-07-27 International Business Machines Corporation Customer statistics based on database lock use
US20060253844A1 (en) * 2005-04-21 2006-11-09 Holt John M Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US20060265705A1 (en) * 2005-04-21 2006-11-23 Holt John M Computer architecture and method of operation for multi-computer distributed processing with finalization of objects
US20060265704A1 (en) * 2005-04-21 2006-11-23 Holt John M Computer architecture and method of operation for multi-computer distributed processing with synchronization

Cited By (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7844665B2 (en) 2004-04-23 2010-11-30 Waratek Pty Ltd. Modified computer architecture having coordinated deletion of corresponding replicated memory locations among plural computers
US20050240737A1 (en) * 2004-04-23 2005-10-27 Waratek (Australia) Pty Limited Modified computer architecture
US20050262313A1 (en) * 2004-04-23 2005-11-24 Waratek Pty Limited Modified computer architecture with coordinated objects
US20050262513A1 (en) * 2004-04-23 2005-11-24 Waratek Pty Limited Modified computer architecture with initialization of objects
US20090198776A1 (en) * 2004-04-23 2009-08-06 Waratek Pty Ltd. Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US7860829B2 (en) 2004-04-23 2010-12-28 Waratek Pty Ltd. Computer architecture and method of operation for multi-computer distributed processing with replicated memory
US20050257219A1 (en) * 2004-04-23 2005-11-17 Holt John M Multiple computer architecture with replicated memory fields
US7849452B2 (en) 2004-04-23 2010-12-07 Waratek Pty Ltd. Modification of computer applications at load time for distributed execution
US20060020913A1 (en) * 2004-04-23 2006-01-26 Waratek Pty Limited Multiple computer architecture with synchronization
US7707179B2 (en) 2004-04-23 2010-04-27 Waratek Pty Limited Multiple computer architecture with synchronization
US7788314B2 (en) 2004-04-23 2010-08-31 Waratek Pty Ltd. Multi-computer distributed processing with replicated local memory exclusive read and write and network value update propagation
US20090235034A1 (en) * 2005-04-21 2009-09-17 John Matthew Holt Computer architecture and method of operation for multi-computer distributed processing with synchronization
US8028299B2 (en) * 2005-04-21 2011-09-27 Waratek Pty, Ltd. Computer architecture and method of operation for multi-computer distributed processing with finalization of objects
US7818296B2 (en) 2005-04-21 2010-10-19 Waratek Pty Ltd. Computer architecture and method of operation for multi-computer distributed processing with synchronization
US20060265704A1 (en) * 2005-04-21 2006-11-23 Holt John M Computer architecture and method of operation for multi-computer distributed processing with synchronization
US20060253844A1 (en) * 2005-04-21 2006-11-09 Holt John M Computer architecture and method of operation for multi-computer distributed processing with initialization of objects
US7958322B2 (en) 2005-10-25 2011-06-07 Waratek Pty Ltd Multiple machine architecture with overhead reduction
US20070101080A1 (en) * 2005-10-25 2007-05-03 Holt John M Multiple machine architecture with overhead reduction
US8015236B2 (en) 2005-10-25 2011-09-06 Waratek Pty. Ltd. Replication of objects having non-primitive fields, especially addresses
US7996627B2 (en) 2005-10-25 2011-08-09 Waratek Pty Ltd Replication of object graphs
US20080189385A1 (en) * 2005-10-25 2008-08-07 Holt John M Multiple machine architecture with overhead reduction
US8122198B2 (en) 2005-10-25 2012-02-21 Waratek Pty Ltd. Modified machine architecture with partial memory updating
US8209393B2 (en) 2005-10-25 2012-06-26 Waratek Pty Ltd. Multiple machine architecture with overhead reduction
US20070174734A1 (en) * 2005-10-25 2007-07-26 Holt John M Failure resistant multiple computer system and method
US7849369B2 (en) 2005-10-25 2010-12-07 Waratek Pty Ltd. Failure resistant multiple computer system and method
US20070126750A1 (en) * 2005-10-25 2007-06-07 Holt John M Replication of object graphs
US20070101057A1 (en) * 2005-10-25 2007-05-03 Holt John M Modified machine architecture with advanced synchronization
US8122200B2 (en) 2005-10-25 2012-02-21 Waratek Pty Ltd. Modified machine architecture with advanced synchronization
US7761670B2 (en) 2005-10-25 2010-07-20 Waratek Pty Limited Modified machine architecture with advanced synchronization
US20070100918A1 (en) * 2005-10-25 2007-05-03 Holt John M Multiple computer system with enhanced memory clean up
US7660960B2 (en) 2005-10-25 2010-02-09 Waratek Pty, Ltd. Modified machine architecture with partial memory updating
US20070100954A1 (en) * 2005-10-25 2007-05-03 Holt John M Modified machine architecture with partial memory updating
US7581069B2 (en) 2005-10-25 2009-08-25 Waratek Pty Ltd. Multiple computer system with enhanced memory clean up
US20070100828A1 (en) * 2005-10-25 2007-05-03 Holt John M Modified machine architecture with machine redundancy
US20080215703A1 (en) * 2005-10-25 2008-09-04 Holt John M Modified machine architecture with partial memory updating
US20080215701A1 (en) * 2005-10-25 2008-09-04 Holt John M Modified machine architecture with advanced synchronization
US20080215593A1 (en) * 2005-10-25 2008-09-04 Holt John M Replication of object graphs
US20080215928A1 (en) * 2005-10-25 2008-09-04 Holt John M Failure resistant multiple computer system and method
US20080195617A1 (en) * 2005-10-25 2008-08-14 Holt John M Modified machine architecture with machine redundancy
US20080195682A1 (en) * 2005-10-25 2008-08-14 Holt John M Multiple computer system with enhanced memory clean up
US20080123642A1 (en) * 2006-10-05 2008-05-29 Holt John M Switch protocol for network communications
US20080126322A1 (en) * 2006-10-05 2008-05-29 Holt John M Synchronization with partial memory replication
US20080140975A1 (en) * 2006-10-05 2008-06-12 Holt John M Contention detection with data consolidation
US20080137662A1 (en) * 2006-10-05 2008-06-12 Holt John M Asynchronous data transmission
US20080140982A1 (en) * 2006-10-05 2008-06-12 Holt John M Redundant multiple computer architecture
US20080140856A1 (en) * 2006-10-05 2008-06-12 Holt John M Multiple communication networks for multiple computers
US20080141092A1 (en) * 2006-10-05 2008-06-12 Holt John M Network protocol for network communications
US20080140805A1 (en) * 2006-10-05 2008-06-12 Holt John M Multiple network connections for multiple computers
US20080155127A1 (en) * 2006-10-05 2008-06-26 Holt John M Multi-path switching networks
US20080151902A1 (en) * 2006-10-05 2008-06-26 Holt John M Multiple network connections for multiple computers
US20080184071A1 (en) * 2006-10-05 2008-07-31 Holt John M Cyclic redundant multiple computer architecture
US20080140863A1 (en) * 2006-10-05 2008-06-12 Holt John M Multiple communication networks for multiple computers
US20080130652A1 (en) * 2006-10-05 2008-06-05 Holt John M Multiple communication networks for multiple computers
US20080133711A1 (en) * 2006-10-05 2008-06-05 Holt John M Advanced contention detection
US20080133859A1 (en) * 2006-10-05 2008-06-05 Holt John M Advanced synchronization and contention resolution
US20080133692A1 (en) * 2006-10-05 2008-06-05 Holt John M Multiple computer system with redundancy architecture
US20080133884A1 (en) * 2006-10-05 2008-06-05 Holt John M Multiple network connections for multiple computers
US20080133694A1 (en) * 2006-10-05 2008-06-05 Holt John M Redundant multiple computer architecture
US20080133871A1 (en) * 2006-10-05 2008-06-05 Holt John M Hybrid replicated shared memory
US20080134189A1 (en) * 2006-10-05 2008-06-05 Holt John M Job scheduling amongst multiple computers
US20080133861A1 (en) * 2006-10-05 2008-06-05 Holt John M Silent memory reclamation
US20080133870A1 (en) * 2006-10-05 2008-06-05 Holt John M Hybrid replicated shared memory
US20080133869A1 (en) * 2006-10-05 2008-06-05 Holt John M Redundant multiple computer architecture
US20080133690A1 (en) * 2006-10-05 2008-06-05 Holt John M Contention detection and resolution
US20080133689A1 (en) * 2006-10-05 2008-06-05 Holt John M Silent memory reclamation
US20080126508A1 (en) * 2006-10-05 2008-05-29 Holt John M Synchronization with partial memory replication
US7831779B2 (en) 2006-10-05 2010-11-09 Waratek Pty Ltd. Advanced contention detection
US20080140801A1 (en) * 2006-10-05 2008-06-12 Holt John M Multiple computer system with dual mode redundancy architecture
US7849151B2 (en) 2006-10-05 2010-12-07 Waratek Pty Ltd. Contention detection
US20080126703A1 (en) * 2006-10-05 2008-05-29 Holt John M Cyclic redundant multiple computer architecture
US20080126516A1 (en) * 2006-10-05 2008-05-29 Holt John M Advanced contention detection
US7852845B2 (en) 2006-10-05 2010-12-14 Waratek Pty Ltd. Asynchronous data transmission
US20080126506A1 (en) * 2006-10-05 2008-05-29 Holt John M Multiple computer system with redundancy architecture
US7894341B2 (en) 2006-10-05 2011-02-22 Waratek Pty Ltd. Switch protocol for network communications
US7949837B2 (en) 2006-10-05 2011-05-24 Waratek Pty Ltd. Contention detection and resolution
US20080126505A1 (en) * 2006-10-05 2008-05-29 Holt John M Multiple computer system with redundancy architecture
US7958329B2 (en) 2006-10-05 2011-06-07 Waratek Pty Ltd Hybrid replicated shared memory
US7962697B2 (en) 2006-10-05 2011-06-14 Waratek Pty Limited Contention detection
US7971005B2 (en) 2006-10-05 2011-06-28 Waratek Pty Ltd. Advanced contention detection
US20080120478A1 (en) * 2006-10-05 2008-05-22 Holt John M Advanced synchronization and contention resolution
US20080120477A1 (en) * 2006-10-05 2008-05-22 Holt John M Contention detection with modified message format
US20080114945A1 (en) * 2006-10-05 2008-05-15 Holt John M Contention detection
US8086805B2 (en) 2006-10-05 2011-12-27 Waratek Pty Ltd. Advanced contention detection
US8090926B2 (en) 2006-10-05 2012-01-03 Waratek Pty Ltd. Hybrid replicated shared memory
US8095616B2 (en) 2006-10-05 2012-01-10 Waratek Pty Ltd. Contention detection
US20080114896A1 (en) * 2006-10-05 2008-05-15 Holt John M Asynchronous data transmission
US20080114943A1 (en) * 2006-10-05 2008-05-15 Holt John M Adding one or more computers to a multiple computer system
US20080114853A1 (en) * 2006-10-05 2008-05-15 Holt John M Network protocol for network communications
US8473564B2 (en) 2006-10-05 2013-06-25 Waratek Pty Ltd. Contention detection and resolution
US8316190B2 (en) 2007-04-06 2012-11-20 Waratek Pty. Ltd. Computer architecture and method of operation for multi-computer distributed processing having redundant array of independent systems with replicated memory and code striping
US10313345B2 (en) 2013-03-11 2019-06-04 Amazon Technologies, Inc. Application marketplace for virtual desktops
US10616129B2 (en) 2013-03-11 2020-04-07 Amazon Technologies, Inc. Automated desktop placement
US20150019733A1 (en) * 2013-06-26 2015-01-15 Amazon Technologies, Inc. Management of computing sessions
US10623243B2 (en) * 2013-06-26 2020-04-14 Amazon Technologies, Inc. Management of computing sessions

Also Published As

Publication number Publication date
US20060265704A1 (en) 2006-11-23
US7818296B2 (en) 2010-10-19
US20090198776A1 (en) 2009-08-06
WO2006110937A1 (en) 2006-10-26
US20090235033A1 (en) 2009-09-17
US7788314B2 (en) 2010-08-31
EP1880303A1 (en) 2008-01-23
US7860829B2 (en) 2010-12-28
US20090235034A1 (en) 2009-09-17
US20060242464A1 (en) 2006-10-26
US20090055603A1 (en) 2009-02-26
EP1880303A4 (en) 2008-12-31
WO2006110957A1 (en) 2006-10-26
US20060265705A1 (en) 2006-11-23
US8028299B2 (en) 2011-09-27
US20100262590A1 (en) 2010-10-14
US20060253844A1 (en) 2006-11-09

Similar Documents

Publication Publication Date Title
US7860829B2 (en) Computer architecture and method of operation for multi-computer distributed processing with replicated memory
EP1763774B1 (en) Multiple computer architecture with replicated memory fields
US20050262513A1 (en) Modified computer architecture with initialization of objects
US7844665B2 (en) Modified computer architecture having coordinated deletion of corresponding replicated memory locations among plural computers
US20060095483A1 (en) Modified computer architecture with finalization of objects
US20060020913A1 (en) Multiple computer architecture with synchronization
US20080140762A1 (en) Job scheduling amongst multiple computers
US7660960B2 (en) Modified machine architecture with partial memory updating
JP2009512030A (en) Duplicate object graph
AU2005236089A1 (en) Multiple computer architecture with replicated memory fields
JP2009512028A (en) Improved machine architecture with partial memory update
AU2006301909B2 (en) Modified machine architecture with partial memory updating
JP2009512029A (en) Multi-computer system with extended memory cleanup
WO2007041760A1 (en) Modified machine architecture with advanced synchronization
AU2005236085A1 (en) Modified computer architecture with initialization of objects
AU2005236086A1 (en) Multiple computer architecture with synchronization
AU2006301907A1 (en) Modified machine architecture with advanced synchronization
AU2006238334A1 (en) Modified computer architecture for a computer to operate in a multiple computer system
AU2005236088A1 (en) Modified computer architecture with finalization of objects
JP2009512081A (en) Multi-machine architecture with reduced overhead

Legal Events

Date Code Title Description
AS Assignment

Owner name: WARATEK PTY LTD, AUSTRALIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HOLT, JOHN M;REEL/FRAME:018133/0314

Effective date: 20060731

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION