US20030163758A1 - Method and system to identify a memory corruption source within a multiprocessor system - Google Patents
Method and system to identify a memory corruption source within a multiprocessor system Download PDFInfo
- Publication number
- US20030163758A1 US20030163758A1 US10/087,920 US8792002A US2003163758A1 US 20030163758 A1 US20030163758 A1 US 20030163758A1 US 8792002 A US8792002 A US 8792002A US 2003163758 A1 US2003163758 A1 US 2003163758A1
- Authority
- US
- United States
- Prior art keywords
- corrupt data
- processor
- computer system
- memory
- memory address
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/0703—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
- G06F11/079—Root cause analysis, i.e. error or fault diagnosis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/0703—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
- G06F11/0706—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment
- G06F11/0712—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment in a virtual computing platform, e.g. logically partitioned systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/0703—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
- G06F11/0706—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment
- G06F11/0721—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment within a central processing unit [CPU]
- G06F11/0724—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment within a central processing unit [CPU] in a multiprocessor or a multi-core unit
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/0703—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation
- G06F11/0706—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment
- G06F11/073—Error or fault processing not based on redundancy, i.e. by taking additional measures to deal with the error or fault not making use of redundancy in operation, in hardware, or in data representation the processing taking place on a specific hardware platform or in a specific software environment in a memory management context, e.g. virtual memory or cache management
Definitions
- the present invention relates in general to the field of computers, and in particular, to the field of data storage. Still more particularly, the present invention relates to an improved method and system for identifying a source of corrupt data in memory.
- An LPAR computer system partitions its multiple processors into discrete processing partitions.
- Each processing partition may be a single processor, or may be a group of processors.
- Each processing partition operates under a single operating system (OS), and typically runs one program at a time, although simultaneous multiprocessing (a.k.a. multitasking) of multiple programs within a processing partition is common.
- the OS of each processing partition may be the same or different OS used by other processing partitions, and the processing partitions may run the same or different programs as other processing partitions.
- Each processing partition has its own private memory, which is either a separate physical memory or a reserved partition of a main memory in the LPAR computer system. When a processing partition has multiple processors executing a single program, this process if referred to as parallel processing.
- a non-LPAR computer system simultaneously uses multiple processors to execute a single program operating under a common OS. Unlike the LPAR computer system, the non-LPAR computer system shares a single memory address space, typically a memory partition in main memory. If each processor takes the same time to access main memory, the non-LPAR computer system is called a uniform memory access (UMA) multiprocessor or symmetric multiprocessor (SMP). If memory accesses are faster for some processors compared to others within the non-LPAR computer system, the computer system is called a nonuniform memory access (NUMA) multiprocessor.
- UMA uniform memory access
- SMP symmetric multiprocessor
- LPAR computer systems are designed such that each processing partition uses a separate memory or, more typically, a partition of main memory.
- the LPAR architecture protocol prohibits one processing partition from using memory in another processing partition's memory partition.
- a hardware or software error can sometimes occur, resulting in corrupt data being stored in an unauthorized memory address location.
- DABR hardware Data Address Break
- a logic analyzer records a processor's operation history, including data storage, by measuring activity on external pins of the processor.
- the logic analyzer is an intelligent piece of hardware that physically fits over a processor to contact the processor's pins, and creates a log of signals at the pins, including data storage instructions.
- most multiprocessor systems do not have the required amount of physical space needed to position a logic analyzer on top of a processor, and thus cannot be used.
- the present invention is a method and system for identifying a source of a corrupt data in a memory in a multiprocessor computer system.
- debugging software locates and identifies the corrupt data that caused a program failure (crash).
- the multiprocessor computer system is shut down, and the corrupt data is cleared from the memory.
- a processor is selected to load and run monitor code designed to monitor the location where the corrupt data was stored.
- the crashed system is then restarted.
- the selected processor detects re-storage of the corrupt data in the same memory address, all system operations are immediately suspended.
- the registers of all suspected processors that may have stored the corrupt data are inspected to determine the source of the corrupt data, thus allowing the problem to be corrected.
- the present invention is particularly useful in logical partition (LPAR) computer systems that prohibit access to memory partitions by processors using an OS that is not permitted by the memory partition.
- LPAR logical partition
- the selected processor used to monitor the memory address for the corrupt data is isolated before being loaded with any OS. Monitoring code, which is independent of any OS, is loaded into the selected processor, which is able to cross different memory partitions. Thus, the selected processor is able to monitor the content of any memory location in any memory partition in an LPAR computer system.
- FIG. 1 depicts an exemplary non-logical partition (non-LPAR) computer system used with the present invention
- FIG. 2 illustrates an exemplary logical partition (LPAR) computer system used with the present invention
- FIG. 3 depicts an exemplary logical partition in the LPAR computer system illustrated in FIG. 2;
- FIG. 4 is a flow chart of a process logic used by the present invention to identify a source of corrupt data in an LPAR computer system
- FIG. 5 is a flow chart of a process logic used by the present invention to identify a source of corrupt data in a non-LPAR computer system.
- non-LPAR non-logical partition
- SMP system 10 includes multiple processors 12 a - 12 n , each processor 12 having a respective cache memory 14 a - 14 n .
- Processors 12 and cache memories 14 are connected to a single bus 16 , which connects to a main memory 18 , an input/output (I/O) interface 20 , and a network interface 21 .
- Cache memories 14 may be level 1, level 2 or higher level cache memory that references main memory 18 using any method known in the art, including direct mapping.
- I/O interface 20 provides an interface to I/O devices, including monitors, keyboards, pointers, printers, etc.
- Network interface 21 provides an interface to connect SMP system 10 to other computer devices via a network (not shown), such as a local-area network (LAN), wide-area network (WAN) or Internet.
- LAN local-area network
- WAN wide-area network
- Internet Internet
- SMP system 10 may encounter corrupt data 19 within main memory 18 .
- Corrupt data 19 may have its source in any processor 12 , but initially that source will be unknown to the user debugging the program.
- one of the processors 12 which is a least affected processor, is utilized to monitor main memory 18 for corrupt data 19 at a specific memory address in main memory 18 .
- LPAR computer system 23 includes multiple logical partitions, one of which, logical partition 25 , is depicted in FIG. 3.
- Logical partition 25 includes a processor partition 21 a , a memory partition 40 a and preferably an I/O 38 a .
- Memory partition 40 a and I/O 38 a connect to processor partition 21 a via a bus 34 as depicted.
- Processor partitions 21 , memory partitions 40 , and I/O's 38 are described in further detail below.
- LPAR computer system 23 includes multiple processor partitions 21 a - 21 n .
- each processor partition 21 has multiple central processor units (CPU's).
- each processor partition may have only a single CPU.
- each CPU in each processor partition 21 has a cache memory 24 .
- processor partition 21 a includes CPU's 22 a - 22 n , each CPU 22 having an associated cache memory 24 a - 24 n ;
- processor partition 21 b includes CPU's 26 a - 26 n , each CPU 26 having an associated cache memory 28 a - 28 n ;
- processor partition 21 n includes CPU's 30 a - 30 n , each CPU 30 having an associated cache memory 32 a - 32 n.
- Each processor partition 21 is connected to bus 34 , which is further connected to a main memory 36 and an I/O interface 38 .
- I/O interface 38 serves an analogous function as I/O interface 20 described for the non-LPAR computer system depicted in FIG. 1.
- each processor partition 21 has its own I/O 38 , such that processor partition 21 a uses I/O 38 a , processor partition 21 b uses I/O 38 b , and processor partition 21 n uses I/O 38 n.
- Main memory 36 is partitioned into memory partitions 40 a - 40 n , such that each processor partition 21 has its own private memory partition 40 .
- processor partition 21 a uses memory partition 40 a
- processor partition 21 b uses memory partition 40 b
- processor partition 21 n uses memory partition 40 n .
- main memory 36 may include corrupt data 41 at a specific address within one of the memory partitions 40 .
- one of the CPU's in one of the processor partitions 21 will be isolated and utilized, free of an operating system (OS), to monitor main memory 36 to identify the source of corrupt data 41 .
- OS operating system
- FIG. 4 there is depicted a flow chart of the method for identifying corrupt data as contemplated by the present invention when used with an LPAR computer system, such as depicted in FIG. 2.
- a query is made in query block 44 as to whether a program failure has occurred.
- the program failure may be a single program running on all processing partitions in the LPAR computer system, or it may be a failure of one program out of several running simultaneously on the LPAR computer system.
- a program failure has not occurred, then no further steps are taken. If a program failure has occurred, such as a system crash or a program crash, the cause of the crash is assumed to be the result of corrupt data being stored in an unauthorized memory partition of main memory by an unauthorized processing partition. For example, as depicted in FIG. 2, one of the CPU's 22 in processing partition 21 a may have caused the storage of corrupt data 41 in memory partition 40 b . Under LPAR protocol, processing partition 21 a should only store data in its private memory partition 40 a . In the example described here, however, a CPU 22 in processing partition 21 a either directly stored corrupt data 41 in memory partition 40 b , or else processing partition 21 a initially stored valid data in memory partition 40 a .
- corrupt data 41 is shown as being in a single location in a specific memory partition 40 of main memory 36 , corrupt data may be in multiple memory locations. That is, there may be corrupt data stored in several unauthorized locations in main memory, or corrupt data may be stored in both unauthorized cache memory locations as well as unauthorized main memory locations. For purposes of explanatory simplicity, it will be assumed that a single main memory address contains a single corrupt data.
- the operation of all processor partitions is suspended.
- the suspension of operations may be limited to only those suspected processor partitions suspected of causing the software failure. For simplicity, it will be assumed that all processing partitions are suspect, and thus are all initially suspended.
- the memory address of the corrupt data that caused the software failure is identified, through the use of debugging software that does not affect the address of the corrupt data, and, as shown in block 48 , the corrupt data and its memory address location are stored in a memory area that will not be overwritten and will not affect the memory address of the corrupt data.
- monitor code to be run by an appropriated processor as described below in block 54 as a monitor processor, is stored, likewise in a memory location that will not be overwritten and will not affect the corrupt data memory address.
- the memory block used is the same as that previously allocated in main memory for the appropriated processor described below in block 54 .
- the memory address location that contains the corrupt data is cleared, and the LPAR computer system is then booted as described in block 52 to a “standby state.”
- the standby state all processors in the LPAR computer system are in a working state, but have not been allocated to an OS. That is, each logical partition in the LPAR computer system is re-booted to a point just before loading a specific OS for each logical partition.
- OS operating system
- Access to memory partition in a LPAR computer system is limited to a processing partition in the same logical partition.
- Each logical partition operates under a single OS.
- no OS is loaded into any logical partition until a processor from one of the logical partitions is selected to operate as a monitor processor, such that the monitor processor is OS independent to allow the monitor processor to access any memory partition.
- a CPU (processor) from one of the processor partitions is appropriated to monitor corrupt data found in one of the memory partitions.
- the processor chosen is selected from a processing partition that is the least affected by the software failure. That is, the processor chosen is preferably from the processing partition that is the least likely to have either caused or been affected by the corrupt data storage. Since access to a specific memory partition would be prohibited if the appropriated processor is running under a prohibited OS, the system re-boot is stopped before an OS is loaded. The appropriated processor is then loaded with monitoring code that is OS independent.
- the monitor code is a short piece of software code that instructs the appropriated processor to monitor a specific address in main memory for the storage of a specific piece of data. The specific address and specific piece of data are those stored earlier as described in block 48 .
- the LPAR computer system's logical partitions are booted, and the program that crashed earlier is restarted, as described in block 57 .
- all logical partition processors except for the appropriated monitoring processor, are rebooted with an operating system to run programs that were running at the time of the software crash.
- the appropriated processor running the monitor code then monitors main memory to identify a storing event of the specific corrupt data at the specified memory address location, as described in block 58 .
- the corrupt data store event is detected, as described in block 60 , all processors suspected of storing corrupt data are suspended as shown in block 62 .
- the registers of the suspect processors are examined, as described in block 64 , to identify which processor in which processor partition is responsible for storing the corrupt data in the main memory. Once the offending processor is identified, then steps are taken, as shown in block 66 , to correct the problem causing the corrupt data, whether that problem is software or hardware related.
- FIG. 5 there is depicted a flow chart of the method for identifying corrupt data as contemplated by the present invention when used with a non-LPAR computer system, such as illustrated in FIG. 1.
- a query is made in query block 70 as to whether a program failure has occurred.
- the program failure may be a single program running on all processors in the non-LPAR computer system, or it may be a failure of one program out of several running simultaneously on the non-LPAR computer system.
- a program failure has not occurred, then no further steps are taken. If a program failure has occurred, such as a system crash or a program crash, the cause of the crash is assumed to be the result of corrupt data being stored in an unauthorized memory address in main memory.
- a program failure such as a system crash or a program crash
- the cause of the crash is assumed to be the result of corrupt data being stored in an unauthorized memory address in main memory.
- one of the processors 12 may have caused the storage of corrupt data 19 in main memory 16 .
- the corrupt data may have been the result of improper function of a memory controller (not shown). While corrupt data 19 is shown as being in a single location in main memory 16 , corrupt data may be in multiple memory locations. That is, there may be corrupt data stored in several unauthorized locations in main memory, or corrupt data may be stored in both unauthorized cache memory locations as well as unauthorized main memory locations.
- a single main memory address contains a single corrupt data.
- the operation of all processors is suspended, as described in block 72 .
- the suspension of operations may be limited to only those processors suspected of causing the software failure. For simplicity, it will be assumed that all processors are suspect, and thus are all initially suspended.
- the memory location of the corrupt data that caused the software failure is identified, preferably through the use of debugging software that does not affect the corrupt data memory address, and, as shown in block 76 and block 77 , the corrupt data and its memory address location, plus the monitor code to be used by a monitor processor described below in block 80 , are stored in a memory area that will not be overwritten and will not affect the corrupt data memory address. As illustrated in block 78 , the memory address location that contains the corrupt data is then cleared.
- the non-LPAR computer system is booted, and the crashed program that caused the corrupt data storage is restored.
- a processor is then appropriated to function as the monitor processor to monitor the corrupt data memory address for re-storage of the corrupt data. Since non-LPAR computer systems do not have the OS constraints described above in FIG. 4 for LPAR computer systems, a processor from the non-LPAR computer system is simply appropriated, and executes the monitoring software stored as described in block 77 , which is capable of comparing the content of a specific memory location with the known corrupt data, which was previously stored as described above in block 76 .
- the appropriated processor running the monitor code then monitors main memory to identify a storing event of the specific corrupt data at the specified memory address location, as described in block 82 .
- the corrupt data store event is detected, as described in block 84 , all processors suspected of storing corrupt data are suspended, as shown in block 86 .
- the registers of the suspect processors are examined, as described in block 88 , to identify which processor in which processor partition is responsible for storing the corrupt data in the main memory. Once the offending processor is identified, then steps are taken, as shown in block 90 , to correct the problem causing the corrupt data, whether that problem is software or hardware related.
- the present invention is therefore able to monitor a specific memory location using a dedicated processor appropriated from multiple processors in a multiprocessor computer system.
- the dedicated processor is able to monitor a specific memory address location using software that is not limited by an OS.
- the invention works well with either a non-LPAR computer system or an LPAR computer system.
- signal bearing media include, without limitation, recordable type media such as floppy disk or compact disk, read-only-memories (CDROMs) and transmission type media such as analog or digital communication links.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Quality & Reliability (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Techniques For Improving Reliability Of Storages (AREA)
Abstract
Description
- 1. Technical Field
- The present invention relates in general to the field of computers, and in particular, to the field of data storage. Still more particularly, the present invention relates to an improved method and system for identifying a source of corrupt data in memory.
- 2. Description of the Related Art
- As computer processing becomes more complex, the need for higher computer performance increases. One method of addressing this need is the use of multiple processors, executing the same or different programs, within a computing system. While many architectures use multiple processors, such architectures may be categorized as either Logical Partition (LPAR) computer systems or non-LPAR computer systems.
- An LPAR computer system partitions its multiple processors into discrete processing partitions. Each processing partition may be a single processor, or may be a group of processors. Each processing partition operates under a single operating system (OS), and typically runs one program at a time, although simultaneous multiprocessing (a.k.a. multitasking) of multiple programs within a processing partition is common. The OS of each processing partition may be the same or different OS used by other processing partitions, and the processing partitions may run the same or different programs as other processing partitions. Each processing partition has its own private memory, which is either a separate physical memory or a reserved partition of a main memory in the LPAR computer system. When a processing partition has multiple processors executing a single program, this process if referred to as parallel processing.
- A non-LPAR computer system simultaneously uses multiple processors to execute a single program operating under a common OS. Unlike the LPAR computer system, the non-LPAR computer system shares a single memory address space, typically a memory partition in main memory. If each processor takes the same time to access main memory, the non-LPAR computer system is called a uniform memory access (UMA) multiprocessor or symmetric multiprocessor (SMP). If memory accesses are faster for some processors compared to others within the non-LPAR computer system, the computer system is called a nonuniform memory access (NUMA) multiprocessor.
- As described above, LPAR computer systems are designed such that each processing partition uses a separate memory or, more typically, a partition of main memory. The LPAR architecture protocol prohibits one processing partition from using memory in another processing partition's memory partition. However, a hardware or software error can sometimes occur, resulting in corrupt data being stored in an unauthorized memory address location.
- During execution of a computer program, valid data may be written several times to a memory address. However, when corrupt data is stored to that memory address, program failure often results. In an LPAR computer system, the corrupt data is often the result of one logical partition storing, either directly or indirectly, data to another logical partition's memory. After program failure, the corrupt data and the main memory address in which the corrupt data is stored can be identified. However, conventional debugging software is unable to determine the cause and source of the corrupt data for several reasons.
- First, loading debugging software in a continuous main memory typically causes an uninitialized pointer problem. That is, loading debugging software in main memory often causes the memory location where the corrupt data originally occurred to move, thus making monitoring future corrupt data stores difficult, if not impossible. Second, in an LPAR computer system, prior art debugging software is OS dependent, and thus is unable to communicate cross logical partitions. That is, debugging software under a specific OS is not able to monitor a memory of a first logical partition operating under a different OS. Further, the debugging software cannot access a processor of a second logical partition that is the source of the corrupt data if it is also under a different OS from that used by the debugging software. Finally, a hardware Data Address Break (DABR) is unusable since many valid data writes to a memory address may occur. That is, the mere storage of data to the corrupt data address may or may not be the storage of corrupt data, thus making use of a DABR flag unhelpful.
- In the prior art, the offending processor that erroneously stored corrupt data to a prohibited memory address is sometimes identified using hardware called a logic analyzer. A logic analyzer records a processor's operation history, including data storage, by measuring activity on external pins of the processor. The logic analyzer is an intelligent piece of hardware that physically fits over a processor to contact the processor's pins, and creates a log of signals at the pins, including data storage instructions. However, most multiprocessor systems do not have the required amount of physical space needed to position a logic analyzer on top of a processor, and thus cannot be used.
- Therefore, there exists a need for a tool that has unrestricted access to all memory on a system and the ability to identify a specific value of a corrupt data at a specific memory address. The tool should have the further ability to identify the source of the corrupt data.
- The present invention is a method and system for identifying a source of a corrupt data in a memory in a multiprocessor computer system. When a computer program fails, debugging software locates and identifies the corrupt data that caused a program failure (crash). The multiprocessor computer system is shut down, and the corrupt data is cleared from the memory. During a restart of the multiprocessor computer system, a processor is selected to load and run monitor code designed to monitor the location where the corrupt data was stored.
- The crashed system is then restarted. When the selected processor detects re-storage of the corrupt data in the same memory address, all system operations are immediately suspended. The registers of all suspected processors that may have stored the corrupt data are inspected to determine the source of the corrupt data, thus allowing the problem to be corrected.
- The present invention is particularly useful in logical partition (LPAR) computer systems that prohibit access to memory partitions by processors using an OS that is not permitted by the memory partition. The selected processor used to monitor the memory address for the corrupt data is isolated before being loaded with any OS. Monitoring code, which is independent of any OS, is loaded into the selected processor, which is able to cross different memory partitions. Thus, the selected processor is able to monitor the content of any memory location in any memory partition in an LPAR computer system.
- The above, as well as additional objectives, features and advantages of the present invention will become apparent in the following detailed written description.
- The novel features believed characteristic of the invention are set forth in the appended claims. The invention itself, however, as well as a preferred mode of use, further objects and advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
- FIG. 1 depicts an exemplary non-logical partition (non-LPAR) computer system used with the present invention;
- FIG. 2 illustrates an exemplary logical partition (LPAR) computer system used with the present invention;
- FIG. 3 depicts an exemplary logical partition in the LPAR computer system illustrated in FIG. 2;
- FIG. 4 is a flow chart of a process logic used by the present invention to identify a source of corrupt data in an LPAR computer system; and
- FIG. 5 is a flow chart of a process logic used by the present invention to identify a source of corrupt data in a non-LPAR computer system.
- With reference now to the drawings and in particular to FIG. 1, there is depicted an exemplary non-logical partition (non-LPAR) computer system configured as a symmetric multiprocessor (SMP)
system 10.SMP system 10 includes multiple processors 12 a-12 n, each processor 12 having a respective cache memory 14 a-14 n. Processors 12 and cache memories 14 are connected to asingle bus 16, which connects to amain memory 18, an input/output (I/O)interface 20, and anetwork interface 21. Cache memories 14 may belevel 1, level 2 or higher level cache memory that referencesmain memory 18 using any method known in the art, including direct mapping. I/O interface 20 provides an interface to I/O devices, including monitors, keyboards, pointers, printers, etc.Network interface 21 provides an interface to connectSMP system 10 to other computer devices via a network (not shown), such as a local-area network (LAN), wide-area network (WAN) or Internet. - When using parallel processing software,
SMP system 10 may encountercorrupt data 19 withinmain memory 18.Corrupt data 19 may have its source in any processor 12, but initially that source will be unknown to the user debugging the program. In a process described further and illustrated in FIG. 4, one of the processors 12, which is a least affected processor, is utilized to monitormain memory 18 forcorrupt data 19 at a specific memory address inmain memory 18. - With reference now to FIG. 2, there is illustrated a block diagram of an exemplary logical partition (LPAR)
computer system 23.LPAR computer system 23 includes multiple logical partitions, one of which,logical partition 25, is depicted in FIG. 3.Logical partition 25 includes aprocessor partition 21 a, amemory partition 40 a and preferably an I/O 38 a.Memory partition 40 a and I/O 38 a connect toprocessor partition 21 a via abus 34 as depicted.Processor partitions 21, memory partitions 40, and I/O's 38 are described in further detail below. - Returning again to FIG. 2,
LPAR computer system 23 includesmultiple processor partitions 21 a-21 n. In the example shown, eachprocessor partition 21 has multiple central processor units (CPU's). Alternatively, each processor partition may have only a single CPU. In the exemplary system depicted, each CPU in eachprocessor partition 21 has a cache memory 24. As depicted,processor partition 21 a includes CPU's 22 a-22 n, each CPU 22 having an associated cache memory 24 a-24 n;processor partition 21 b includes CPU's 26 a-26 n, each CPU 26 having an associated cache memory 28 a-28 n; andprocessor partition 21 n includes CPU's 30 a-30 n, each CPU 30 having an associated cache memory 32 a-32 n. - Each
processor partition 21 is connected tobus 34, which is further connected to amain memory 36 and an I/O interface 38. I/O interface 38 serves an analogous function as I/O interface 20 described for the non-LPAR computer system depicted in FIG. 1. As depicted in FIG. 2, eachprocessor partition 21 has its own I/O 38, such thatprocessor partition 21 a uses I/O 38 a,processor partition 21 b uses I/O 38 b, andprocessor partition 21 n uses I/O 38 n. -
Main memory 36 is partitioned into memory partitions 40 a-40 n, such that eachprocessor partition 21 has its own private memory partition 40. Thusprocessor partition 21 auses memory partition 40 a,processor partition 21 b usesmemory partition 40 b, andprocessor partition 21 n usesmemory partition 40 n. As will be described below,main memory 36 may includecorrupt data 41 at a specific address within one of the memory partitions 40. As discussed below, one of the CPU's in one of theprocessor partitions 21 will be isolated and utilized, free of an operating system (OS), to monitormain memory 36 to identify the source ofcorrupt data 41. - With reference now to FIG. 4, there is depicted a flow chart of the method for identifying corrupt data as contemplated by the present invention when used with an LPAR computer system, such as depicted in FIG. 2. Starting at
block 42, a query is made inquery block 44 as to whether a program failure has occurred. The program failure may be a single program running on all processing partitions in the LPAR computer system, or it may be a failure of one program out of several running simultaneously on the LPAR computer system. - If a program failure has not occurred, then no further steps are taken. If a program failure has occurred, such as a system crash or a program crash, the cause of the crash is assumed to be the result of corrupt data being stored in an unauthorized memory partition of main memory by an unauthorized processing partition. For example, as depicted in FIG. 2, one of the CPU's22 in
processing partition 21 a may have caused the storage ofcorrupt data 41 inmemory partition 40 b. Under LPAR protocol,processing partition 21 a should only store data in itsprivate memory partition 40 a. In the example described here, however, a CPU 22 inprocessing partition 21 a either directly storedcorrupt data 41 inmemory partition 40 b, or else processingpartition 21 a initially stored valid data inmemory partition 40 a. The valid data then migrated tomemory partition 40 b to be stored ascorrupt data 41. Whilecorrupt data 41 is shown as being in a single location in a specific memory partition 40 ofmain memory 36, corrupt data may be in multiple memory locations. That is, there may be corrupt data stored in several unauthorized locations in main memory, or corrupt data may be stored in both unauthorized cache memory locations as well as unauthorized main memory locations. For purposes of explanatory simplicity, it will be assumed that a single main memory address contains a single corrupt data. - As described in
block 45, the operation of all processor partitions is suspended. In an alternative embodiment, the suspension of operations may be limited to only those suspected processor partitions suspected of causing the software failure. For simplicity, it will be assumed that all processing partitions are suspect, and thus are all initially suspended. - As described in
block 46, the memory address of the corrupt data that caused the software failure is identified, through the use of debugging software that does not affect the address of the corrupt data, and, as shown in block 48, the corrupt data and its memory address location are stored in a memory area that will not be overwritten and will not affect the memory address of the corrupt data. As depicted inblock 49, monitor code, to be run by an appropriated processor as described below inblock 54 as a monitor processor, is stored, likewise in a memory location that will not be overwritten and will not affect the corrupt data memory address. In a preferred embodiment, the memory block used is the same as that previously allocated in main memory for the appropriated processor described below inblock 54. - As illustrated in
block 50, the memory address location that contains the corrupt data is cleared, and the LPAR computer system is then booted as described inblock 52 to a “standby state.” In the standby state, all processors in the LPAR computer system are in a working state, but have not been allocated to an OS. That is, each logical partition in the LPAR computer system is re-booted to a point just before loading a specific OS for each logical partition. Thus, a specific CPU in the LPAR computer system can be isolated and free of any operating system (OS), which significance is now described. - Access to memory partition in a LPAR computer system is limited to a processing partition in the same logical partition. Each logical partition operates under a single OS. Thus, preferably no OS is loaded into any logical partition until a processor from one of the logical partitions is selected to operate as a monitor processor, such that the monitor processor is OS independent to allow the monitor processor to access any memory partition.
- Thus, as described in
block 54, a CPU (processor) from one of the processor partitions is appropriated to monitor corrupt data found in one of the memory partitions. The processor chosen is selected from a processing partition that is the least affected by the software failure. That is, the processor chosen is preferably from the processing partition that is the least likely to have either caused or been affected by the corrupt data storage. Since access to a specific memory partition would be prohibited if the appropriated processor is running under a prohibited OS, the system re-boot is stopped before an OS is loaded. The appropriated processor is then loaded with monitoring code that is OS independent. The monitor code is a short piece of software code that instructs the appropriated processor to monitor a specific address in main memory for the storage of a specific piece of data. The specific address and specific piece of data are those stored earlier as described in block 48. - Continuing with FIG. 4, the LPAR computer system's logical partitions are booted, and the program that crashed earlier is restarted, as described in
block 57. Thus all logical partition processors, except for the appropriated monitoring processor, are rebooted with an operating system to run programs that were running at the time of the software crash. The appropriated processor running the monitor code then monitors main memory to identify a storing event of the specific corrupt data at the specified memory address location, as described inblock 58. When the corrupt data store event is detected, as described inblock 60, all processors suspected of storing corrupt data are suspended as shown inblock 62. The registers of the suspect processors are examined, as described inblock 64, to identify which processor in which processor partition is responsible for storing the corrupt data in the main memory. Once the offending processor is identified, then steps are taken, as shown inblock 66, to correct the problem causing the corrupt data, whether that problem is software or hardware related. - With reference now to FIG. 5, there is depicted a flow chart of the method for identifying corrupt data as contemplated by the present invention when used with a non-LPAR computer system, such as illustrated in FIG. 1. Starting at
block 68, a query is made inquery block 70 as to whether a program failure has occurred. The program failure may be a single program running on all processors in the non-LPAR computer system, or it may be a failure of one program out of several running simultaneously on the non-LPAR computer system. - If a program failure has not occurred, then no further steps are taken. If a program failure has occurred, such as a system crash or a program crash, the cause of the crash is assumed to be the result of corrupt data being stored in an unauthorized memory address in main memory. For example, as depicted in FIG. 1, one of the processors12 may have caused the storage of
corrupt data 19 inmain memory 16. The corrupt data may have been the result of improper function of a memory controller (not shown). Whilecorrupt data 19 is shown as being in a single location inmain memory 16, corrupt data may be in multiple memory locations. That is, there may be corrupt data stored in several unauthorized locations in main memory, or corrupt data may be stored in both unauthorized cache memory locations as well as unauthorized main memory locations. For purposes of explanatory simplicity, it will be assumed that a single main memory address contains a single corrupt data. - Referring again to FIG. 5, the operation of all processors is suspended, as described in
block 72. In an alternative embodiment, the suspension of operations may be limited to only those processors suspected of causing the software failure. For simplicity, it will be assumed that all processors are suspect, and thus are all initially suspended. - As described in
block 74, the memory location of the corrupt data that caused the software failure is identified, preferably through the use of debugging software that does not affect the corrupt data memory address, and, as shown inblock 76 andblock 77, the corrupt data and its memory address location, plus the monitor code to be used by a monitor processor described below inblock 80, are stored in a memory area that will not be overwritten and will not affect the corrupt data memory address. As illustrated inblock 78, the memory address location that contains the corrupt data is then cleared. - As illustrated in
block 79, the non-LPAR computer system is booted, and the crashed program that caused the corrupt data storage is restored. As described inblock 80, a processor is then appropriated to function as the monitor processor to monitor the corrupt data memory address for re-storage of the corrupt data. Since non-LPAR computer systems do not have the OS constraints described above in FIG. 4 for LPAR computer systems, a processor from the non-LPAR computer system is simply appropriated, and executes the monitoring software stored as described inblock 77, which is capable of comparing the content of a specific memory location with the known corrupt data, which was previously stored as described above inblock 76. - The appropriated processor running the monitor code then monitors main memory to identify a storing event of the specific corrupt data at the specified memory address location, as described in
block 82. When the corrupt data store event is detected, as described inblock 84, all processors suspected of storing corrupt data are suspended, as shown inblock 86. The registers of the suspect processors are examined, as described inblock 88, to identify which processor in which processor partition is responsible for storing the corrupt data in the main memory. Once the offending processor is identified, then steps are taken, as shown inblock 90, to correct the problem causing the corrupt data, whether that problem is software or hardware related. - The present invention is therefore able to monitor a specific memory location using a dedicated processor appropriated from multiple processors in a multiprocessor computer system. The dedicated processor is able to monitor a specific memory address location using software that is not limited by an OS. Thus, the invention works well with either a non-LPAR computer system or an LPAR computer system.
- It should be appreciated that the method described above for identifying the source of corrupt data can be embodied in a computer program product in a variety of forms, and that the present invention applies equally regardless of the particular type of signal bearing media utilized to actually carry out the method described in the invention. Examples of signal bearing media include, without limitation, recordable type media such as floppy disk or compact disk, read-only-memories (CDROMs) and transmission type media such as analog or digital communication links.
- While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention.
Claims (18)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/087,920 US6845470B2 (en) | 2002-02-27 | 2002-02-27 | Method and system to identify a memory corruption source within a multiprocessor system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/087,920 US6845470B2 (en) | 2002-02-27 | 2002-02-27 | Method and system to identify a memory corruption source within a multiprocessor system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20030163758A1 true US20030163758A1 (en) | 2003-08-28 |
US6845470B2 US6845470B2 (en) | 2005-01-18 |
Family
ID=27753955
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/087,920 Expired - Fee Related US6845470B2 (en) | 2002-02-27 | 2002-02-27 | Method and system to identify a memory corruption source within a multiprocessor system |
Country Status (1)
Country | Link |
---|---|
US (1) | US6845470B2 (en) |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060036909A1 (en) * | 2004-08-13 | 2006-02-16 | Seakr Engineering, Incorporated | Soft error detection and recovery |
US20060184717A1 (en) * | 2005-02-17 | 2006-08-17 | Intel Corporation | Integrated circuit capable of flash memory storage management |
US20060212762A1 (en) * | 2005-03-21 | 2006-09-21 | Zimmer Vincent J | Error management topologies |
US20070294584A1 (en) * | 2006-04-28 | 2007-12-20 | Microsoft Corporation | Detection and isolation of data items causing computer process crashes |
US20110173501A1 (en) * | 2008-09-26 | 2011-07-14 | Microsoft Corporation | Memory management techniques selectively using mitigations to reduce errors |
US20120151251A1 (en) * | 2010-12-08 | 2012-06-14 | Advanced Micro Devices, Inc. | Queue freeze on protocol error |
US8443066B1 (en) | 2004-02-13 | 2013-05-14 | Oracle International Corporation | Programmatic instantiation, and provisioning of servers |
US8458390B2 (en) | 2004-02-13 | 2013-06-04 | Oracle International Corporation | Methods and systems for handling inter-process and inter-module communications in servers and server clusters |
US8601053B2 (en) | 2004-02-13 | 2013-12-03 | Oracle International Corporation | Multi-chassis fabric-backplane enterprise servers |
US8713295B2 (en) | 2004-07-12 | 2014-04-29 | Oracle International Corporation | Fabric-backplane enterprise servers with pluggable I/O sub-system |
US8743872B2 (en) | 2004-02-13 | 2014-06-03 | Oracle International Corporation | Storage traffic communication via a switch fabric in accordance with a VLAN |
US8848727B2 (en) | 2004-02-13 | 2014-09-30 | Oracle International Corporation | Hierarchical transport protocol stack for data transfer between enterprise servers |
US8868790B2 (en) | 2004-02-13 | 2014-10-21 | Oracle International Corporation | Processor-memory module performance acceleration in fabric-backplane enterprise servers |
US9772894B2 (en) * | 2016-01-29 | 2017-09-26 | Netapp, Inc. | Systems, methods, and machine-readable media to perform state data collection |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7111200B2 (en) * | 2003-04-17 | 2006-09-19 | International Business Machines Corporation | Apparatus and method for debugging a logical partition |
US20050097141A1 (en) * | 2003-10-30 | 2005-05-05 | International Business Machines Corporation | Autonomic filesystem recovery |
US7363544B2 (en) * | 2003-10-30 | 2008-04-22 | International Business Machines Corporation | Program debug method and apparatus |
US20050165862A1 (en) * | 2004-01-12 | 2005-07-28 | International Business Machines Corporation | Autonomic and fully recovering filesystem operations |
US7613954B2 (en) * | 2004-12-21 | 2009-11-03 | National Instruments Corporation | Test executive with stack corruption detection |
US7391312B2 (en) * | 2005-04-22 | 2008-06-24 | Microsoft Corporation | Method and system for an incidental feedback platform |
US20070083867A1 (en) * | 2005-09-09 | 2007-04-12 | International Business Machines Corporation | Method and system to recover from control block hangs in a heterogenous multiprocessor environment |
US7502957B2 (en) * | 2005-09-09 | 2009-03-10 | International Business Machines Corporation | Method and system to execute recovery in non-homogeneous multi processor environments |
US7457985B2 (en) * | 2005-09-09 | 2008-11-25 | International Business Machines Corporation | Method to detect errors in computer systems by using state tracking |
US7996585B2 (en) * | 2005-09-09 | 2011-08-09 | International Business Machines Corporation | Method and system for state tracking and recovery in multiprocessing computing systems |
JP2007226413A (en) * | 2006-02-22 | 2007-09-06 | Hitachi Ltd | Memory dump method, memory dump program and computer system |
US7647509B2 (en) * | 2006-05-12 | 2010-01-12 | Intel Corporation | Method and apparatus for managing power in a processing system with multiple partitions |
US7797555B2 (en) * | 2006-05-12 | 2010-09-14 | Intel Corporation | Method and apparatus for managing power from a sequestered partition of a processing system |
US7840946B2 (en) * | 2006-06-02 | 2010-11-23 | International Business Machines Corporation | System and method for matching a plurality of ordered sequences with applications to call stack analysis to identify known software problems |
US8140908B2 (en) * | 2007-06-22 | 2012-03-20 | Microsoft Corporation | System and method of client side analysis for identifying failing RAM after a user mode or kernel mode exception |
US20090006902A1 (en) * | 2007-06-29 | 2009-01-01 | International Business Machines Corporation | Methods, systems, and computer program products for reporting fru failures in storage device enclosures |
US8141045B2 (en) * | 2007-12-14 | 2012-03-20 | International Business Machines Corporation | Automatically identifying the source of copied software |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5513315A (en) * | 1992-12-22 | 1996-04-30 | Microsoft Corporation | System and method for automatic testing of computer software |
US6009258A (en) * | 1997-09-26 | 1999-12-28 | Symantec Corporation | Methods and devices for unwinding stack of frozen program and for restarting the program from unwound state |
US6230284B1 (en) * | 1998-04-24 | 2001-05-08 | Intel Corporation | Initiating corrective action after the occurrence of a program error |
US6412082B1 (en) * | 1997-12-17 | 2002-06-25 | Sony Corporation | Method and apparatus for selecting computer programs based on an error detection mechanism |
US20020133738A1 (en) * | 1999-11-10 | 2002-09-19 | Art Zeigler | Methods and systems for saving data potentially modified by a crashed computer program executing in a preemptive multitasking operating system environment |
US6502208B1 (en) * | 1997-03-31 | 2002-12-31 | International Business Machines Corporation | Method and system for check stop error handling |
US20030046612A1 (en) * | 2001-08-31 | 2003-03-06 | James Grey | System and method enabling execution stop and restart of a test executive sequence(s) |
US20030084377A1 (en) * | 2001-10-31 | 2003-05-01 | Parks Jeff A. | Process activity and error monitoring system and method |
US6591379B1 (en) * | 2000-06-23 | 2003-07-08 | Microsoft Corporation | Method and system for injecting an exception to recover unsaved data |
US6675295B1 (en) * | 2000-06-19 | 2004-01-06 | Microsoft Corporation | Method and computer system for detecting and correcting a failure in a computer application program during startup |
US6701454B1 (en) * | 2000-06-05 | 2004-03-02 | Microsoft Corporation | Method and system for recovering information during a program failure |
US6772367B1 (en) * | 1999-10-13 | 2004-08-03 | Board Of Regents, The University Of Texas System | Software fault tolerance of concurrent programs using controlled re-execution |
US6795916B2 (en) * | 2000-11-29 | 2004-09-21 | Mitac International Corp. | Method for updating microcode in a system wherein a keyboard BIOS expects an updated signal within a time period after microcode update commences |
-
2002
- 2002-02-27 US US10/087,920 patent/US6845470B2/en not_active Expired - Fee Related
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5513315A (en) * | 1992-12-22 | 1996-04-30 | Microsoft Corporation | System and method for automatic testing of computer software |
US6502208B1 (en) * | 1997-03-31 | 2002-12-31 | International Business Machines Corporation | Method and system for check stop error handling |
US6009258A (en) * | 1997-09-26 | 1999-12-28 | Symantec Corporation | Methods and devices for unwinding stack of frozen program and for restarting the program from unwound state |
US6412082B1 (en) * | 1997-12-17 | 2002-06-25 | Sony Corporation | Method and apparatus for selecting computer programs based on an error detection mechanism |
US6230284B1 (en) * | 1998-04-24 | 2001-05-08 | Intel Corporation | Initiating corrective action after the occurrence of a program error |
US6772367B1 (en) * | 1999-10-13 | 2004-08-03 | Board Of Regents, The University Of Texas System | Software fault tolerance of concurrent programs using controlled re-execution |
US20020133738A1 (en) * | 1999-11-10 | 2002-09-19 | Art Zeigler | Methods and systems for saving data potentially modified by a crashed computer program executing in a preemptive multitasking operating system environment |
US6701454B1 (en) * | 2000-06-05 | 2004-03-02 | Microsoft Corporation | Method and system for recovering information during a program failure |
US6675295B1 (en) * | 2000-06-19 | 2004-01-06 | Microsoft Corporation | Method and computer system for detecting and correcting a failure in a computer application program during startup |
US6591379B1 (en) * | 2000-06-23 | 2003-07-08 | Microsoft Corporation | Method and system for injecting an exception to recover unsaved data |
US6795916B2 (en) * | 2000-11-29 | 2004-09-21 | Mitac International Corp. | Method for updating microcode in a system wherein a keyboard BIOS expects an updated signal within a time period after microcode update commences |
US20030046612A1 (en) * | 2001-08-31 | 2003-03-06 | James Grey | System and method enabling execution stop and restart of a test executive sequence(s) |
US20030084377A1 (en) * | 2001-10-31 | 2003-05-01 | Parks Jeff A. | Process activity and error monitoring system and method |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8443066B1 (en) | 2004-02-13 | 2013-05-14 | Oracle International Corporation | Programmatic instantiation, and provisioning of servers |
US8868790B2 (en) | 2004-02-13 | 2014-10-21 | Oracle International Corporation | Processor-memory module performance acceleration in fabric-backplane enterprise servers |
US8848727B2 (en) | 2004-02-13 | 2014-09-30 | Oracle International Corporation | Hierarchical transport protocol stack for data transfer between enterprise servers |
US8458390B2 (en) | 2004-02-13 | 2013-06-04 | Oracle International Corporation | Methods and systems for handling inter-process and inter-module communications in servers and server clusters |
US8601053B2 (en) | 2004-02-13 | 2013-12-03 | Oracle International Corporation | Multi-chassis fabric-backplane enterprise servers |
US8743872B2 (en) | 2004-02-13 | 2014-06-03 | Oracle International Corporation | Storage traffic communication via a switch fabric in accordance with a VLAN |
US8713295B2 (en) | 2004-07-12 | 2014-04-29 | Oracle International Corporation | Fabric-backplane enterprise servers with pluggable I/O sub-system |
US7263631B2 (en) | 2004-08-13 | 2007-08-28 | Seakr Engineering, Incorporated | Soft error detection and recovery |
US20060036909A1 (en) * | 2004-08-13 | 2006-02-16 | Seakr Engineering, Incorporated | Soft error detection and recovery |
US20060184717A1 (en) * | 2005-02-17 | 2006-08-17 | Intel Corporation | Integrated circuit capable of flash memory storage management |
US20060212762A1 (en) * | 2005-03-21 | 2006-09-21 | Zimmer Vincent J | Error management topologies |
US7543179B2 (en) * | 2005-03-21 | 2009-06-02 | Intel Corporation | Error management topologies |
US20070294584A1 (en) * | 2006-04-28 | 2007-12-20 | Microsoft Corporation | Detection and isolation of data items causing computer process crashes |
US8417999B2 (en) * | 2008-09-26 | 2013-04-09 | Microsoft Corporation | Memory management techniques selectively using mitigations to reduce errors |
US20110173501A1 (en) * | 2008-09-26 | 2011-07-14 | Microsoft Corporation | Memory management techniques selectively using mitigations to reduce errors |
US8645762B2 (en) * | 2010-12-08 | 2014-02-04 | Advanced Micro Devices, Inc. | Queue freeze on protocol error |
US20120151251A1 (en) * | 2010-12-08 | 2012-06-14 | Advanced Micro Devices, Inc. | Queue freeze on protocol error |
US9772894B2 (en) * | 2016-01-29 | 2017-09-26 | Netapp, Inc. | Systems, methods, and machine-readable media to perform state data collection |
Also Published As
Publication number | Publication date |
---|---|
US6845470B2 (en) | 2005-01-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6845470B2 (en) | Method and system to identify a memory corruption source within a multiprocessor system | |
US7805636B2 (en) | Bootable post crash analysis environment | |
JP5579354B2 (en) | Method and apparatus for storing track data cross-reference for related applications | |
US6934879B2 (en) | Method and apparatus for backing up and restoring data from nonvolatile memory | |
AU671543B2 (en) | Processor interface chip for dual-microprocessor processor system | |
US7594143B2 (en) | Analysis engine for analyzing a computer system condition | |
US8407515B2 (en) | Partition transparent memory error handling in a logically partitioned computer system with mirrored memory | |
US6883116B2 (en) | Method and apparatus for verifying hardware implementation of a processor architecture in a logically partitioned data processing system | |
US7895477B2 (en) | Resilience to memory errors with firmware assistance | |
US20060010344A1 (en) | System and method for predictive processor failure recovery | |
US6789048B2 (en) | Method, apparatus, and computer program product for deconfiguring a processor | |
US6725396B2 (en) | Identifying field replaceable units responsible for faults detected with processor timeouts utilizing IPL boot progress indicator status | |
US7953914B2 (en) | Clearing interrupts raised while performing operating system critical tasks | |
CN116483600A (en) | Memory fault processing method and computer equipment | |
US7430683B2 (en) | Method and apparatus for enabling run-time recovery of a failed platform | |
CN115576734B (en) | Multi-core heterogeneous log storage method and system | |
US8195981B2 (en) | Memory metadata used to handle memory errors without process termination | |
US6934888B2 (en) | Method and apparatus for enhancing input/output error analysis in hardware sub-systems | |
US7139954B1 (en) | Method and apparatus for testing a computing device with memory using test program code | |
JP2016076152A (en) | Error detection system, error detection method, and error detection program | |
AU706450B2 (en) | A processor interface circuit | |
WO2015015591A1 (en) | Software debugging method, information processing device, and program | |
JPS60207937A (en) | Data processor | |
JPS63159952A (en) | Diagnosis system for directory checking address comparator | |
JPH11282634A (en) | Duplexed device i/o controlling system and its program record medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AUSTEN, CHRISTOPHER HARRY;LEE, VAN HOA;MILLER II, MILTON DEVON;AND OTHERS;REEL/FRAME:012694/0403;SIGNING DATES FROM 20020218 TO 20020222 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees | ||
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20130118 |