US8024542B1 - Allocating background workflows in a data storage system using historical data - Google Patents
Allocating background workflows in a data storage system using historical data Download PDFInfo
- Publication number
- US8024542B1 US8024542B1 US12/101,251 US10125108A US8024542B1 US 8024542 B1 US8024542 B1 US 8024542B1 US 10125108 A US10125108 A US 10125108A US 8024542 B1 US8024542 B1 US 8024542B1
- Authority
- US
- United States
- Prior art keywords
- load
- history
- data
- periodic samples
- computer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/008—Reliability or availability analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5083—Techniques for rebalancing the load in a distributed system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2206/00—Indexing scheme related to dedicated interfaces for computers
- G06F2206/10—Indexing scheme related to storage interfaces for computers, indexing schema related to group G06F3/06
- G06F2206/1012—Load balancing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0653—Monitoring storage devices or systems
Definitions
- a virtualized cluster is a cluster of different storage nodes that together expose a single storage device. Input/output (I/O) operations sent to the cluster are internally re-routed to read and write data to the appropriate locations.
- I/O Input/output
- a virtualized cluster of storage nodes can be considered analogous to collection of disks in a Redundant Array of Inexpensive Disks (RAID) configuration, since a virtualized cluster hides the internal details of the cluster's operation from initiators and presents a unified device instead.
- RAID Redundant Array of Inexpensive Disks
- a virtualized cluster can field I/Os from one or more initiators.
- the system internally processes internal operations including I/O operations.
- I/O operations can be processed in the background to provide functionality such as data migration, adaptive network compression, boot data analysis, volume mirroring, defragmentation, or other internal maintenance operations.
- These background operations can compete with initiator I/O operations for resources such as disk access, CPU cycles, memory, network bandwidth, and so on. If the background operations are scheduled during a period of very high I/O activity from initiators, a drop in performance may be experienced as decreased throughput and increased latency of the storage system.
- storage systems do not generally run at full load all the time. Instead, the load on a storage system can fluctuate with usage patterns of end users and the applications that store and access data in the storage system. More often than not, these patterns conform to human patterns, with the load on data systems often peaking during working hours, and having lulls during nights and weekends. Additionally, there may be patterns that recur on a weekly basis, such as a higher load every Friday, or on a monthly or yearly basis.
- Predicting the amount of load that is expected on the system during various time periods in the future can provide for intelligent scheduling of background I/O operations. Improved allocation of background operations may reduce system overload and also may reduce the impact on production performance. For example, background I/O operations may be intelligently scheduled to coincide with periods of otherwise low system load.
- the background workflows may be automated maintenance tasks that perform adaptive network compression, boot data analysis, data migration, mirroring, or other tasks that would otherwise compete with normal I/O operations for various storage system resources.
- Historical load statistics for the system can be collected and maintained.
- the historical load data can be analyzed to estimate preferred times in the future for scheduling background workflows.
- a number of averages can be calculated from the statistics. Such averages can correspond to the hourly average load at different natural periods of system utilization. For example, daily, weekly, or other periodicities.
- a weighted sum of these averages can be established for each hour of the day. The weighting can be imposed such that more weight is given to more recent data.
- the weighted sums may be sorted in ascending order so that the hour of the day with the lowest estimated load is given the maximum priority when allocating future background I/O operations.
- the expected I/O load on a data storage system can be estimated for the next hour, or any given hour.
- a module can operate periodically to collect and maintain historical load data about the data access patterns of the system.
- An hourly average and a weighted weekly average can both be calculated from the historical load data. These two averages can be used to estimate the load for the next hour.
- a background workflow quota can be provided for the next hour.
- This quota can represent the maximum number of background operations that can be performed with reduced impact on primary I/O operations, such as initiator data access.
- a peak load can be detected for the system. Such a peak load can serve as a proxy for the practical maximum of operations supported by the system.
- the data access load for the next hour can be estimated.
- a permitted background load can be calculated by subtracting the estimated load from the peak load. When the permitted load is determined in capacity per time units it can be converted to a number of migrations per hour by factoring in the size of a territory and the fact that a migration requires a double access to both read and write the data to be migrated.
- One example of allocating background operations is to specify a quota, or maximum, of territories that can be migrated. This quota can be provided to a module that migrates territories of data around the storage system in attempts to optimally position the data within the storage system.
- reducing the likelihood of background operations occurring during periods of high initiator load on the data storage system may improve the average I/O performance of the system.
- the load on the underlying storage devices may be leveled or more evenly distributed over time. Such load balancing over time may contribute to longer average disk life.
- FIG. 1 is a network architecture diagram illustrating aspects of a storage system that includes several virtualized clusters according to one exemplary embodiment
- FIG. 2 is a statistical plot illustrating data access load statistics over time within a data storage system according to one exemplary embodiment
- FIG. 3 is a data structure diagram illustrating a set of load statistics and the generation of daily and weekly load averages in a data storage system according to one exemplary embodiment
- FIG. 4 is a data structure diagram illustrating a set of load statistics for a data storage system and the generation of weighted average loads over multiple weeks according to one exemplary embodiment
- FIG. 5 is a logical flow diagram illustrating a process performed by a data storage system for determining a background load permitted for a given period according to one exemplary embodiment
- FIG. 6 is a logical flow diagram illustrating a process performed by a data storage system for determining the expected load for a given period according to one exemplary embodiment
- FIG. 7 is a computer architecture diagram illustrating a computer hardware architecture for a computing system capable of serving as a storage node according to one exemplary embodiment.
- program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types.
- program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types.
- program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types.
- the subject matter described herein may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, and the like.
- FIG. 1 is a network architecture diagram showing aspects of a storage system 100 that includes several virtualized clusters 5 A- 5 B.
- a virtualized cluster is a cluster of different storage nodes that together expose a single storage device.
- FIG. 1 is a network architecture diagram showing aspects of a storage system 100 that includes several virtualized clusters 5 A- 5 B.
- a virtualized cluster is a cluster of different storage nodes that together expose a single storage device.
- FIG. 1 is a network architecture diagram showing aspects of a storage system 100 that includes several virtualized clusters 5 A- 5 B.
- a virtualized cluster is a cluster of different storage nodes that together expose a single storage device.
- the clusters 5 A- 5 B include the storage server computers 2 A- 2 G (also referred to herein as “storage nodes” or a “node”) that are operative to read and write data to one or more mass storage devices, such as hard disk drives.
- the cluster 5 A includes the nodes 2 A- 2 D and the cluster 5 B includes the nodes 2 E- 2 G. All of the nodes 2 in a cluster 5 can be physically housed in the same rack, located in the same building, or distributed over geographically diverse locations, such as various buildings, cities, or countries.
- the nodes within a cluster may be housed in a one rack space unit storing up to four hard disk drives.
- the node 2 A is a one rack space computing system that includes four hard disk drives 4 A- 4 D.
- each node may be housed in a three rack space unit storing up to fifteen hard disk drives.
- the node 2 E includes fourteen hard disk drives 4 A- 4 N.
- Other types of enclosures may also be utilized that occupy more or fewer rack units and that store fewer or more hard disk drives.
- the type of storage enclosure and number of hard disk drives utilized is not generally significant to the implementation of the embodiments described herein. Any type of storage enclosure and virtually any number of hard disk devices or other types of mass storage devices may be utilized.
- multiple storage nodes may be configured together as a virtualized storage cluster.
- the nodes 2 A- 2 D have been configured as a storage cluster 5 A and the nodes 2 E- 2 G have been configured as a storage cluster 5 B.
- each of the storage nodes 2 A- 2 G is utilized to field I/O operations independently, but are exposed to the initiator of the I/O operation as a single device.
- a storage cluster may include any number of storage nodes.
- a virtualized cluster in which each node contains an independent processing unit, and in which each node can field I/Os independently (and route them according to the cluster layout) is called a horizontally virtualized or peer cluster.
- a cluster in which each node provides storage but the processing and mapping is done completely or primarily in a single node, is called a vertically virtualized cluster.
- Data may be striped across the nodes of each storage cluster.
- the cluster 5 A may stripe data across the storage nodes 2 A, 2 B, 2 C, and 2 D.
- the cluster 5 B may similarly stripe data across the storage nodes 2 E, 2 F, and 2 G. Striping data across nodes generally ensures that different I/O operations are fielded by different nodes, thereby utilizing all of the nodes simultaneously, and that the same I/O operation is not split between multiple nodes. Striping the data in this manner provides a boost to random I/O performance without decreasing sequential I/O performance.
- each storage server computer 2 A- 2 G includes one or more network ports operatively connected to a network switch 6 using appropriate network cabling. It should be appreciated that, according to embodiments of the invention, Ethernet or Gigabit Ethernet may be utilized. However, it should also be appreciated that other types of suitable physical connections may be utilized to form a network of which each storage server computer 2 A- 2 G is a part. Through the use of the network ports and other appropriate network cabling and equipment, each node within a cluster is communicatively connected to the other nodes within the cluster. Many different types and number of connections may be made between the nodes of each cluster. Furthermore, each of the storage server computers 2 A- 2 G need not be connected to the same switch 6 .
- the storage server computers 2 A- 2 G can be interconnected by any type of network or communication links, such as a LAN, a WAN, a MAN, a fiber ring, a fiber star, wireless, optical, satellite, or any other network technology, topology, protocol, or combination thereof.
- Each cluster 5 A- 5 B is also connected to a network switch 6 .
- the network switch 6 is connected to one or more client computers 8 A- 8 N (also referred to herein as “initiators”). It should be appreciated that other types of networking topologies may be utilized to interconnect the clients and the clusters 5 A- 5 B. It should also be appreciated that the initiators 8 A- 8 N may be connected to the same local area network (“LAN”) as the clusters 5 A- 5 B or may be connected to the clusters 5 A- 5 B via a distributed wide area network, such as the Internet. An appropriate protocol, such as the Internet Small Computer Systems Interface (“iSCSI”) protocol may be utilized to enable the initiators 8 A- 8 D to communicate with and utilize the various functions of the storage clusters 5 A- 5 B over a wide area network such as the Internet.
- LAN local area network
- iSCSI Internet Small Computer Systems Interface
- Two or more disks 4 within each cluster 5 A- 5 B or across clusters 5 A- 5 B may be mirrored for data redundancy and protection against failure of one, or more, of the disks 4 .
- the disks 4 may include hard drives, spinning disks, stationary media, non-volatile memories, or optically scanned media; each, or in combination, employing magnetic, capacitive, optical, semiconductor, electrical, quantum, dynamic, static, or any other data storage technology.
- the disks 4 may use IDE, ATA, SATA, PATA, SCSI, USB, PCI, Firewire, or any other bus, link, connection, protocol, network, controller, or combination thereof for I/O transfers.
- a statistical plot illustrates data access load statistics over time within a data storage system according to one exemplary embodiment.
- the horizontal axis of the plot 200 represents time and the vertical axis represents the load.
- the load can encompass any combination of various parameters such as CPU utilization, memory utilization, network bandwidth load, or I/O load.
- the best time to process background operations may be when the load on the system is minimal.
- the various background modules or plug-ins can be allocated system resources during a low period, or trough 220 , in the load as opposed to during the time of a peak 210 in the load.
- the load on the system may be a direct function of the load in the past. That is, the load for a future period can be predicted, in a statistical sense, by examining loads from the past. More specifically, the system load may be a superposition of various periodic loads. These periodic loads may correspond to a particular human-centric periodicity. For example, over the course of a day, roughly periodic troughs and peaks in the load may be observed depending, for example, on when employees come to work and when they leave. Over the course of a week, periods of relaxed load such as nights and weekend may be observed. Over the course of a year, certain days may be holidays year after year and display low loading. In this manner, an expected system load can be predicted, or estimated, from past loads.
- the total work capacity of the system can be estimated. For example, the highest recent peak, or a local maxima 230 , in the load can serve as a proxy for the total work capacity of the system. Subtracting the expected load from this peak capacity can predict the available system capacity during some period in the future for which the expected load was predicted.
- a data structure diagram 300 illustrates a set of load statistics and the generation of daily and weekly load averages in a data storage system according to one exemplary embodiment.
- the data structure 300 can store an hourly load average for every hour of the day for a week.
- the data structure 300 can be visualized as a table. In FIG. 3 , the hours of the day, in 24-hour format, are shown along the top of the table 300 and the days of the week are shown down the side of the table 300 .
- Data related to the table 300 can be collected by a background operation. For example, a periodic collection thread can store system load data from the for every hour. Of course, data collection, or sampling, periods other than one hour may also be used.
- Adding up all of the entries across a horizontal row of the table 300 and dividing by 24 can provide an hourly average load for that day.
- the third row of the table 300 can be averaged to calculate an hourly average 320 for Tuesday of that week.
- adding up all of the entries down a vertical column of the table 300 and dividing by seven can provide an hourly average load for a given hour of each day over the course of a week.
- the ninth column of the table 300 can be averaged to calculate an 8 AM average 310 over one week.
- the hourly average 320 for Tuesday and the 8 AM average 310 over one week can both be averaged together to establish a predicted load value for a future Tuesday at 8 AM.
- the load data that is collected for every hour can be maintained on multiple granularities.
- the data collected every hour can be maintained in its entirety for the last X days and then daily averages can be maintained for the Y days prior to that.
- X may be 100 and Y may be 1000.
- hourly data for the last 100 days can be maintained along with 1000 days worth of daily averages for the period ranging from 100 to 1100 days ago.
- Other granularities of maintained data, or number of different granularities, or quantities of maintained data for each granularity may also be used.
- a data structure diagram 400 illustrates a set of load statistics for a data storage system and the generation of weighted average loads over multiple weeks according to one exemplary embodiment.
- the data structure 400 can include four sets of the data structure 300 , each of which can store an hourly load average for every hour of the day for one week. In other words, the data structure 400 can store hourly load data for four weeks.
- the data structure 400 can be visualized as four layers 300 A- 300 D. Each one of the four layers 300 A- 300 D can store one week worth of hourly load data. Each one of the four layers 300 A- 300 D can be analyzed as discussed with respect to FIG. 3 .
- each layer 300 A- 300 D can yield an hourly average 320 for Tuesday, an 8 AM average 310 over one week, or a predicted load value for a future Tuesday at 8 AM.
- various other times, days, or other combinations can be averaged and estimated from each layer 300 A- 300 D as well.
- the data averaged or predicted from each of the four layers 300 A- 300 D can also be averaged together to obtain predictions from a larger data set.
- the value from each layer 300 A- 300 D can also be weighted before averaging to yield a four week weighted average 410 .
- the weights can add up to one so that the magnitude of the prediction remains to scale.
- Such a weighting can put more emphasis on recent data and may be useful for adaptive load prediction in a storage system undergoing changes in system load dynamics over time. Changes affecting the present may have reached more effectiveness in recent load histories than in distant past histories. Thus, the recent data may better predict the present, or near future, than would older data in a changing system.
- One example of a system undergoing rapidly changing load patterns may be a new system coming on line.
- Much of the system activity may be generated by the formatting of disks, installing of software, and so forth. These activities may gradually decrease into the steady state load dynamics for the system.
- Other weightings can be used as well. While the weights may be established in linear step fashion as in the example above, the weights may also be established based on examining fluctuations in the load data, or an adaptive system may vary the weights over time.
- FIG. 5 is a flow diagram showing a routine 500 that illustrates aspects of an exemplary process performed by a storage system for determining a background load permitted for a given period.
- routine 500 that illustrates aspects of an exemplary process performed by a storage system for determining a background load permitted for a given period.
- the logical operations described herein are implemented (1) as a sequence of computer implemented acts or program modules running on a computing system and/or (2) as interconnected machine logic circuits or circuit modules within the computing system. The implementation is a matter of choice dependent on the performance and other requirements of the computing system. Accordingly, the logical operations described herein are referred to variously as operations, structural devices, acts, or modules.
- the routine 500 can be part of a workflow module that determines when to perform certain background tasks. For example, data migration, restriping, defragmentation, and so forth.
- the determination of permitted background load can be made periodically. For example, the permitted background load may be determined once per hour to allocate resources to background threads or modules for the next hour. Other time period granularities may be used as well.
- the routine 500 can begin with operation 510 where the peak load of the system can be established.
- the historical load data can be examined to find the peak load, or local maxima, during a time period. For example, the time period may be over the last week.
- This actual peak load can serve as a proxy for the maximum load of the system. Since the peak can be determined from real production loads, it may be a more reasonable estimate of the maximum allowable load given the particular kind of load that the system is subjected to. For example, when considering I/O specific loads, the typical mix of random access versus sequential access I/O operations.
- the peak load in terms of I/O loading, may be determined as a data rate such as MB/s (megabytes per second) or as a more aggregated rate such as sectors per second. Other types of system loading may be addressed in terms of the appropriate units, such as instruction counts, machine cycles, memory consumption, other similar quantities, or any combinations thereof.
- the expected load can be determined.
- the load can be estimated from examining historical load data, an example of which is illustrated as a subroutine 600 . Further details of estimating the load are discussed with respect to FIG. 6 below.
- the permitted load is determined by subtracting the expected load from the peak load. That is, the amount of permitted background load is determined from the maximum load (as given by the local maxima load) less the expected load as determined from historical load data.
- the permitted background load can be subjected to predetermined minimum and maximum permitted load limits. This can insure that there are always at least a minimum amount of system resources dedicated to the background operations and also that the system is never overly dedicated to performing only background operations. The imposing of minimum and maximum permitted background loads is optional.
- the permitted load can be translated from the original MB/s or sectors per second into territories per hour.
- This permitted load in territories per hour can be provided, for example, to a data migration module as a maximum number of territories to be migrated over the next hour. This may be referred to as the data migration quota.
- the load data is maintained on a per-volume basis, the total permitted load may be determined for all volumes in a given container or in the storage system.
- the routine 500 may terminate after operation 540 .
- a storage resource management (SRM) module can operate periodically, for example every hour, to manage the allocation of background workflow to various other modules or plug-ins.
- data migration modules may be supplied with quotas offering each module an opportunity to move a set number of territories. If a module is unable to make use of its allocated number of data migrations, the quota of other modules or plug-ins may be increased.
- territory quotas and I/O loads are discussed herein, it should be appreciated that data migration is only one example of a module performing background operations.
- workflows of various other background threads or plug-in modules can be allocated using the technology discussed herein, as the permitted loads that are calculated can related to not only I/O loads, but also system loads on memory, CPU cycles, network bandwidth, caching, and so forth.
- the type of system loading to be considered may vary based on the system demands of the respective background operation for which resources are being allocated.
- FIG. 6 is a flow diagram illustrating a routine 600 that shows aspects of an exemplary process performed by a storage system for using historical load data to estimate a load on the storage system for a future period.
- the routine 600 can begin with operation 610 where historical load data is collected.
- the load data can be collected on a fixed periodic cycle. For example, every hour.
- the load data can also be maintained on multiple granularities. For example, the data collected every hour can be maintained in its entirety for the last 100 days while 1000 days worth of daily averages may be maintained from periods prior to the last 100 days. Other granularities of maintained data or quantities of maintained data for each granularity may also be used.
- an hourly average can be calculated.
- the hourly average can be the average for a particular hour every day of the last week. For example, if it is currently 9 AM and the load for 10 AM is being estimated, then the 10 AM average for the last week can be calculated.
- a weighted weekly average can be calculated.
- the weekly average can be the average for that particular hour and that particular day of the week, for every week of the last several weeks. For example, four weeks may be used.
- the weekly average can be a weighted average where the values to be averaged may first be multiplied by weights. The weighting can be applied as to give more weight to recent load data and less weight to older load data.
- the hourly average and the weighted weekly average can be averaged together to provide a predicted load for the given hour and day of the week specified.
- the load may be predicted for the next hour so that system resources may be scheduled for background operations based on the expected load for the next hour.
- FIG. 7 and the following discussion are intended to provide a brief, general description of a suitable computing environment in which the embodiments described herein may be implemented. While the technical details are presented herein in the general context of program modules that execute in conjunction with the execution of an operating system, those skilled in the art will recognize that the embodiments may also be implemented in combination with other program modules.
- program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types.
- program modules may be located in both local and remote memory storage devices.
- FIG. 7 shows an illustrative computer architecture for a storage node computer 2 that may be utilized in the implementations described herein.
- the storage node computer 2 includes a baseboard, or “motherboard”, which is a printed circuit board to which a multitude of components or devices may be connected by way of a system bus or other electrical communication paths.
- a CPU 22 operates in conjunction with a chipset 52 .
- the CPU 22 is a standard central processor that performs arithmetic and logical operations necessary for the operation of the computer.
- the storage node computer 2 may include a multitude of CPUs 22 .
- the chipset 52 includes a north bridge 24 and a south bridge 26 .
- the north bridge 24 provides an interface between the CPU 22 and the remainder of the computer 2 .
- the north bridge 24 also provides an interface to a random access memory (“RAM”) used as the main memory 54 in the computer 2 and, possibly, to an on-board graphics adapter 30 .
- the north bridge 24 may also include functionality for providing networking functionality through a gigabit Ethernet adapter 28 .
- the gigabit Ethernet adapter 28 is capable of connecting the computer 2 to another computer via a network. Connections which may be made by the network adapter 28 may include LAN or WAN connections. LAN and WAN networking environments are commonplace in offices, enterprise-wide computer networks, intranets, and the internet.
- the north bridge 24 is connected to the south bridge 26 .
- the south bridge 26 is responsible for controlling many of the input/output functions of the computer 2 .
- the south bridge 26 may provide one or more universal serial bus (“USB”) ports 32 , a sound adapter 46 , an Ethernet controller 60 , and one or more general purpose input/output (“GPIO”) pins 34 .
- the south bridge 26 may also provide a bus for interfacing peripheral card devices such as a graphics adapter 62 .
- the bus comprises a peripheral component interconnect (“PCI”) bus.
- PCI peripheral component interconnect
- the south bridge 26 may also provide a system management bus 64 for use in managing the various components of the computer 2 . Additional details regarding the operation of the system management bus 64 and its connected components are provided below.
- the south bridge 26 is also operative to provide one or more interfaces for connecting mass storage devices to the computer 2 .
- the south bridge 26 includes a serial advanced technology attachment (“SATA”) adapter for providing one or more serial ATA ports 36 and an ATA 100 adapter for providing one or more ATA 100 ports 44 .
- the serial ATA ports 36 and the ATA 100 ports 44 may be, in turn, connected to one or more mass storage devices storing an operating system 40 and application programs, such as the SATA disk drive 38 .
- an operating system 40 comprises a set of programs that control operations of a computer and allocation of resources.
- An application program is software that runs on top of the operating system software, or other runtime environment, and uses computer resources to perform application specific tasks desired by the user.
- the operating system 40 comprises the LINUX operating system. According to another embodiment of the invention the operating system 40 comprises the WINDOWS SERVER operating system from MICROSOFT CORPORATION. According to another embodiment, the operating system 40 comprises the UNIX or SOLARIS operating system. It should be appreciated that other operating systems may also be utilized.
- the mass storage devices connected to the south bridge 26 , and their associated computer-readable media, provide non-volatile storage for the computer 2 .
- computer-readable media can be any available media that can be accessed by the computer 2 .
- computer-readable media may comprise computer storage media and communication media.
- Computer storage media includes volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data.
- Computer storage media includes, but is not limited to, RAM, ROM, EPROM, EEPROM, flash memory or other solid state memory technology, CD-ROM, DVD, HD-DVD, BLU-RAY, or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computer.
- a low pin count (“LPC”) interface may also be provided by the south bridge 6 for connecting a “Super I/O” device 70 .
- the Super I/O device 70 is responsible for providing a number of input/output ports, including a keyboard port, a mouse port, a serial interface 72 , a parallel port, and other types of input/output ports.
- the LPC interface may also connect a computer storage media such as a ROM or a flash memory such as a NVRAM 48 for storing the firmware 50 that includes program code containing the basic routines that help to start up the computer 2 and to transfer information between elements within the computer 2 .
- the south bridge 26 may include a system management bus 64 .
- the system management bus 64 may include a BMC 66 .
- the BMC 66 is a microcontroller that monitors operation of the computer system 2 .
- the BMC 66 monitors health-related aspects associated with the computer system 2 , such as, but not limited to, the temperature of one or more components of the computer system 2 , speed of rotational components (e.g., spindle motor, CPU Fan, etc.) within the system, the voltage across or applied to one or more components within the system 2 , and the available or used capacity of memory devices within the system 2 .
- speed of rotational components e.g., spindle motor, CPU Fan, etc.
- the BMC 66 is communicatively connected to one or more components by way of the management bus 64 .
- these components include sensor devices for measuring various operating and performance-related parameters within the computer system 2 .
- the sensor devices may be either hardware or software based components configured or programmed to measure or detect one or more of the various operating and performance-related parameters.
- the BMC 66 functions as the master on the management bus 64 in most circumstances, but may also function as either a master or a slave in other circumstances.
- Each of the various components communicatively connected to the BMC 66 by way of the management bus 64 is addressed using a slave address.
- the management bus 64 is used by the BMC 66 to request and/or receive various operating and performance-related parameters from one or more components, which are also communicatively connected to the management bus 64 .
- the computer 2 may comprise other types of computing devices, including hand-held computers, embedded computer systems, personal digital assistants, and other types of computing devices known to those skilled in the art. It is also contemplated that the computer 2 may not include all of the components shown in FIG. 7 , may include other components that are not explicitly shown in FIG. 7 , or may utilize an architecture completely different than that shown in FIG. 7 .
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Quality & Reliability (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Debugging And Monitoring (AREA)
Abstract
Description
Claims (17)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/101,251 US8024542B1 (en) | 2007-04-13 | 2008-04-11 | Allocating background workflows in a data storage system using historical data |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US92322507P | 2007-04-13 | 2007-04-13 | |
US12/101,251 US8024542B1 (en) | 2007-04-13 | 2008-04-11 | Allocating background workflows in a data storage system using historical data |
Publications (1)
Publication Number | Publication Date |
---|---|
US8024542B1 true US8024542B1 (en) | 2011-09-20 |
Family
ID=44587168
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/101,251 Active 2030-05-04 US8024542B1 (en) | 2007-04-13 | 2008-04-11 | Allocating background workflows in a data storage system using historical data |
Country Status (1)
Country | Link |
---|---|
US (1) | US8024542B1 (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9141508B2 (en) * | 2010-12-21 | 2015-09-22 | Oracle International Corporation | Assigning read requests based on busyness of devices |
US9519438B1 (en) | 2007-04-13 | 2016-12-13 | American Megatrends, Inc. | Data migration between multiple tiers in a storage system using age and frequency statistics |
US9870296B1 (en) * | 2011-06-17 | 2018-01-16 | Mark A. Parenti | Evaluating system performance |
US10055170B2 (en) * | 2015-04-30 | 2018-08-21 | International Business Machines Corporation | Scheduling storage unit maintenance tasks in a dispersed storage network |
US20190220217A1 (en) * | 2018-01-12 | 2019-07-18 | International Business Machines Corporation | Automated predictive tiered storage system |
US10528295B2 (en) | 2017-04-29 | 2020-01-07 | International Business Machines Corporation | Intra-tier data migration to segregate zombie storage from dead storage |
US10705853B2 (en) | 2008-05-06 | 2020-07-07 | Amzetta Technologies, Llc | Methods, systems, and computer-readable media for boot acceleration in a data storage system by consolidating client-specific boot data in a consolidated boot volume |
US10810229B2 (en) | 2018-10-19 | 2020-10-20 | Oracle International Corporation | Database replication based on data access scores |
CN113110805A (en) * | 2021-05-06 | 2021-07-13 | 杭州海康威视数字技术股份有限公司 | Disk array internal data interaction method and device and electronic equipment |
CN113791909A (en) * | 2021-09-18 | 2021-12-14 | 平安普惠企业管理有限公司 | Server capacity adjusting method and device, computer equipment and storage medium |
Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4942579A (en) | 1987-06-02 | 1990-07-17 | Cab-Tek, Inc. | High-speed, high-capacity, fault-tolerant error-correcting storage system |
US5257367A (en) | 1987-06-02 | 1993-10-26 | Cab-Tek, Inc. | Data storage system with asynchronous host operating system communication link |
US5502836A (en) | 1991-11-21 | 1996-03-26 | Ast Research, Inc. | Method for disk restriping during system operation |
US5720027A (en) | 1996-05-21 | 1998-02-17 | Storage Computer Corporation | Redundant disc computer having targeted data broadcast |
US5732238A (en) | 1996-06-12 | 1998-03-24 | Storage Computer Corporation | Non-volatile cache for providing data integrity in operation with a volatile demand paging cache in a data storage system |
US5790774A (en) | 1996-05-21 | 1998-08-04 | Storage Computer Corporation | Data storage system with dedicated allocation of parity storage and parity reads and writes only on operations requiring parity information |
US5893919A (en) | 1996-09-27 | 1999-04-13 | Storage Computer Corporation | Apparatus and method for storing data with selectable data protection using mirroring and selectable parity inhibition |
US6098128A (en) | 1995-09-18 | 2000-08-01 | Cyberstorage Systems Corporation | Universal storage management system |
US6275898B1 (en) | 1999-05-13 | 2001-08-14 | Lsi Logic Corporation | Methods and structure for RAID level migration within a logical unit |
US20050055402A1 (en) | 2003-09-09 | 2005-03-10 | Eiichi Sato | File sharing device and inter-file sharing device data migration method |
US20050165925A1 (en) * | 2004-01-22 | 2005-07-28 | International Business Machines Corporation | System and method for supporting transaction and parallel services across multiple domains based on service level agreenments |
US20050188075A1 (en) * | 2004-01-22 | 2005-08-25 | International Business Machines Corporation | System and method for supporting transaction and parallel services in a clustered system based on a service level agreement |
US20060031649A1 (en) | 2000-05-24 | 2006-02-09 | Hitachi, Ltd. | Data storage system and method of hierarchical control thereof |
US20060031648A1 (en) * | 2004-08-09 | 2006-02-09 | Atsushi Ishikawa | Storage device |
US20060130042A1 (en) * | 2004-12-15 | 2006-06-15 | Dias Daniel M | Method and apparatus for dynamic application upgrade in cluster and grid systems for supporting service level agreements |
US20060248273A1 (en) | 2005-04-29 | 2006-11-02 | Network Appliance, Inc. | Data allocation within a storage system architecture |
US20060243056A1 (en) * | 2005-04-28 | 2006-11-02 | Sundermeyer Jeffry N | Systems and methods for maintaining load histories |
US7360051B2 (en) | 2004-09-10 | 2008-04-15 | Hitachi, Ltd. | Storage apparatus and method for relocating volumes thereof |
US20080104343A1 (en) | 2006-10-30 | 2008-05-01 | Hitachi, Ltd. | Storage control device and data migration method for storage control device |
US7404102B2 (en) | 2003-08-14 | 2008-07-22 | Compellent Technologies | Virtual disk drive system and method |
US20080320247A1 (en) | 2005-05-12 | 2008-12-25 | Morfey Alistair G | Processor and interface |
US7730531B2 (en) * | 2005-04-15 | 2010-06-01 | Microsoft Corporation | System and method for detection of artificially generated system load |
-
2008
- 2008-04-11 US US12/101,251 patent/US8024542B1/en active Active
Patent Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5257367A (en) | 1987-06-02 | 1993-10-26 | Cab-Tek, Inc. | Data storage system with asynchronous host operating system communication link |
US4942579A (en) | 1987-06-02 | 1990-07-17 | Cab-Tek, Inc. | High-speed, high-capacity, fault-tolerant error-correcting storage system |
US5502836A (en) | 1991-11-21 | 1996-03-26 | Ast Research, Inc. | Method for disk restriping during system operation |
US6098128A (en) | 1995-09-18 | 2000-08-01 | Cyberstorage Systems Corporation | Universal storage management system |
US5720027A (en) | 1996-05-21 | 1998-02-17 | Storage Computer Corporation | Redundant disc computer having targeted data broadcast |
US5790774A (en) | 1996-05-21 | 1998-08-04 | Storage Computer Corporation | Data storage system with dedicated allocation of parity storage and parity reads and writes only on operations requiring parity information |
US5732238A (en) | 1996-06-12 | 1998-03-24 | Storage Computer Corporation | Non-volatile cache for providing data integrity in operation with a volatile demand paging cache in a data storage system |
US5893919A (en) | 1996-09-27 | 1999-04-13 | Storage Computer Corporation | Apparatus and method for storing data with selectable data protection using mirroring and selectable parity inhibition |
US6275898B1 (en) | 1999-05-13 | 2001-08-14 | Lsi Logic Corporation | Methods and structure for RAID level migration within a logical unit |
US20060031649A1 (en) | 2000-05-24 | 2006-02-09 | Hitachi, Ltd. | Data storage system and method of hierarchical control thereof |
US7404102B2 (en) | 2003-08-14 | 2008-07-22 | Compellent Technologies | Virtual disk drive system and method |
US20050055402A1 (en) | 2003-09-09 | 2005-03-10 | Eiichi Sato | File sharing device and inter-file sharing device data migration method |
US20050165925A1 (en) * | 2004-01-22 | 2005-07-28 | International Business Machines Corporation | System and method for supporting transaction and parallel services across multiple domains based on service level agreenments |
US20050188075A1 (en) * | 2004-01-22 | 2005-08-25 | International Business Machines Corporation | System and method for supporting transaction and parallel services in a clustered system based on a service level agreement |
US20060031648A1 (en) * | 2004-08-09 | 2006-02-09 | Atsushi Ishikawa | Storage device |
US7360051B2 (en) | 2004-09-10 | 2008-04-15 | Hitachi, Ltd. | Storage apparatus and method for relocating volumes thereof |
US20060130042A1 (en) * | 2004-12-15 | 2006-06-15 | Dias Daniel M | Method and apparatus for dynamic application upgrade in cluster and grid systems for supporting service level agreements |
US7730531B2 (en) * | 2005-04-15 | 2010-06-01 | Microsoft Corporation | System and method for detection of artificially generated system load |
US20060243056A1 (en) * | 2005-04-28 | 2006-11-02 | Sundermeyer Jeffry N | Systems and methods for maintaining load histories |
US20060248273A1 (en) | 2005-04-29 | 2006-11-02 | Network Appliance, Inc. | Data allocation within a storage system architecture |
US20080320247A1 (en) | 2005-05-12 | 2008-12-25 | Morfey Alistair G | Processor and interface |
US20080104343A1 (en) | 2006-10-30 | 2008-05-01 | Hitachi, Ltd. | Storage control device and data migration method for storage control device |
Non-Patent Citations (7)
Title |
---|
Burtscher, Martin, et al., "Prediction Outcome History-based Confidence Estimation for Load Value Prediction," Department of Computer Science, University of Colorado., Journal of Instruction-Level Parallelism 1, published May 1999, p. 1-25. * |
U.S. Appl. No. 12/101,236, filed Apr. 11, 2008, entitled "Data Migration Between Multiple Tiers in a Storage System Using Age and Frequency Statistics," Inventors: Chatterjee et al. |
U.S. Appl. No. 12/101,238, filed Apr. 11, 2008, entitled "Data Migration Between Multiple Tiers in a Storage System Using Pivot Tables," Inventors: Chatterjee et al. |
U.S. Appl. No. 12/101,241, filed Apr. 11, 2008, entitled "Allocating Background Workflows in a Data Storage System Using Autocorrelation," Inventors: Chatterjee et al. |
U.S. Official Action dated Dec. 2, 2010 in U.S. Appl. No. 12/104,123. |
U.S. Official Action dated Dec. 22, 2010 in U.S. Appl. No. 12/101,236. |
U.S. Official Action dated Oct. 28, 2010 in U.S. Appl. No. 12/104,135. |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9519438B1 (en) | 2007-04-13 | 2016-12-13 | American Megatrends, Inc. | Data migration between multiple tiers in a storage system using age and frequency statistics |
US10705853B2 (en) | 2008-05-06 | 2020-07-07 | Amzetta Technologies, Llc | Methods, systems, and computer-readable media for boot acceleration in a data storage system by consolidating client-specific boot data in a consolidated boot volume |
US9141508B2 (en) * | 2010-12-21 | 2015-09-22 | Oracle International Corporation | Assigning read requests based on busyness of devices |
US9875040B2 (en) | 2010-12-21 | 2018-01-23 | Oracle International Corporation | Assigning read requests based on busyness of devices |
US9870296B1 (en) * | 2011-06-17 | 2018-01-16 | Mark A. Parenti | Evaluating system performance |
US10055170B2 (en) * | 2015-04-30 | 2018-08-21 | International Business Machines Corporation | Scheduling storage unit maintenance tasks in a dispersed storage network |
US10528295B2 (en) | 2017-04-29 | 2020-01-07 | International Business Machines Corporation | Intra-tier data migration to segregate zombie storage from dead storage |
US20190220217A1 (en) * | 2018-01-12 | 2019-07-18 | International Business Machines Corporation | Automated predictive tiered storage system |
US11157194B2 (en) * | 2018-01-12 | 2021-10-26 | International Business Machines Corporation | Automated predictive tiered storage system |
US10810229B2 (en) | 2018-10-19 | 2020-10-20 | Oracle International Corporation | Database replication based on data access scores |
US11243981B2 (en) | 2018-10-19 | 2022-02-08 | Oracle International Corporation | Database replication based on data access scores |
CN113110805A (en) * | 2021-05-06 | 2021-07-13 | 杭州海康威视数字技术股份有限公司 | Disk array internal data interaction method and device and electronic equipment |
CN113110805B (en) * | 2021-05-06 | 2023-06-02 | 杭州海康威视数字技术股份有限公司 | Method and device for interaction of internal data of disk array and electronic equipment |
CN113791909A (en) * | 2021-09-18 | 2021-12-14 | 平安普惠企业管理有限公司 | Server capacity adjusting method and device, computer equipment and storage medium |
CN113791909B (en) * | 2021-09-18 | 2023-10-03 | 深圳市华智世纪科技有限公司 | Method and device for adjusting server capacity, computer equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8024542B1 (en) | Allocating background workflows in a data storage system using historical data | |
US8271757B1 (en) | Container space management in a data storage system | |
US8006061B1 (en) | Data migration between multiple tiers in a storage system using pivot tables | |
US9703500B2 (en) | Reducing power consumption by migration of data within a tiered storage system | |
US9104316B2 (en) | Runtime dynamic performance skew elimination | |
EP3087503B1 (en) | Cloud compute scheduling using a heuristic contention model | |
US6912635B2 (en) | Distributing workload evenly across storage media in a storage array | |
US11256595B2 (en) | Predictive storage management system | |
US10887176B2 (en) | Predicting resource demand in computing environments | |
GB2496807B (en) | Computer system, management method of the computer system, and program | |
US8140775B1 (en) | Allocating background workflows in a data storage system using autocorrelation | |
US20120144008A1 (en) | System and Method for Analyzing Computing System Resources | |
US9953276B2 (en) | Method and system that measures and reports computational-resource usage in a data center | |
JP2021056955A (en) | Distributed storage system and data migration method | |
CN111966449B (en) | Virtual machine backup management method, system, terminal and storage medium | |
EP4258096A1 (en) | Predictive block storage size provisioning for cloud storage volumes | |
JP2021197010A (en) | Distributed storage system and rebalancing method | |
Mustyala et al. | Cost Optimization Strategies for Kubernetes Deployments in Cloud Environments | |
Kambatla et al. | Optimistic scheduling with service guarantees | |
Shetti et al. | E-VM: An elastic virtual machine scheduling algorithm to minimize the total cost of ownership in a hybrid cloud | |
US20230325125A1 (en) | Predictive block storage size provisioning for cloud storage volumes | |
Murugan et al. | Software Defined Energy Adaptation in Scale-Out Storage Systems | |
Wu et al. | POS: A popularity-based online scaling scheme for raid-structured storage systems | |
US20220147370A1 (en) | Management of computing devices | |
Kelley | Resource Allocation using Adaptive Characterization of Online, Data-Intensive Workloads |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: AMERICAN MEGATRENDS, INC., GEORGIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHATTERJEE, PARESH;NARAYANAN, AJIT;RANGANATHAN, LOGANATHAN;AND OTHERS;SIGNING DATES FROM 20080403 TO 20080408;REEL/FRAME:020788/0562 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
AS | Assignment |
Owner name: AMZETTA TECHNOLOGIES, LLC,, GEORGIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AMERICAN MEGATRENDS INTERNATIONAL, LLC,;REEL/FRAME:053007/0151 Effective date: 20190308 Owner name: AMERICAN MEGATRENDS INTERNATIONAL, LLC, GEORGIA Free format text: CHANGE OF NAME;ASSIGNOR:AMERICAN MEGATRENDS, INC.;REEL/FRAME:053007/0233 Effective date: 20190211 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2553); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 12 |