Nothing Special   »   [go: up one dir, main page]

US8392670B2 - Performance management of access to flash memory in a storage device - Google Patents

Performance management of access to flash memory in a storage device Download PDF

Info

Publication number
US8392670B2
US8392670B2 US12/742,783 US74278310A US8392670B2 US 8392670 B2 US8392670 B2 US 8392670B2 US 74278310 A US74278310 A US 74278310A US 8392670 B2 US8392670 B2 US 8392670B2
Authority
US
United States
Prior art keywords
flash
controller
storage
flash memory
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/742,783
Other versions
US20110283046A1 (en
Inventor
Hideyuki Koseki
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hitachi Ltd
Original Assignee
Hitachi Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi Ltd filed Critical Hitachi Ltd
Assigned to HITACHI, LTD. reassignment HITACHI, LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KOSEKI, HIDEYUKI
Publication of US20110283046A1 publication Critical patent/US20110283046A1/en
Priority to US13/760,144 priority Critical patent/US8924659B2/en
Application granted granted Critical
Publication of US8392670B2 publication Critical patent/US8392670B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/0223User address space allocation, e.g. contiguous or non contiguous base addressing
    • G06F12/023Free address space management
    • G06F12/0238Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory
    • G06F12/0246Memory management in non-volatile memory, e.g. resistive RAM or ferroelectric memory in block erasable memory, e.g. flash memory
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/061Improving I/O performance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0638Organizing or formatting or addressing of data
    • G06F3/0644Management of space entities, e.g. partitions, extents, pools
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0683Plurality of storage devices
    • G06F3/0688Non-volatile semiconductor memory arrays

Definitions

  • the present invention relates to a storage device.
  • a storage device is a device that has a disk for storing data and a controller for controlling the disk and is intended to provide a high-capacity data storage space to a host or a computer that is connected to the storage device.
  • the I/O processing speed can be increased and high fault tolerance against disk failures can be achieved.
  • Patent Literature 1 discloses a storage device having an HDD (Hard Disk Drive) as a storage medium.
  • HDD Hard Disk Drive
  • Flash memory has been gaining increased attention as a new storage medium. Flash memory is characterized by its faster read and write times than those of HDDs. For example, although HDDs have read and write times each on the order of several milliseconds, there has been reported a performance data example of NAND flash memory that has a read time of 25 microsecond and a write time of 300 microsecond.
  • flash memory unlike HDDs, has another characteristic that its physical area to which data has been written once cannot be overwritten. In order to rewrite data in flash memory, it is necessary to once erase the data on a “block” basis. However, flash memory has a limited number of write cycles due to the physical reasons of memory cells for storing information. Further, there has been reported data that flash memory requires an erase time of about 1 ms.
  • flash memory typically employs a method in which the correspondence relationship between the address and the cell location that is shown to a host system is retained, and when data stored in a block is to be updated, the update data is stored into an empty page, in which data is not written, and the original data stored in the block is invalidated.
  • a technique of addressing a shortage of empty pages a technique called “garbage collection (hereinafter referred to as GC)” has been laid open to the public.
  • GC garbage collection
  • WL wear leveling
  • flash memory has a very fast read time, it can be said to be a device having slow write and erase times. Further, from the aspect of control, flash memory should be controlled with GC, WL, or the like that has not been used for the conventional HDDs.
  • write if there are no sufficient empty pages for storing update data, empty pages are secured by executing GC once; thereafter, data is stored into the secured empty pages.
  • time required for the completion of the write processing corresponds to the sum of the GC execution time+the actual data write time.
  • flash memory since flash memory requires long data write and erase times, it may take several milliseconds until the completion of the write processing in some cases. This, however, could result in degraded performance down to a level of, or below the performance of HDDs.
  • FIG. 1 is a diagram illustrating the aforementioned problem with flash memory.
  • GC+write which is described as “Program” in FIG. 1
  • the storage controller must wait for a long time until the completion of the write. In this period, the storage controller would keep consuming resources of devices such as cache memory for temporally storing data, which could result in degraded performance.
  • the present invention has been made in order to solve the aforementioned problem. It is an object of the present invention to improve the performance of accessing flash memory used as a storage medium in a storage device.
  • a storage controller before accessing flash memory, queries a flash controller as to whether the flash memory is accessible.
  • a storage device of the present invention whether flash memory is accessible is queried in advance, whereby the wait time in which the flash memory is inaccessible can be reduced. Accordingly, the substantial access performance to the flash memory can be improved.
  • FIG. 1 is a diagram illustrating the aforementioned problem with flash memory.
  • FIG. 2 is a diagram showing the configuration of a storage system 10000 in accordance with Embodiment 1.
  • FIG. 3 is a structure diagram of memory 140 .
  • FIG. 4 is a structure diagram of an LU management table 1422 for managing the address spaces of LUs.
  • FIG. 5 is a structure diagram of a memory management table 1423 .
  • FIG. 6 is a structure diagram of a transfer list 1424 .
  • FIG. 7 is a structure diagram of a transfer completion list 1425 .
  • FIG. 8 is a configuration diagram of a flash box 210 .
  • FIG. 9 is a configuration diagram of a flash memory chip 231 .
  • FIG. 10 is a diagram illustrating address assignment.
  • FIG. 11 is a structure diagram of memory 224 .
  • FIG. 12 is a structure diagram of an address management table 22422 .
  • FIG. 13 is a structure diagram of a block management table 22423 .
  • FIG. 14 is a flow chart illustrating the basic I/O sequence of a storage device 10 .
  • FIG. 15 is an operation flow chart of an I/O processing program executed by a storage controller 100 .
  • FIG. 16 is a detailed flow chart of step S 150 in FIG. 15 .
  • FIG. 17 is a structure diagram of a request command 1000 and a reply command 2000 transferred between the storage controller 100 and a flash controller 220 .
  • FIG. 18 is a flow chart showing the details of step S 153 .
  • FIG. 19 is a detailed flow chart of step S 1533 .
  • FIG. 20 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 1.
  • FIG. 21 is an operation flow chart of an I/O processing program executed by the storage controller 100 in Embodiment 2.
  • FIG. 22 shows a detailed flow of step S 240 .
  • FIG. 23 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 2.
  • FIG. 24 is a flow chart showing the processing in which the storage controller 100 requests the flash controller 220 to free a block 232 .
  • FIG. 25 is a structure diagram of the memory 140 in accordance with Embodiment 4.
  • FIG. 26 is a structure diagram of a segment management table 1426 .
  • FIG. 27 is a structure diagram of an LU management table 1427 .
  • FIG. 28 is a structure diagram of a statistical information management table 1428 .
  • FIG. 29 is a flow chart showing the processing in which the storage controller 100 makes equal the load on each flash box 210 .
  • FIG. 2 is a diagram showing the configuration of a storage system 10000 in accordance with Embodiment 1 of the present invention.
  • the storage system 10000 includes a storage device 10 and a host computer 40 .
  • the host computer 40 is a host system that uses the storage device 10 , such as an application server.
  • the storage device 10 includes a storage controller 100 and a flash box unit 200 .
  • the storage controller 100 controls the operation of the storage device 10 .
  • the flash box unit 200 includes flash memory for storing data.
  • the host computer 40 is connected to the storage device 10 via an I/F chip 110 of the storage controller 100 .
  • a SAN (Storage Area Network) 20 is used as an interface for connecting the storage device 10 and the host computer 40 .
  • Fibre Channel, SCSI (Small Computer System Interface), iSCSI (internet Small Computer System Interface), USB (Universal Serial Bus), IEEE 1394 bus, SAS (Serial Attached SCSI), or the like can be used, for example.
  • More than one host computer 40 can be connected to the storage device 10 .
  • other types of interfaces can be used instead of the SAN 20 .
  • the host computer 40 has installed thereon control software that issues control commands and the like to the storage device 10 .
  • the storage controller 100 of the storage device 10 upon receipt of such control commands and the like, executes its own control software to thereby control the operation of the storage device 10 .
  • the storage controller 100 has an MC (Network Interface Card) 120 that is an interface for connecting the storage device 10 to the LAN 30 .
  • a computer that manages, controls, or maintains the storage device 10 a different computer other than the host computer 40 can be used.
  • other types of communication networks can be used instead of the LAN 30 .
  • the storage controller 100 includes ASICs (Application Specific Integrated Circuits) 150 , a CPU (Central Processing Unit) 130 , memory 140 , and an I/F chip 160 .
  • the ASICs 150 perform a parity operation and the like when RAID is constructed.
  • the CPU 130 controls the entire storage device 10 .
  • the memory 140 has stored therein computer programs and data that are necessary for the CPU 130 to control the storage device 10 .
  • the I/F chip 160 connects the storage controller 100 and the flash box unit 200 .
  • the ASICs 150 include a DMA (Direct Memory Access) transfer unit 151 .
  • the DMA transfer unit 151 transfers data in accordance with a transfer list stored in the memory 140 . It should be noted that the DMA transfer unit 151 may be incorporated in a device other than the ASICs 150 , such as the I/F chip 110 or 160 .
  • Each of the I/F chip 110 , the MC 120 , the CPU 130 , and the I/F chip 160 is connected to the ASICs 150 via a dedicated connection bus 170 such as a PCI (Peripheral Component Interconnect).
  • the memory 140 is connected to the ASICs 150 via a dedicated connection bus 190 such as a DDR3 (Double Data Rate 3).
  • the storage controller 100 accesses flash memory chips 231 via a flash controller 220 described below. It should be noted that although the storage controller 100 will be hereinafter described as accessing the flash memory chips 231 for the sake of convenience of the description, what actually accesses the flash memory chips 231 is the flash controller 220 .
  • the flash box unit 200 includes a plurality of flash boxes 210 .
  • the flash box unit 200 and the storage controller 100 are connected via the I/F chip 160 and a dedicated bus 180 such as Fibre Channel or SAS.
  • Each flash box 210 includes a plurality of flash memory chips 231 . Though not shown, the plurality of memory chips 231 constitute an RAID group.
  • FIG. 3 is a structure diagram of the memory 140 .
  • the memory 140 includes a data cache area 141 and a system area 142 .
  • the data cache area 141 is an area used to temporarily store user data.
  • the system area 142 has stored therein programs and various tables for controlling the storage device 10 .
  • the data cache area 141 is divided into a plurality of cache segments of a given size so that user data 1411 is stored on a cache segment basis.
  • the user data 1411 is, for example, data read from or written to the storage device 10 by the host computer 40 .
  • the system area 142 has stored therein a program group 1421 , an LU management table 1422 , a memory management table 1423 , a transfer list 1424 , and a transfer completion list 1425 .
  • the program group 1421 includes various programs executed by the CPU 130 .
  • the LU management table 1422 manages the address spaces of LUs (Logical Units) created in the RAID group of the flash memory 231 .
  • LU represents a unit for providing a storage area of the storage device 10 to the outside of the storage device 10 .
  • the memory management table 1423 manages the cache segments in the data cache area 141 .
  • the transfer list 1424 is used for the CPU 130 to instruct the ASICs 150 to execute DMA transfer.
  • the transfer completion list 1425 is used for the ASICs 150 to inform the CPU 130 of the completion of the DMA transfer.
  • FIG. 4 is a structure diagram of the LU management table 1422 for managing the address spaces of LUs.
  • the LU management table 1422 contains a LUN (Logical Unit Number) 14221 , flash box No. 14222 , start LBA 14223 , and last LBA 14224 .
  • LUN Logical Unit Number
  • the LUN 14221 indicates the identifier of a created LU, and a unique number is assigned to each LU.
  • the flash box No. 14222 indicates the number of the flash box 210 in which the LU is created.
  • the start LBA 14223 and the last LBA 14224 indicate the location of the LU address space in the flash box 210 , that is, each of the start address and the end address of the LU is indicated by an LBA (Logical Block Address).
  • FIG. 5 is a structure diagram of the memory management table 1423 .
  • the memory management table 1423 contains a cache segment No. 14231 , LUN 14232 , LBA 14233 , and length 14234 .
  • the cache segment No. 14231 is the identifier of each cache segment in the data cache area 141 , and a unique number is assigned to each cache segment.
  • Each of the LUN 14232 , the LBA 14233 , and the length 14234 is information for managing user data stored in the cache segment.
  • the LUN 14232 indicates the identification number of an LU in which the relevant data is stored.
  • the LBA 14233 indicates the start address of the data.
  • the length 14234 indicates the length of the data.
  • FIG. 6 is a structure diagram of the transfer list 1424 .
  • the transfer list 1424 contains a list No. 14241 , transfer source 14242 , transfer destination 14243 , cache segment No. 14244 , length 14245 , LUN 14246 , and LBA 14247 .
  • the list No. 14241 is the identifier of each transfer list, and a unique number is assigned to each list.
  • the transfer source 14242 and the transfer destination 14243 are the parameters for managing the data transfer direction.
  • the transfer source indicating “Mem” and the transfer destination indicating “FE (Front End)” means the data transfer from the memory 140 to the I/F chip 110 .
  • the transfer source indicating “Mem” and the transfer destination indicating “BE (Back End)” means the data transfer from the memory 140 to the I/F chip 160 .
  • the cache segment No. 14244 and the length 14245 indicate the location of the cache segment in which data to be DMA-transferred is to be stored (or stored) and the size of the data to be transferred, respectively.
  • the LUN 14246 and the LBA 14247 indicate the identification number and the start address of an LU to which the data to be transferred belongs, respectively.
  • FIG. 7 is a structure diagram of the transfer completion list 1425 .
  • the transfer completion list 1425 contains a transfer list No. 14251 .
  • the transfer list No. 14251 has stored therein the value of the list No. 14241 whose transfer has been completed. By referring to the transfer completion list 1425 , it is possible to check which transfer request has been completed.
  • FIG. 8 is a configuration diagram of each flash box 210 .
  • the flash box 210 includes a flash controller 220 and a memory array 230 .
  • the flash controller 220 controls the entire flash box 210 .
  • the memory array 230 includes the plurality of flash memory chips 231 .
  • the flash controller 220 includes an I/F chip 221 , a CPU 222 , ASICs 223 , and memory 224 .
  • the I/F chip 221 is used to connect the flash controller 220 and the storage controller 100 .
  • the I/F chip 221 and the storage controller 100 are connected via the connecting bus 180 described above.
  • the CPU 222 controls the entire flash box 210 .
  • the memory 224 has stored therein computer programs and data that are necessary for the CPU 222 to control the flash box 210 .
  • the ASICs 223 provide functions for controlling the flash memory chips 231 such as GC or WL.
  • the ASICs 223 include a DMA transfer unit 2231 and a memory controller (Mem CTL) 2232 .
  • the memory controller 2232 manages the basic operations such as read, write, and erasure performed to the flash memory chips 231 .
  • Such functional unit may be incorporated in a device other than the ASICs 223 .
  • the memory controller 2232 may be provided separately as a single unit.
  • the ASICs 223 are connected to each of the CPU 222 and the I/F chip 221 via a dedicated bus 225 such as a PCI.
  • the ASICs 223 are also connected to the memory 224 via a dedicated bus 226 such as a DDR 3. Further, the ASICs 223 are also connected to the flash memory chips 231 with a dedicated bus 227 via the memory controller 2232 .
  • FIG. 9 is a configuration diagram of the flash memory chip 231 .
  • the flash memory chip 231 includes a plurality of blocks 232 , and the memory is erased per block 232 .
  • Each block 232 includes a plurality of pages 233 that are the minimum read/write units of the flash memory chip 231 .
  • FIG. 10 is a diagram illustrating address assignment.
  • the blocks 232 are divided in terms of attributes into a normal block 236 and an alternate block 237 .
  • the normal block 232 is a block recognized as an LU space by the storage controller 100 .
  • the alternate block 237 is used as a data migration destination when GC or WL is executed or as a replacement block when the normal block has crashed.
  • Such attribute is not specific to each block 232 , and is changed as needed in accordance with the usage state of each block 232 .
  • the pages 233 in the normal block 236 are divided into a base area 234 and an update area 235 .
  • the base area 234 is an area provided as a real address space to the host system
  • the update area 235 is an area for, upon request to update pages, storing the target update data. Therefore, in a strict sense, the sum of the physical address spaces of the pages 233 that belong to the base area 234 is provided as a logical address space to the storage controller 100 .
  • Embodiment 1 illustrates an example in which the base area 234 and the update area 235 reside in the same block, such areas can be provided in separate blocks.
  • FIG. 11 is a structure diagram of the memory 224 .
  • the memory 224 includes a data cache area 2241 and a system area 2242 .
  • the data cache area 2241 is an area used to temporarily store user data.
  • the system area 2242 has stored therein programs and various tables for controlling the flash box 210 .
  • the data cache area 2241 is divided into a plurality of segments of a given size so that user data 22411 is stored on a cache segment basis.
  • the system area 2242 has stored therein a program group 22421 , an address management table 22422 , a block management table 22423 , and a transfer list 22424 .
  • the program group 22421 includes various programs executed by the CPU 222 .
  • the address management table 22422 manages the correspondence between the logical address and the physical address.
  • the block management table 22423 manages the status of each block 232 .
  • the transfer list 22424 is used when the ASICs 223 execute DMA transfer.
  • FIG. 12 is a structure diagram of the address management table 22422 .
  • the address management table 22422 contains start LBA 224221 , chip No. 224222 , block No. 224223 , and page No. 224224 .
  • the start LBA 224221 indicates the offset value of LBA recognized by the storage controller 100 . This value is determined by the size of the page 233 that is the minimum read/write unit of the flash memory chip 231 . For example, when each page 233 has an address space of 20 h, the start LBA 224221 has values at intervals of 20 h.
  • the chip No. 224222 , the block No. 224223 , and the page No. 224224 are information for identifying the page 233 corresponding to the logical address.
  • FIG. 13 is a structure diagram of the block management table 22423 .
  • the block management table 22423 contains a chip No. 224231 , block No. 224232 , attribute 224233 , status 224234 , the residual number of valid pages 224235 , and the number of erasures 224236 .
  • the chip No. 224231 indicates the number of the flash memory chip 231 to which the block 232 belongs.
  • the block No. 224232 indicates the number of the block in the flash memory chip 231 .
  • Each block 232 is identified by the chip No. 224231 and the block No. 224232 .
  • the attribute 224233 indicates the attribute of the block 232 described with reference to FIG. 10 , and either “Normal” or “Alternate” is set.
  • the status 224234 indicates the operating status of each block 232 . For example, “WL” is set when WL is executed to the relevant block, “GC” is set when GC is executed to the relevant block, “Normal” is set when read/write access to the relevant block is allowed, and “n/a” is set for other statuses.
  • the residual number of valid pages 224235 indicates the residual number of the pages 233 that are assigned as the update area 235 within the relevant block 232 . It should be noted that if the attribute of the relevant block is “Alternate,” or if the processing of freeing pages such as WL or GC is executed, “n/a” is set.
  • the number of erasures 224236 indicates the number of erase operations performed to the relevant block 232 , and is mainly used to manage information about whether the relevant block 232 has been accessed up to its maximum allowable write access limit.
  • FIG. 14 is a flow chart illustrating the basic I/O sequence of the storage device 10 . Hereinafter, each step of FIG. 14 will be described.
  • the storage device 10 receives a data write request or a data read request (an I/O request) from the host computer 40 , upon which the storage controller 100 starts the present flow chart.
  • the storage controller 100 determines whether the request received in step S 10 is a read (RD) request or a write (WR) request. If the request is determined to be a read request, the flow proceeds to step S 30 , and if the request is determined to be a write request, the flow proceeds to step S 50 .
  • RD read
  • WR write
  • the storage controller 100 executes staging processing in which data requested to be read is transferred from the flash box 210 to the memory 140 .
  • staging processing in which data requested to be read is transferred from the flash box 210 to the memory 140 .
  • the details of the present step will be described below with reference to FIGS. 15 and 16 .
  • the storage controller 100 transmits the data, which has been transferred from the flash box 210 to the memory 140 in step S 30 , to the host computer 40 . By this step, the processing of the read request issued from the host computer 40 terminates.
  • the storage controller 100 requests the host computer 40 to transfer data that is requested to be written. Then, the storage controller 100 receives the data and stores it into the memory 140 .
  • the storage controller 100 informs the host computer 40 of the successful completion of the transfer.
  • the storage controller 100 executes destaging processing in which data retained in the memory 140 is stored into the flash box 210 .
  • the destaging processing is executed each time a write request is issued in this drawing, the destaging processing can be executed at any given time. In that case, destaging processing is collectively executed after steps S 50 to S 60 are repeated a plurality of times. Accordingly, the target data to be written can be collectively transferred to the flash box 210 through single destaging processing, whereby the processing efficiency can be expected to be increased.
  • FIG. 15 is an operation flow chart of an I/O processing program executed by the storage controller 100 .
  • This flow chart illustrates a sequence for executing the read processing and the write processing described with reference to FIG. 14 .
  • each step of FIG. 15 will be described. It should be noted that description will be made first of the read processing, and then of the write processing for the sake of convenience of the description.
  • the storage controller 100 determines the presence or absence of an I/O request issued from the host computer 40 . If the answer to step S 100 is Yes, the flow proceeds to step S 110 , and if the answer to step S 100 is No, the flow proceeds to step S 140 .
  • the storage controller 100 analyzes the command issued from the host computer 40 and determines whether the type of the request is RD or WR. If the type of the request is determined to be RD, the flow proceeds to step S 120 , and if the type of the request is determined to be WR, the flow proceeds to step S 130 .
  • steps performed when the type of the request is RD will be described.
  • the storage controller 100 creates a data transfer request (a BE transfer request (1)) to execute the staging processing (processing of reading data stored in the flash memory chip 231 ). Specifically, the storage controller 100 refers to the memory management table 1423 to secure a cache segment in which data read from the flash memory chip 231 is to be stored. Next, the storage controller 100 creates the transfer list 1424 . In this case, “BE” is set as the transfer source 14242 , “Mem” is set as the transfer destination 14243 , and the number of the secured cache segment or the like is set as the cache segment No. 14244 . Then, the flow again returns to step S 100 so that the storage controller 100 determines the presence or absence of any other I/O requests. Herein, the flow proceeds to step S 140 on the assumption that there are no other I/O requests.
  • the storage controller 100 determines the presence or absence of the transfer list 1424 to determine the presence or absence of a data transfer request. If the transfer list 1424 has been created, the storage controller 100 determines a transfer request to be present, and the flow proceeds to step S 150 . If the transfer list 1424 has not been created, the flow proceeds to step S 160 .
  • the storage controller 100 executes data transfer processing in accordance with the description of the transfer list 1424 . This step will be described in detail with reference to FIG. 16 .
  • the storage controller 100 determines the presence or absence of the transfer completion list 1425 to determine if there is any data transfer executed in step S 150 that has been completed. If the answer to step S 160 is Yes, the flow proceeds to step S 180 , and if the answer to step S 160 is No, the flow proceeds to step S 170 .
  • step S 100 the flow returns to step S 100 to repeat the same processing, and if the storage controller 100 terminates the operation, the flow ends.
  • the storage controller 100 refers to the transfer list 1424 based on the list No. 14251 whose data transfer processing has been completed as a key, and acquires the values of the transfer source 14242 and the transfer destination 14243 . If the transfer source 14242 indicates “BE” and the transfer destination 14243 indicates “Mem,” the flow proceeds to step S 190 , and if the transfer source 14242 indicates “Mem” and the transfer destination 14243 indicates “FE,” the flow proceeds to step S 220 .
  • the transfer source 14242 indicates “BE” and the transfer destination 14243 indicates “Mem,” it can be confirmed that the processing of reading data from the flash box 210 into the memory 140 has been completed. In that case, it is further necessary to transfer the data from the memory 140 to the host computer 40 . Thus, a data transfer request to that end should be created in step S 190 . If the transfer source 14242 indicates “Mem” and the transfer destination 14243 indicates “FE,” it can be confirmed that the processing of transferring data from the memory 140 to the host computer 40 , that is, the read processing has been completed.
  • the storage controller 100 creates as the transfer list 1424 a data transfer request (an PE transfer request (3)) for transferring to the host computer 40 the data that has been read from the flash box 210 into the memory 140 in step S 150 .
  • a data transfer request (an PE transfer request (3)) for transferring to the host computer 40 the data that has been read from the flash box 210 into the memory 140 in step S 150 .
  • “Mem” is set as the transfer source 14242 and “FE” is set as the transfer destination 14243 in the transfer list 1424 .
  • the flow again returns to step S 160 so that the storage controller 100 determines the presence or absence of any other data transfer that has been completed.
  • the storage controller 100 determines that the processing of transferring the data to the host computer 40 is completed.
  • FIG. 15 Steps S 100 to S 110 )
  • the storage controller 100 executes the same processing as that in steps S 100 to S 110 described with respect to the read processing.
  • steps S 100 to S 110 described with respect to the read processing.
  • steps performed when the type of the request is WR will be described.
  • the storage controller 100 creates a data transfer request (an FE transfer request (2)) to execute data receiving processing (processing of receiving the target data to be written from the host computer 40 ). Specifically, the storage controller 100 refers to the memory management table 1423 to secure a cache segment in which data received from the host computer 40 is to be stored. Then, the storage controller 100 creates the transfer list 1424 . In this case, “FE” is set as the transfer source 14242 , “Mem” is set as the transfer destination 14243 , and the number of the secured cache segment or the like is set as the cache segment No. 14244 . Then, the flow again returns to step S 100 so that the storage controller 100 determines the presence or absence of any other I/O requests. Herein, the flow proceeds to step S 140 on the assumption that there are no other I/O requests.
  • FIG. 15 Steps S 140 to S 170
  • the storage controller 100 refers to the transfer list 1424 based on the list No. 14251 whose data transfer processing has been completed as a key, and acquires the values of the transfer source 14242 and the transfer destination 14243 . If the transfer source 14242 indicates “FE” and the transfer destination 14243 indicates “Mem,” the flow proceeds to step S 200 , and if the transfer source 14242 indicates “Mem” and the transfer destination 14243 indicates “BE,” the flow proceeds to step S 230 .
  • the storage controller 100 informs the host computer 40 of the successful completion of the data writing. Although data is actually written in a subsequent step, the storage controller 100 herein is configured to return a response to the host computer 40 upon reception of the target data to be written in order to promptly return a response to the host computer 40 .
  • the storage controller 100 creates as the transfer list 1424 a data transfer request (a BE transfer request (4)) for transferring to the flash box 210 the data that has been received from the host computer 40 and stored into the memory 140 in step S 150 .
  • a data transfer request (a BE transfer request (4)) for transferring to the flash box 210 the data that has been received from the host computer 40 and stored into the memory 140 in step S 150 .
  • “Mem” is set as the transfer source 14242 and “BE” is set as the transfer destination 14243 in the transfer list 1424 .
  • the flow again returns to step S 160 so that the storage controller 100 determines the presence or absence of any other data transfer that has been completed.
  • the storage controller 100 determines that the processing of transferring the data to the flash box 210 has been completed.
  • FIG. 16 is a detailed flow chart of step S 150 in FIG. 15 . Hereinafter, each step of FIG. 16 will be described.
  • the CPU 130 refers to the transfer list 1424 and determines the list No. 14241 with the lowest number to be the target command to be processed.
  • the CPU 130 determines whether the command that has been determined to be the target command to be processed in step S 151 is an I/O request to the flash box 210 . Specifically, the CPU 130 determines “BE” set as the transfer destination 14243 to be an I/O request to the flash box 210 . If the answer to step S 152 is Yes, the flow proceeds to step S 153 , and if the answer to step S 152 is No, the flow proceeds to step S 155 .
  • a command that is not an I/O request to the flash box 210 is, for example, a data transfer request to the host computer 40 .
  • the CPU 130 queries the flash box 210 as to whether data transfer is executable. This step will be described in detail with reference to FIGS. 17 to 19 .
  • the CPU 130 determines whether data transfer is executable. If data transfer is determined to be executable, the flow proceeds to step S 155 , and if data transfer is determined to be not executable, the flow proceeds to step S 156 .
  • the CPU 130 sends the transfer list 1424 to the ASICs 150 .
  • the ASICs 150 execute the data transfer described in the transfer list 1424 using the function of the DMA transfer unit 151 .
  • the CPU 130 if data transfer is determined to be not executable in step S 154 , determines that write access to the flash box 210 cannot be executed promptly. Thus, the CPU 130 waits for a predetermined period of time without starting data transfer. With regard to data that is not transferred in this step, the CPU 130 will attempt to transfer such data when the present flow is executed again.
  • the CPU 130 determines the presence or absence of any other transfer requests that have not been executed yet. If the answer to step S 157 is Yes, the flow returns to step S 151 , and if the answer to step S 157 is No, the flow ends.
  • FIG. 17 is a structure diagram of a request command 1000 and a reply command 2000 exchanged between the storage controller 100 and the flash controller 220 in step S 153 .
  • the structure of each command will be described.
  • the request command 1000 is a command sent from the storage controller 100 to the flash controller 220 , and includes a command No. 1001 and request content 1002 . Depending on the type of the request content 1002 , the request command 1000 may further include additional information. Hereinafter, an example will be described in which the storage controller 100 queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. In that case, the request command 1000 further includes a start address 1003 and write request size 1004 .
  • the value of the list No. 14241 of the transfer list 1424 that is currently processed by the ASICs 150 is set.
  • the request content 1002 a value indicating a command that causes the request command 1000 to query as to whether data transfer is executable is set.
  • the value of the LBA 14247 is set as the start address 1003
  • the value of the length 14245 is set as the write request size 1004 .
  • the reply command 2000 is a command sent from the flash controller 220 to the storage controller 100 , and includes information to be returned in response to the request command 1000 .
  • the reply command 2000 includes a command No. 2001 and reply content 2002 .
  • the value of the command No. 1001 that has been set for the corresponding request command 1000 is set.
  • the reply content 2002 a value indicating a reply content of the reply command 2000 is set. For example, if the request command 1000 is a command that queries as to whether data transfer is executable, a value indicating that data transfer is executable or not is set as the reply content 2002 .
  • the configurations of the commands are not limited to the examples shown in FIG. 17 .
  • the flash controller 220 can be configured to send the request command 1000 to the storage controller 100
  • the storage controller 100 can be configured to send the reply command 2000 to the flash controller 220 .
  • FIG. 18 is a flow chart showing the details of step S 153 . Hereinafter, each step of FIG. 18 will be described.
  • the storage controller 100 creates the request command 1000 .
  • the content of the transfer list 1424 that is currently processed by the ASICs 150 is set as each item of the request command 1000 .
  • the storage controller 100 sends the created request command 1000 to the flash controller 220 .
  • the flash controller 220 receives the request command 1000 .
  • the flash controller 220 checks the content of the request command 1000 . This step will be described in detail with reference to FIG. 19 .
  • the flash controller 220 based on the result of step S 1533 , creates the reply command 2000 and sends it to the storage controller 100 .
  • the storage controller 100 receives the reply command 2000 and checks the content, and then terminates the present flow.
  • FIG. 19 is a detailed flow chart of step S 1533 . Hereinafter, each step of FIG. 19 will be described.
  • the flash controller 220 refers to the address management table 22422 to search for the chip No. 224222 and the block No. 224223 in which data specified by the request command 1000 is stored.
  • the flash controller 220 refers to the status information 224234 in the block management table 22423 , and determines whether GC is currently executed to the relevant block 232 . If the answer to step S 15331 is Yes, the flow proceeds to step S 15336 , and if the answer to step S 15331 is No, the flow proceeds to step S 15332 .
  • the flash controller 220 refers to the status information 224234 in the block management table 22423 , and determines whether WL is currently executed to the relevant block. If the answer to step S 15332 is Yes, the flow proceeds to step S 15336 , and if the answer to step S 15332 is No, the flow proceeds to step S 15333 .
  • the flash controller 220 determines whether GC will be executed to the block for which the data transfer request has been received. Specifically, the flash controller 220 determines from the write request size 1004 included in the request command 1000 how many pages 233 in the update area would be necessary to store the data. If the number of the necessary pages 233 is determined to be insufficient, the flash controller 220 determines that GC will be executed, and if not, the flash controller 220 determines that GC will not be executed. If the answer to step S 15333 is Yes, the flow proceeds to step S 15335 , and if the answer to step S 15333 is No, the flow proceeds to step S 15334 .
  • Examples of the determination methods in the present step include a method of calculating the number of necessary pages by dividing the write request size by the size of the page 233 .
  • the flash controller 220 if the quotient is greater than the residual number of valid pages 224235 of the block, determines that GC will be executed, and if the quotient is less than or equal to the residual number of valid pages 224235 of the block, determines that GC will not be executed.
  • the flash controller 220 determines the data transfer to be executable and terminates the present flow.
  • the flash controller 220 starts execution of GC to the relevant block.
  • the flash controller 220 determines the data transfer to be non-executable, and terminates the present flow.
  • FIG. 20 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 1. Hereinafter, each step of FIG. 20 will be described.
  • FIG. 20 Steps S 3101 to S 3102 )
  • the storage controller 100 starts the processing of querying the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. At this time, the storage controller 100 sends to the flash controller 220 the request command 1000 to query as to whether data write access to the flash memory chip 231 is allowed.
  • FIG. 20 Steps S 3103 to S 3104 )
  • the flash controller 220 is executing GC to the flash memory chip 231 .
  • the flash controller 220 sends to the storage controller 100 the reply command 2000 indicating that data write access to the flash memory chip 231 is not allowed.
  • the storage controller 100 suspends the data write processing to the flash memory chip 231 and executes another processing while waiting.
  • the wait time herein can be either fixed or determined as needed.
  • the flash controller 220 after the completion of the GC, enters a state ready to receive a request to write data to the flash memory chip 231 .
  • FIG. 20 Steps S 3107 to S 3108 )
  • the storage controller 100 restarts the data write processing that has been suspended in step S 3105 , and executes the same processing as that in steps S 3101 to S 3102 .
  • FIG. 20 Steps S 3109 to S 3110 )
  • the flash controller 220 sends the reply command 2000 to the storage controller 100 as in steps S 3103 to S 3104 .
  • the flash controller 220 sends the reply command 2000 indicating that data write access to the flash memory chip 231 is allowed.
  • FIG. 20 Steps S 3111 to S 3112 )
  • the storage controller 100 starts the data write processing to the flash memory chip 231 . At this time, the storage controller 100 sends to the flash controller 220 the request command 1000 to write data to the flash memory chip 231 .
  • FIG. 20 Steps S 3113 to S 3116 )
  • the flash controller 220 executes data writing to the flash memory chip 231 . Upon completion of the data writing, the flash controller 220 sends to the storage controller 100 the reply command 2000 indicating the completion of the data writing. The storage controller 100 , upon receipt of the reply command 2000 , terminates the data write processing.
  • the storage controller 100 and the flash controller 220 operate cooperatively, whereby the time required for the storage controller 100 to execute destaging can be minimized. Specifically, the time required for the storage controller 100 to execute destaging can be made substantially equal to the time in which data is actually written to the flash memory chip 231 . Accordingly, the time in which the storage controller 100 can execute another processing (step S 3105 ) can be increased with the result that the entire performance of the storage system 10000 can be improved.
  • the storage controller 100 before writing data to the flash memory chip 231 , queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed using the request command 1000 . Accordingly, the storage controller 100 can actually start data writing after the flash memory chip 231 becomes write-accessible, and thus the substantial data write time can be suppressed.
  • the storage controller 100 if it has received the reply command 2000 indicating that data write access to the flash memory chip 231 is not allowed, starts another processing. Accordingly, it becomes possible for the storage controller 100 to continue the processing without wasting the time in which the data write access to the flash memory chip 231 is not allowed, whereby the performance of the storage device 10 can be improved.
  • the storage controller 100 having completed another processing after the receipt of the reply command 2000 indicating that data write access to the flash memory chip 231 is not allowed, again queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. Accordingly, even if data write access to the flash memory chip 231 is not allowed at a given moment in time, data writing can successfully be completed without being abandoned.
  • Embodiment 1 description has been made of an example in which the storage controller 100 queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. However, it is also possible to configure the storage controller 100 to query as to whether data read access to the flash memory chip 231 is allowed. Accordingly, the storage controller 100 , during execution of GC, can execute another processing without issuing a data read request, and can, after the completion of the GC, read the data from the flash memory chip 231 .
  • the flash controller 220 can be configured to execute GC not only when it has received an access request from the storage controller 100 but also at any given time as needed. This is also true of the following embodiments.
  • Embodiment 1 when the flash controller 220 sends to the storage controller 100 the reply command 2000 indicating that the flash memory chip 231 is inaccessible, information on the wait time can be included in the reply command 2000 .
  • the storage controller 100 after having waited for the wait time described in the reply command 2000 , retries data writing.
  • the wait time herein depends on the time required for the flash controller 220 to execute GC.
  • a value indicating an appropriate wait time can be stored in advance in a storage device such as the memory 224 in accordance with the performance of each device so that the flash controller 220 will read such value and inform the storage controller 100 of the value.
  • a value indicating a wait time can be embedded in a program executed by the flash controller 220 .
  • Embodiment 2 of the present invention an operation example will be described in which GC is executed in advance before the storage controller 100 issues a data write request to the flash controller 220 .
  • FIG. 21 is an operation flow chart of an I/O processing program executed by the storage controller 100 in Embodiment 2. This flow differs from that in FIG. 15 in that, in addition to the flow described with reference to FIG. 15 , step S 240 is provided after step S 130 . Steps other than that are the same as those in FIG. 15 .
  • FIG. 22 shows a detailed flow of step S 240 . Hereinafter, each step of FIG. 22 will be described.
  • the storage controller 100 determines whether the currently processed I/O request is a sequential write request. If the answer to step S 241 is Yes, the flow proceeds to step S 242 , and if the answer to step S 241 is No, the flow ends.
  • the reason for determining whether the I/O request is a sequential write request in this step is that sequential write commands are typically write commands that are executed in writing large-size data.
  • determining if an I/O request is a sequential write request is a method of determining if addresses of data to be written are consecutive. Specifically, the storage controller 100 , with reference to the memory management table 1423 , determines whether data, which has an immediately preceding address of the address (LBA) contained in the I/O request received from the host computer 40 , resides in the memory 140 . If such data is determined to reside in the memory 140 , the storage controller 100 determines the relevant I/O request to be a sequential write request.
  • LBA immediately preceding address of the address
  • the storage controller 100 creates the request command 1000 to execute GC, and sends it to the flash controller 220 .
  • FIG. 23 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 2. Hereinafter, each step of FIG. 23 will be described.
  • FIG. 23 Steps S 3201 to S 3205 )
  • the storage controller 100 has received a sequential write request from the host computer 40 .
  • the storage controller 100 sends the request command 1000 to start execution of GC to the flash controller 220 .
  • the storage controller 100 suspends the data write processing to the flash memory chip 231 and executes another processing while waiting.
  • FIG. 23 Steps S 3206 to S 3208 )
  • the flash controller 220 executes GC to the flash memory chip 231 . Upon completion of the GC, the flash controller 220 enters a state ready to receive a request to write data to the flash memory chip 231 .
  • FIG. 23 Steps S 3209 to S 3218 )
  • the storage controller 100 before starting destaging processing to the flash controller 220 , requests the flash controller 220 to start GC. Accordingly, sufficient empty pages can be created in the flash memory chip 231 before data writing is actually started. Thus, even when processing of destaging large-size data is subsequently generated along with a sequential write request, the flash controller 220 can promptly store such data without executing GC to the flash memory chip 231 .
  • the storage controller 100 is configured to, upon receipt of a sequential write request from the host computer 40 , determine that processing of writing large-size data will be generated. However, if it is possible to know in advance that the size of data to be written is large through the use of other means, the storage controller 100 need not necessarily determine the necessity of the execution of GC based on whether a received I/O request is a sequential write request or not. For example, if the storage controller 100 is informed of a data size together with a data write request from the host computer 40 , it can determine the necessity of the execution of GC based on the data size.
  • the storage controller 100 is configured to issue the request command 1000 to execute GC in advance to the flash controller 220 .
  • the storage controller 100 can be configured to send a data write request directly to the flash controller 220 so that the flash controller 220 determines the necessity of the execution of GC.
  • the flash controller 220 in order to execute GC, sends to the storage controller 100 the reply command 2000 indicating that data write access is not allowed, and starts the GC.
  • the subsequent operations are the same as those in Embodiment 2.
  • Embodiment 3 of the present invention when an area of the flash memory chip 231 that has been used so far becomes no longer used, for example, at the timing when an LU is deleted, the storage controller 100 instructs the flash controller 220 to free the unused area. Accordingly, alternate blocks 237 that are available in the flash memory chip 231 can be secured.
  • a storage controller frees an LU that is a logical address space at the time of deleting the LU, but does not execute the processing of freeing a physical area like the format processing performed to HDDs. That is, a storage area appears to be freed when an LU is deleted, but in reality, data keeps on residing in the physical area.
  • Embodiment 3 aims to address such a problem.
  • FIG. 24 is a flow chart showing the processing in which the storage controller 100 , after deleting an LU, requests the flash controller 220 to free the block 232 .
  • the storage controller 100 after deleting an LU, requests the flash controller 220 to free the block 232 .
  • FIG. 24 each step of FIG. 24 will be described.
  • the storage controller 100 executes the present processing flow during or after the processing of deleting an LU.
  • the storage controller 100 issues the request command 1000 to request the flash controller 220 to free the block 232 that has been assigned to the deleted LU.
  • the start LBA, the last LBA, or the like of the area to be freed is set.
  • Such a range to be freed can be identified by referring to the start LBA 14223 and the last LBA 14224 of the deleted LU on the LU management table 1422 .
  • the flash controller 220 receives the request command 1000 .
  • the flash controller 220 frees the requested block 232 in accordance with the content of the request command 1000 . Specifically, the flash controller 220 identifies the block 232 to be freed with reference to the address management table 22422 , and issues an erase command to the block 232 . In addition, the flash controller 220 changes the attribute 224233 of the block in the block management table 22423 to “Alternate.”
  • the flash controller 220 creates the reply command 2000 to inform the storage controller 100 of the completion of the processing of freeing the block 232 , and sends it to the storage controller 100 .
  • the storage controller 100 receives the reply command 2000 from the flash controller 220 , and terminates the present processing flow.
  • the processing of freeing a storage area of the flash memory chip 231 is executed at the point when the storage controller 100 has deleted an LU. Accordingly, it is possible to reduce the possibility that a shortage of the alternate blocks 237 may occur when the flash controller 220 attempts to execute GC, in which case the GC cannot be executed.
  • the frequency of execution of GC will increase under the operating conditions that GC is executed in advance before large-size data is written as described in Embodiment 2. In such a case, the advantageous effect of the method described in Embodiment 3 can be maximized.
  • Embodiment 4 of the present invention a method of securing available alternate blocks 237 when the storage device 10 uses a volume virtualization technique will be described.
  • the reason that the available alternate blocks 237 should be secured is the same as that described in Embodiment 3.
  • the volume virtualization technique is a technique of presenting a virtual volume, which is larger than the physical volume, of the storage device 10 to the host computer 40 in advance, and adding a storage capacity only after the virtual volume is actually needed.
  • JP Patent Publication (Kokai) No. 2007-66259 A discloses an example of a storage device that uses a volume virtualization technique.
  • Embodiment 4 a storage area that is assigned when data is written on the storage device 10 with a virtualized storage capacity is referred to as a “segment” in the present invention.
  • JP Patent Publication (Kokai) No. 2007-66259 A discloses, as a means of improving the performance of a storage device that uses a volume virtualization technique, migrating a segment (the term “segment” is also used in this document) from one storage device to another.
  • FIG. 25 is a structure diagram of the memory 140 in accordance with Embodiment 4. This structure differs from that described with reference to FIG. 3 in Embodiment 1 in that a segment management table 1426 , an LU management table 1427 , and a statistical information management table 1428 are newly provided instead of the LU management table 1422 . Structures other than that of the memory 140 are the same as those in Embodiments 1 to 3.
  • the segment management table 1426 manages the structures of segments.
  • the LU management table 1427 manages the structures of LUs.
  • the statistical information management table 1428 has recorded thereon the statistical information such as the number of commands issued to each segment.
  • FIG. 26 is a structure diagram of the segment management table 1426 .
  • the segment management table 1426 contains a segment No. 14261 , flash box No. 14262 , start LBA 14263 , and last LBA 14264 .
  • the segment No. 14261 is the identifier of a created segment.
  • the flash box No. 14262 is the identification number of the flash box 210 in which the segment is created.
  • the start LBA 14263 and the last LBA 14264 indicate the location of the segment's address space in the flash box 210 , that is, each of the start address and the end address of the segment is indicated by an LBA.
  • FIG. 27 is a structure diagram of the LU management table 1427 .
  • the LU management table 1427 is provided in a number corresponding to the number of created LUs.
  • the LU management table 1427 contains a LUN 14271 , LU size 14272 , and segment No. 14273 .
  • the LUN 14271 is the identifier of an LU.
  • the LU size 14272 indicates the size of the LU in units of GB (GigaByte).
  • the segment No. 14273 indicates the identification number of a segment assigned to the LU, and is provided in a number corresponding to the LU size.
  • n/a that indicates “not assigned” is set for an area that is not assigned a segment, whereas for an area that has been assigned a segment, the identification number of the segment is set.
  • FIG. 28 is a structure diagram of the statistical information management table 1428 .
  • the statistical information management table 1428 contains a segment No. 14281 , the number of write commands 14282 , and the number of read commands 14283 .
  • the segment No. 14281 is the identifier of a segment.
  • the number of write commands 14282 and the number of read commands 14283 indicate the number of write I/O requests and the number of read I/O requests, respectively, that have been issued to the segment from the host computer 40 .
  • the storage controller 100 after migrating a segment or deleting an LU, requests the flash controller 220 to free the block 232 corresponding to the segment. This step is the same as that shown in FIG. 24 , but differs in the method of identifying the target block 232 to be freed.
  • the storage controller 100 When migrating a segment, the storage controller 100 identifies from the segment management table 1426 the start LBA 14263 and the last LBA 14264 of the segment, based on the segment No. 14261 of the migration source as a key.
  • the storage controller 100 When deleting an LU, the storage controller 100 first refers to the LU management table 1427 based on the LUN 14271 as a key, and identifies the segment No. 14273 that constitutes the deleted LU. Next, the storage controller 100 refers to the segment management table 1426 based on the segment No. as a key, and identifies the start LBA 14263 and the last LBA 14264 of the segment.
  • Embodiment 4 the processing of freeing a storage area of the flash memory chip 231 is executed at the point when the storage controller 100 has migrated a segment or deleted an LU. Accordingly, even when the storage device 10 employs a volume virtualization technique, an advantageous effect similar to that in Embodiment 3 can be exerted.
  • Embodiment 4 even when the storage controller 100 has deleted a segment, a storage area of the flash memory chip 231 can be freed in a similar way.
  • Embodiment 5 of the present invention a method of evenly distributing the frequency of write I/O access or the number of data erasure cycles across each of the plurality of flash boxes 210 will be described.
  • the configuration of Embodiment 5 is based on the configuration described in Embodiment 4 that uses a volume virtualization technique, a similar method can also be used under the configuration described in any of Embodiments 1 to 3.
  • the frequency of GC or WL executed to the flash box 210 would correspondingly increase, and thus the number of erasure cycles of such a flash memory chip 231 would also increase.
  • the flash memory chip 231 is a device with a finite number of erasure cycles, if the erasure cycles of a particular block 232 have exceeded a threshold limit, resulting in a crash, such a crashed block 232 is replaced with the alternate block 237 .
  • the number of the available alternate blocks 237 in the flash box 210 that is write I/O accessed with high frequency would easily decrease. The resulting shortage of the available alternate blocks 237 can lead to a decrease in the write processing performance.
  • the storage controller 100 based on statistical information such as the frequency of write I/O access to or the number of erasure cycles of each flash box 210 , migrates segments between the flash boxes 210 . Accordingly, the frequency of write I/O access, the number of erasure cycles, and the like can be distributed evenly across each of the plurality of flash boxes 210 , so that the system performance can be maximized and the lifetime of each flash box 210 can be extended.
  • the storage controller 100 and the flash controller 220 exchange statistical information such as the number of data erasure cycles of each flash box 210 .
  • FIG. 29 is a flow chart showing the processing in which the storage controller 100 makes equal the load on each flash box 210 .
  • the storage controller 100 makes equal the load on each flash box 210 .
  • the storage controller 100 creates the request command 100 to request the flash controller 220 to send statistical information on the flash boxes 210 , and sends it to the flash controller 220 .
  • the flash controller 220 receives the request command 1000 .
  • the flash controller 220 acquires the statistical information on the flash boxes 210 based on the content of the request command 1000 . Specifically, the flash controller 220 first acquires the status of each block 232 with reference to the block management table 22423 . Next, the flash controller 220 acquires statistical information such as the sum of the numbers of erasure cycles of all of the blocks 232 from the statistical information management table 1428 .
  • the flash controller 220 creates the reply command 2000 corresponding to the request content of the request command 1000 , and sends it to the storage controller 100 .
  • the following information can be considered: the numbers of the assigned normal blocks 236 and alternate blocks 237 , the number of the blocks 232 to which GC or WL is executed, the residual number of valid pages, the number of data erasure cycles, and the like.
  • the flash controller 220 is configured to be capable of sending a variety of statistical information on the flash boxes 210 in response to a request from the storage controller 100 .
  • the storage controller 100 receives the reply command 2000 from the flash controller 220 .
  • the storage controller 100 determines the target segments to be exchanged. For example, the storage controller 100 compares the number of data erasure cycles of each flash box 210 acquired from the flash controller 220 with the number of write commands 14282 in the statistical information management table 1428 , and determines a segment with the greatest number of write commands to be allocated to the flash box 210 with the smallest number of data erasure cycles.
  • the storage controller 100 exchanges the segments based on the result of step S 405 . Accordingly, the frequency of access to each segment can be made equal and I/O processing can be distributed evenly across each flash box 210 .
  • the storage controller 100 can acquire statistical information on the flash boxes 210 from the flash controller 220 , and thus can make equal the load on each flash box 210 based on the acquired information. Accordingly, the number of data write cycles or data erasure cycles can be distributed evenly not only within each flash memory chip 231 but also across each flash box 210 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The present invention aims to improve the performance of accessing flash memory used as a storage medium in a storage device. In the storage device in accordance with the present invention, a storage controller, before accessing the flash memory, queries a flash controller as to whether the flash memory is accessible.

Description

TECHNICAL FIELD
The present invention relates to a storage device.
BACKGROUND ART
A storage device is a device that has a disk for storing data and a controller for controlling the disk and is intended to provide a high-capacity data storage space to a host or a computer that is connected to the storage device.
When a storage device is constructed from a plurality of disks with a RAID (Redundant Array of Independent (or Inexpensive) Disks) configuration, the I/O processing speed can be increased and high fault tolerance against disk failures can be achieved.
Patent Literature 1 below discloses a storage device having an HDD (Hard Disk Drive) as a storage medium.
Meanwhile, in recent years, flash memory has been gaining increased attention as a new storage medium. Flash memory is characterized by its faster read and write times than those of HDDs. For example, although HDDs have read and write times each on the order of several milliseconds, there has been reported a performance data example of NAND flash memory that has a read time of 25 microsecond and a write time of 300 microsecond.
In addition, flash memory, unlike HDDs, has another characteristic that its physical area to which data has been written once cannot be overwritten. In order to rewrite data in flash memory, it is necessary to once erase the data on a “block” basis. However, flash memory has a limited number of write cycles due to the physical reasons of memory cells for storing information. Further, there has been reported data that flash memory requires an erase time of about 1 ms.
In order to address the aforementioned limitations, flash memory typically employs a method in which the correspondence relationship between the address and the cell location that is shown to a host system is retained, and when data stored in a block is to be updated, the update data is stored into an empty page, in which data is not written, and the original data stored in the block is invalidated. However, as the data is updated repeatedly, the number of empty pages will decrease. As a technique of addressing a shortage of empty pages, a technique called “garbage collection (hereinafter referred to as GC)” has been laid open to the public. Other than the GC, there has also been known a technique called “wear leveling (hereinafter referred to as WL)” which makes equal the number of write operations performed to each cell.
CITATION LIST Patent Literature
  • PTL 1: JP Patent Publication (Kokai) No. 2005-182104 A
SUMMARY OF INVENTION Technical Problem
As described above, although flash memory has a very fast read time, it can be said to be a device having slow write and erase times. Further, from the aspect of control, flash memory should be controlled with GC, WL, or the like that has not been used for the conventional HDDs.
For example, in data writing (hereinafter referred to as “write”), if there are no sufficient empty pages for storing update data, empty pages are secured by executing GC once; thereafter, data is stored into the secured empty pages. In such a case, time required for the completion of the write processing corresponds to the sum of the GC execution time+the actual data write time. In particular, since flash memory requires long data write and erase times, it may take several milliseconds until the completion of the write processing in some cases. This, however, could result in degraded performance down to a level of, or below the performance of HDDs.
FIG. 1 is a diagram illustrating the aforementioned problem with flash memory. When GC+write (which is described as “Program” in FIG. 1) are executed to flash memory in response to a write request issued from a storage controller that controls a storage system to a flash controller that controls the flash memory, the storage controller must wait for a long time until the completion of the write. In this period, the storage controller would keep consuming resources of devices such as cache memory for temporally storing data, which could result in degraded performance.
The present invention has been made in order to solve the aforementioned problem. It is an object of the present invention to improve the performance of accessing flash memory used as a storage medium in a storage device.
Solution to Problem
In a storage device in accordance with the present invention, a storage controller, before accessing flash memory, queries a flash controller as to whether the flash memory is accessible.
Advantageous Effects of Invention
According to a storage device of the present invention, whether flash memory is accessible is queried in advance, whereby the wait time in which the flash memory is inaccessible can be reduced. Accordingly, the substantial access performance to the flash memory can be improved.
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is a diagram illustrating the aforementioned problem with flash memory.
FIG. 2 is a diagram showing the configuration of a storage system 10000 in accordance with Embodiment 1.
FIG. 3 is a structure diagram of memory 140.
FIG. 4 is a structure diagram of an LU management table 1422 for managing the address spaces of LUs.
FIG. 5 is a structure diagram of a memory management table 1423.
FIG. 6 is a structure diagram of a transfer list 1424.
FIG. 7 is a structure diagram of a transfer completion list 1425.
FIG. 8 is a configuration diagram of a flash box 210.
FIG. 9 is a configuration diagram of a flash memory chip 231.
FIG. 10 is a diagram illustrating address assignment.
FIG. 11 is a structure diagram of memory 224.
FIG. 12 is a structure diagram of an address management table 22422.
FIG. 13 is a structure diagram of a block management table 22423.
FIG. 14 is a flow chart illustrating the basic I/O sequence of a storage device 10.
FIG. 15 is an operation flow chart of an I/O processing program executed by a storage controller 100.
FIG. 16 is a detailed flow chart of step S150 in FIG. 15.
FIG. 17 is a structure diagram of a request command 1000 and a reply command 2000 transferred between the storage controller 100 and a flash controller 220.
FIG. 18 is a flow chart showing the details of step S153.
FIG. 19 is a detailed flow chart of step S1533.
FIG. 20 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 1.
FIG. 21 is an operation flow chart of an I/O processing program executed by the storage controller 100 in Embodiment 2.
FIG. 22 shows a detailed flow of step S240.
FIG. 23 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 2.
FIG. 24 is a flow chart showing the processing in which the storage controller 100 requests the flash controller 220 to free a block 232.
FIG. 25 is a structure diagram of the memory 140 in accordance with Embodiment 4.
FIG. 26 is a structure diagram of a segment management table 1426.
FIG. 27 is a structure diagram of an LU management table 1427.
FIG. 28 is a structure diagram of a statistical information management table 1428.
FIG. 29 is a flow chart showing the processing in which the storage controller 100 makes equal the load on each flash box 210.
DESCRIPTION OF EMBODIMENTS Embodiment 1
FIG. 2 is a diagram showing the configuration of a storage system 10000 in accordance with Embodiment 1 of the present invention. The storage system 10000 includes a storage device 10 and a host computer 40. The host computer 40 is a host system that uses the storage device 10, such as an application server.
The storage device 10 includes a storage controller 100 and a flash box unit 200. The storage controller 100 controls the operation of the storage device 10. The flash box unit 200 includes flash memory for storing data.
The host computer 40 is connected to the storage device 10 via an I/F chip 110 of the storage controller 100. A SAN (Storage Area Network) 20, for example, is used as an interface for connecting the storage device 10 and the host computer 40. As the SAN 20, Fibre Channel, SCSI (Small Computer System Interface), iSCSI (internet Small Computer System Interface), USB (Universal Serial Bus), IEEE 1394 bus, SAS (Serial Attached SCSI), or the like can be used, for example. More than one host computer 40 can be connected to the storage device 10. In addition, other types of interfaces can be used instead of the SAN 20.
The host computer 40 has installed thereon control software that issues control commands and the like to the storage device 10. The storage controller 100 of the storage device 10, upon receipt of such control commands and the like, executes its own control software to thereby control the operation of the storage device 10. As the storage device 10 and the host computer 40 communicate via a LAN (Local Area Network) 30, for example, the storage controller 100 has an MC (Network Interface Card) 120 that is an interface for connecting the storage device 10 to the LAN 30. As a computer that manages, controls, or maintains the storage device 10, a different computer other than the host computer 40 can be used. In addition, other types of communication networks can be used instead of the LAN 30.
Further, the storage controller 100 includes ASICs (Application Specific Integrated Circuits) 150, a CPU (Central Processing Unit) 130, memory 140, and an I/F chip 160. The ASICs 150 perform a parity operation and the like when RAID is constructed. The CPU 130 controls the entire storage device 10. The memory 140 has stored therein computer programs and data that are necessary for the CPU 130 to control the storage device 10. The I/F chip 160 connects the storage controller 100 and the flash box unit 200.
The ASICs 150 include a DMA (Direct Memory Access) transfer unit 151. The DMA transfer unit 151 transfers data in accordance with a transfer list stored in the memory 140. It should be noted that the DMA transfer unit 151 may be incorporated in a device other than the ASICs 150, such as the I/ F chip 110 or 160.
Each of the I/F chip 110, the MC 120, the CPU 130, and the I/F chip 160 is connected to the ASICs 150 via a dedicated connection bus 170 such as a PCI (Peripheral Component Interconnect). The memory 140 is connected to the ASICs 150 via a dedicated connection bus 190 such as a DDR3 (Double Data Rate 3).
The storage controller 100 accesses flash memory chips 231 via a flash controller 220 described below. It should be noted that although the storage controller 100 will be hereinafter described as accessing the flash memory chips 231 for the sake of convenience of the description, what actually accesses the flash memory chips 231 is the flash controller 220.
The flash box unit 200 includes a plurality of flash boxes 210. The flash box unit 200 and the storage controller 100 are connected via the I/F chip 160 and a dedicated bus 180 such as Fibre Channel or SAS. Each flash box 210 includes a plurality of flash memory chips 231. Though not shown, the plurality of memory chips 231 constitute an RAID group.
FIG. 3 is a structure diagram of the memory 140. The memory 140 includes a data cache area 141 and a system area 142. The data cache area 141 is an area used to temporarily store user data. The system area 142 has stored therein programs and various tables for controlling the storage device 10.
The data cache area 141 is divided into a plurality of cache segments of a given size so that user data 1411 is stored on a cache segment basis. The user data 1411 is, for example, data read from or written to the storage device 10 by the host computer 40.
The system area 142 has stored therein a program group 1421, an LU management table 1422, a memory management table 1423, a transfer list 1424, and a transfer completion list 1425. The program group 1421 includes various programs executed by the CPU 130. The LU management table 1422 manages the address spaces of LUs (Logical Units) created in the RAID group of the flash memory 231. LU represents a unit for providing a storage area of the storage device 10 to the outside of the storage device 10. The memory management table 1423 manages the cache segments in the data cache area 141. The transfer list 1424 is used for the CPU 130 to instruct the ASICs 150 to execute DMA transfer. The transfer completion list 1425 is used for the ASICs 150 to inform the CPU 130 of the completion of the DMA transfer.
FIG. 4 is a structure diagram of the LU management table 1422 for managing the address spaces of LUs. The LU management table 1422 contains a LUN (Logical Unit Number) 14221, flash box No. 14222, start LBA 14223, and last LBA 14224.
The LUN 14221 indicates the identifier of a created LU, and a unique number is assigned to each LU. The flash box No. 14222 indicates the number of the flash box 210 in which the LU is created. The start LBA 14223 and the last LBA 14224 indicate the location of the LU address space in the flash box 210, that is, each of the start address and the end address of the LU is indicated by an LBA (Logical Block Address).
FIG. 5 is a structure diagram of the memory management table 1423. The memory management table 1423 contains a cache segment No. 14231, LUN 14232, LBA 14233, and length 14234.
The cache segment No. 14231 is the identifier of each cache segment in the data cache area 141, and a unique number is assigned to each cache segment. Each of the LUN 14232, the LBA 14233, and the length 14234 is information for managing user data stored in the cache segment. The LUN 14232 indicates the identification number of an LU in which the relevant data is stored. The LBA 14233 indicates the start address of the data. The length 14234 indicates the length of the data.
FIG. 6 is a structure diagram of the transfer list 1424. The transfer list 1424 contains a list No. 14241, transfer source 14242, transfer destination 14243, cache segment No. 14244, length 14245, LUN 14246, and LBA 14247.
The list No. 14241 is the identifier of each transfer list, and a unique number is assigned to each list. The transfer source 14242 and the transfer destination 14243 are the parameters for managing the data transfer direction. For example, the transfer source indicating “Mem” and the transfer destination indicating “FE (Front End)” means the data transfer from the memory 140 to the I/F chip 110. Meanwhile, the transfer source indicating “Mem” and the transfer destination indicating “BE (Back End)” means the data transfer from the memory 140 to the I/F chip 160. The cache segment No. 14244 and the length 14245 indicate the location of the cache segment in which data to be DMA-transferred is to be stored (or stored) and the size of the data to be transferred, respectively. The LUN 14246 and the LBA 14247 indicate the identification number and the start address of an LU to which the data to be transferred belongs, respectively.
FIG. 7 is a structure diagram of the transfer completion list 1425. The transfer completion list 1425 contains a transfer list No. 14251. The transfer list No. 14251 has stored therein the value of the list No. 14241 whose transfer has been completed. By referring to the transfer completion list 1425, it is possible to check which transfer request has been completed.
FIG. 8 is a configuration diagram of each flash box 210. The flash box 210 includes a flash controller 220 and a memory array 230. The flash controller 220 controls the entire flash box 210. The memory array 230 includes the plurality of flash memory chips 231.
The flash controller 220 includes an I/F chip 221, a CPU 222, ASICs 223, and memory 224.
The I/F chip 221 is used to connect the flash controller 220 and the storage controller 100. The I/F chip 221 and the storage controller 100 are connected via the connecting bus 180 described above. The CPU 222 controls the entire flash box 210. The memory 224 has stored therein computer programs and data that are necessary for the CPU 222 to control the flash box 210. The ASICs 223 provide functions for controlling the flash memory chips 231 such as GC or WL.
The ASICs 223 include a DMA transfer unit 2231 and a memory controller (Mem CTL) 2232. The memory controller 2232 manages the basic operations such as read, write, and erasure performed to the flash memory chips 231. Such functional unit may be incorporated in a device other than the ASICs 223. Alternatively, the memory controller 2232 may be provided separately as a single unit.
The ASICs 223 are connected to each of the CPU 222 and the I/F chip 221 via a dedicated bus 225 such as a PCI. The ASICs 223 are also connected to the memory 224 via a dedicated bus 226 such as a DDR 3. Further, the ASICs 223 are also connected to the flash memory chips 231 with a dedicated bus 227 via the memory controller 2232.
FIG. 9 is a configuration diagram of the flash memory chip 231. The flash memory chip 231 includes a plurality of blocks 232, and the memory is erased per block 232. Each block 232 includes a plurality of pages 233 that are the minimum read/write units of the flash memory chip 231.
FIG. 10 is a diagram illustrating address assignment. The blocks 232 are divided in terms of attributes into a normal block 236 and an alternate block 237. The normal block 232 is a block recognized as an LU space by the storage controller 100. The alternate block 237 is used as a data migration destination when GC or WL is executed or as a replacement block when the normal block has crashed. Such attribute is not specific to each block 232, and is changed as needed in accordance with the usage state of each block 232.
Further, the pages 233 in the normal block 236 are divided into a base area 234 and an update area 235. The base area 234 is an area provided as a real address space to the host system, whereas the update area 235 is an area for, upon request to update pages, storing the target update data. Therefore, in a strict sense, the sum of the physical address spaces of the pages 233 that belong to the base area 234 is provided as a logical address space to the storage controller 100.
Although the logical address space seen from the storage controller 100 is unique, physical blocks that provide a particular logical address space are updated as needed. Therefore, the logical address recognized as an LBA by the storage controller 100 is always the same. Although Embodiment 1 illustrates an example in which the base area 234 and the update area 235 reside in the same block, such areas can be provided in separate blocks.
FIG. 11 is a structure diagram of the memory 224. The memory 224 includes a data cache area 2241 and a system area 2242. The data cache area 2241 is an area used to temporarily store user data. The system area 2242 has stored therein programs and various tables for controlling the flash box 210.
The data cache area 2241 is divided into a plurality of segments of a given size so that user data 22411 is stored on a cache segment basis.
The system area 2242 has stored therein a program group 22421, an address management table 22422, a block management table 22423, and a transfer list 22424. The program group 22421 includes various programs executed by the CPU 222. The address management table 22422 manages the correspondence between the logical address and the physical address. The block management table 22423 manages the status of each block 232. The transfer list 22424 is used when the ASICs 223 execute DMA transfer.
FIG. 12 is a structure diagram of the address management table 22422. The address management table 22422 contains start LBA 224221, chip No. 224222, block No. 224223, and page No. 224224.
The start LBA 224221 indicates the offset value of LBA recognized by the storage controller 100. This value is determined by the size of the page 233 that is the minimum read/write unit of the flash memory chip 231. For example, when each page 233 has an address space of 20 h, the start LBA 224221 has values at intervals of 20 h. The chip No. 224222, the block No. 224223, and the page No. 224224 are information for identifying the page 233 corresponding to the logical address.
FIG. 13 is a structure diagram of the block management table 22423. The block management table 22423 contains a chip No. 224231, block No. 224232, attribute 224233, status 224234, the residual number of valid pages 224235, and the number of erasures 224236.
The chip No. 224231 indicates the number of the flash memory chip 231 to which the block 232 belongs. The block No. 224232 indicates the number of the block in the flash memory chip 231. Each block 232 is identified by the chip No. 224231 and the block No. 224232. The attribute 224233 indicates the attribute of the block 232 described with reference to FIG. 10, and either “Normal” or “Alternate” is set. The status 224234 indicates the operating status of each block 232. For example, “WL” is set when WL is executed to the relevant block, “GC” is set when GC is executed to the relevant block, “Normal” is set when read/write access to the relevant block is allowed, and “n/a” is set for other statuses. The residual number of valid pages 224235 indicates the residual number of the pages 233 that are assigned as the update area 235 within the relevant block 232. It should be noted that if the attribute of the relevant block is “Alternate,” or if the processing of freeing pages such as WL or GC is executed, “n/a” is set. The number of erasures 224236 indicates the number of erase operations performed to the relevant block 232, and is mainly used to manage information about whether the relevant block 232 has been accessed up to its maximum allowable write access limit.
FIG. 14 is a flow chart illustrating the basic I/O sequence of the storage device 10. Hereinafter, each step of FIG. 14 will be described.
(FIG. 14: Step S10)
The storage device 10 receives a data write request or a data read request (an I/O request) from the host computer 40, upon which the storage controller 100 starts the present flow chart.
(FIG. 14: Step S20)
The storage controller 100 determines whether the request received in step S10 is a read (RD) request or a write (WR) request. If the request is determined to be a read request, the flow proceeds to step S30, and if the request is determined to be a write request, the flow proceeds to step S50.
(FIG. 14: Step S30)
The storage controller 100 executes staging processing in which data requested to be read is transferred from the flash box 210 to the memory 140. The details of the present step will be described below with reference to FIGS. 15 and 16.
(FIG. 14: Step S40)
The storage controller 100 transmits the data, which has been transferred from the flash box 210 to the memory 140 in step S30, to the host computer 40. By this step, the processing of the read request issued from the host computer 40 terminates.
(FIG. 14: Step S50)
The storage controller 100 requests the host computer 40 to transfer data that is requested to be written. Then, the storage controller 100 receives the data and stores it into the memory 140.
(FIG. 14: Step S60)
The storage controller 100 informs the host computer 40 of the successful completion of the transfer.
(FIG. 14: Step S70)
The storage controller 100 executes destaging processing in which data retained in the memory 140 is stored into the flash box 210. Although the destaging processing is executed each time a write request is issued in this drawing, the destaging processing can be executed at any given time. In that case, destaging processing is collectively executed after steps S50 to S60 are repeated a plurality of times. Accordingly, the target data to be written can be collectively transferred to the flash box 210 through single destaging processing, whereby the processing efficiency can be expected to be increased.
FIG. 15 is an operation flow chart of an I/O processing program executed by the storage controller 100. This flow chart illustrates a sequence for executing the read processing and the write processing described with reference to FIG. 14. Hereinafter, each step of FIG. 15 will be described. It should be noted that description will be made first of the read processing, and then of the write processing for the sake of convenience of the description.
Embodiment 1: FIG. 15: Read Processing
(FIG. 15: Step S100)
The storage controller 100 determines the presence or absence of an I/O request issued from the host computer 40. If the answer to step S100 is Yes, the flow proceeds to step S110, and if the answer to step S100 is No, the flow proceeds to step S140.
(FIG. 15: Step S110)
The storage controller 100 analyzes the command issued from the host computer 40 and determines whether the type of the request is RD or WR. If the type of the request is determined to be RD, the flow proceeds to step S120, and if the type of the request is determined to be WR, the flow proceeds to step S130. Hereinafter, steps performed when the type of the request is RD will be described.
(FIG. 15: Step S120)
The storage controller 100 creates a data transfer request (a BE transfer request (1)) to execute the staging processing (processing of reading data stored in the flash memory chip 231). Specifically, the storage controller 100 refers to the memory management table 1423 to secure a cache segment in which data read from the flash memory chip 231 is to be stored. Next, the storage controller 100 creates the transfer list 1424. In this case, “BE” is set as the transfer source 14242, “Mem” is set as the transfer destination 14243, and the number of the secured cache segment or the like is set as the cache segment No. 14244. Then, the flow again returns to step S100 so that the storage controller 100 determines the presence or absence of any other I/O requests. Herein, the flow proceeds to step S140 on the assumption that there are no other I/O requests.
(FIG. 15: Step S140)
The storage controller 100 determines the presence or absence of the transfer list 1424 to determine the presence or absence of a data transfer request. If the transfer list 1424 has been created, the storage controller 100 determines a transfer request to be present, and the flow proceeds to step S150. If the transfer list 1424 has not been created, the flow proceeds to step S160.
(FIG. 15: Step S150)
The storage controller 100 executes data transfer processing in accordance with the description of the transfer list 1424. This step will be described in detail with reference to FIG. 16.
(FIG. 15: Step S160)
The storage controller 100 determines the presence or absence of the transfer completion list 1425 to determine if there is any data transfer executed in step S150 that has been completed. If the answer to step S160 is Yes, the flow proceeds to step S180, and if the answer to step S160 is No, the flow proceeds to step S170.
(FIG. 15: Step S170)
If the storage controller 100 continues the operation, the flow returns to step S100 to repeat the same processing, and if the storage controller 100 terminates the operation, the flow ends.
(FIG. 15: Step S180)
The storage controller 100 refers to the transfer list 1424 based on the list No. 14251 whose data transfer processing has been completed as a key, and acquires the values of the transfer source 14242 and the transfer destination 14243. If the transfer source 14242 indicates “BE” and the transfer destination 14243 indicates “Mem,” the flow proceeds to step S190, and if the transfer source 14242 indicates “Mem” and the transfer destination 14243 indicates “FE,” the flow proceeds to step S220.
(FIG. 15: Step S180: Supplement)
If the transfer source 14242 indicates “BE” and the transfer destination 14243 indicates “Mem,” it can be confirmed that the processing of reading data from the flash box 210 into the memory 140 has been completed. In that case, it is further necessary to transfer the data from the memory 140 to the host computer 40. Thus, a data transfer request to that end should be created in step S190. If the transfer source 14242 indicates “Mem” and the transfer destination 14243 indicates “FE,” it can be confirmed that the processing of transferring data from the memory 140 to the host computer 40, that is, the read processing has been completed.
(FIG. 15: Step S190)
The storage controller 100 creates as the transfer list 1424 a data transfer request (an PE transfer request (3)) for transferring to the host computer 40 the data that has been read from the flash box 210 into the memory 140 in step S150. Specifically, “Mem” is set as the transfer source 14242 and “FE” is set as the transfer destination 14243 in the transfer list 1424. Then, the flow again returns to step S160 so that the storage controller 100 determines the presence or absence of any other data transfer that has been completed.
(FIG. 15: Step S220)
The storage controller 100 determines that the processing of transferring the data to the host computer 40 is completed.
Embodiment 1: FIG. 15: Write Processing
(FIG. 15: Steps S100 to S110)
The storage controller 100 executes the same processing as that in steps S100 to S110 described with respect to the read processing. Hereinafter, steps performed when the type of the request is WR will be described.
(FIG. 15: Step S130)
The storage controller 100 creates a data transfer request (an FE transfer request (2)) to execute data receiving processing (processing of receiving the target data to be written from the host computer 40). Specifically, the storage controller 100 refers to the memory management table 1423 to secure a cache segment in which data received from the host computer 40 is to be stored. Then, the storage controller 100 creates the transfer list 1424. In this case, “FE” is set as the transfer source 14242, “Mem” is set as the transfer destination 14243, and the number of the secured cache segment or the like is set as the cache segment No. 14244. Then, the flow again returns to step S100 so that the storage controller 100 determines the presence or absence of any other I/O requests. Herein, the flow proceeds to step S140 on the assumption that there are no other I/O requests.
(FIG. 15: Steps S140 to S170)
These steps are the same as those of the read processing.
(FIG. 15: Step S180)
The storage controller 100 refers to the transfer list 1424 based on the list No. 14251 whose data transfer processing has been completed as a key, and acquires the values of the transfer source 14242 and the transfer destination 14243. If the transfer source 14242 indicates “FE” and the transfer destination 14243 indicates “Mem,” the flow proceeds to step S200, and if the transfer source 14242 indicates “Mem” and the transfer destination 14243 indicates “BE,” the flow proceeds to step S230.
(FIG. 15: Step S200)
The storage controller 100 informs the host computer 40 of the successful completion of the data writing. Although data is actually written in a subsequent step, the storage controller 100 herein is configured to return a response to the host computer 40 upon reception of the target data to be written in order to promptly return a response to the host computer 40.
(FIG. 15: Step S210)
The storage controller 100 creates as the transfer list 1424 a data transfer request (a BE transfer request (4)) for transferring to the flash box 210 the data that has been received from the host computer 40 and stored into the memory 140 in step S150. Specifically, “Mem” is set as the transfer source 14242 and “BE” is set as the transfer destination 14243 in the transfer list 1424. Then, the flow again returns to step S160 so that the storage controller 100 determines the presence or absence of any other data transfer that has been completed.
(FIG. 15: Step S230)
The storage controller 100 determines that the processing of transferring the data to the flash box 210 has been completed.
FIG. 16 is a detailed flow chart of step S150 in FIG. 15. Hereinafter, each step of FIG. 16 will be described.
(FIG. 16: Step S151)
The CPU 130 refers to the transfer list 1424 and determines the list No. 14241 with the lowest number to be the target command to be processed.
(FIG. 16: Step S152)
The CPU 130, with reference to the transfer destination 14243, determines whether the command that has been determined to be the target command to be processed in step S151 is an I/O request to the flash box 210. Specifically, the CPU 130 determines “BE” set as the transfer destination 14243 to be an I/O request to the flash box 210. If the answer to step S152 is Yes, the flow proceeds to step S153, and if the answer to step S152 is No, the flow proceeds to step S155.
(FIG. 16: Step S152: Supplement)
A command that is not an I/O request to the flash box 210 is, for example, a data transfer request to the host computer 40.
(FIG. 16: Step S153)
The CPU 130 queries the flash box 210 as to whether data transfer is executable. This step will be described in detail with reference to FIGS. 17 to 19.
(FIG. 16: Step S154)
The CPU 130, in accordance with the reply from the flash box 210, determines whether data transfer is executable. If data transfer is determined to be executable, the flow proceeds to step S155, and if data transfer is determined to be not executable, the flow proceeds to step S156.
(FIG. 16: Step S155)
The CPU 130 sends the transfer list 1424 to the ASICs 150. The ASICs 150 execute the data transfer described in the transfer list 1424 using the function of the DMA transfer unit 151.
(FIG. 16: Step S156)
The CPU 130, if data transfer is determined to be not executable in step S154, determines that write access to the flash box 210 cannot be executed promptly. Thus, the CPU 130 waits for a predetermined period of time without starting data transfer. With regard to data that is not transferred in this step, the CPU 130 will attempt to transfer such data when the present flow is executed again.
(FIG. 16: Step S157)
The CPU 130, with reference to the transfer list 1424, determines the presence or absence of any other transfer requests that have not been executed yet. If the answer to step S157 is Yes, the flow returns to step S151, and if the answer to step S157 is No, the flow ends.
FIG. 17 is a structure diagram of a request command 1000 and a reply command 2000 exchanged between the storage controller 100 and the flash controller 220 in step S153. Hereinafter, the structure of each command will be described.
The request command 1000 is a command sent from the storage controller 100 to the flash controller 220, and includes a command No. 1001 and request content 1002. Depending on the type of the request content 1002, the request command 1000 may further include additional information. Hereinafter, an example will be described in which the storage controller 100 queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. In that case, the request command 1000 further includes a start address 1003 and write request size 1004.
As the command No. 1001, the value of the list No. 14241 of the transfer list 1424 that is currently processed by the ASICs 150 is set. As the request content 1002, a value indicating a command that causes the request command 1000 to query as to whether data transfer is executable is set. The value of the LBA 14247 is set as the start address 1003, and the value of the length 14245 is set as the write request size 1004.
The reply command 2000 is a command sent from the flash controller 220 to the storage controller 100, and includes information to be returned in response to the request command 1000. Specifically, the reply command 2000 includes a command No. 2001 and reply content 2002.
As the command No. 2001, the value of the command No. 1001 that has been set for the corresponding request command 1000 is set. As the reply content 2002, a value indicating a reply content of the reply command 2000 is set. For example, if the request command 1000 is a command that queries as to whether data transfer is executable, a value indicating that data transfer is executable or not is set as the reply content 2002.
The configurations of the commands are not limited to the examples shown in FIG. 17. In addition, the flash controller 220 can be configured to send the request command 1000 to the storage controller 100, and the storage controller 100 can be configured to send the reply command 2000 to the flash controller 220.
FIG. 18 is a flow chart showing the details of step S153. Hereinafter, each step of FIG. 18 will be described.
(FIG. 18: Step S1530)
The storage controller 100 creates the request command 1000. As described with reference to FIG. 17, the content of the transfer list 1424 that is currently processed by the ASICs 150 is set as each item of the request command 1000.
(FIG. 18: Step S1531)
The storage controller 100 sends the created request command 1000 to the flash controller 220.
(FIG. 18: Step S1532)
The flash controller 220 receives the request command 1000.
(FIG. 18: Step S1533)
The flash controller 220 checks the content of the request command 1000. This step will be described in detail with reference to FIG. 19.
(FIG. 18: Step S1534)
The flash controller 220, based on the result of step S1533, creates the reply command 2000 and sends it to the storage controller 100.
(FIG. 18: Step S1535)
The storage controller 100 receives the reply command 2000 and checks the content, and then terminates the present flow.
FIG. 19 is a detailed flow chart of step S1533. Hereinafter, each step of FIG. 19 will be described.
(FIG. 19: Step S15330)
The flash controller 220 refers to the address management table 22422 to search for the chip No. 224222 and the block No. 224223 in which data specified by the request command 1000 is stored.
(FIG. 19: Step S15331)
The flash controller 220 refers to the status information 224234 in the block management table 22423, and determines whether GC is currently executed to the relevant block 232. If the answer to step S15331 is Yes, the flow proceeds to step S15336, and if the answer to step S15331 is No, the flow proceeds to step S15332.
(FIG. 19: Step S15332)
The flash controller 220 refers to the status information 224234 in the block management table 22423, and determines whether WL is currently executed to the relevant block. If the answer to step S15332 is Yes, the flow proceeds to step S15336, and if the answer to step S15332 is No, the flow proceeds to step S15333.
(FIG. 19: Step S15333)
The flash controller 220 determines whether GC will be executed to the block for which the data transfer request has been received. Specifically, the flash controller 220 determines from the write request size 1004 included in the request command 1000 how many pages 233 in the update area would be necessary to store the data. If the number of the necessary pages 233 is determined to be insufficient, the flash controller 220 determines that GC will be executed, and if not, the flash controller 220 determines that GC will not be executed. If the answer to step S15333 is Yes, the flow proceeds to step S15335, and if the answer to step S15333 is No, the flow proceeds to step S15334.
(FIG. 19: Step S15333: Supplement)
Examples of the determination methods in the present step include a method of calculating the number of necessary pages by dividing the write request size by the size of the page 233. In that case, the flash controller 220, if the quotient is greater than the residual number of valid pages 224235 of the block, determines that GC will be executed, and if the quotient is less than or equal to the residual number of valid pages 224235 of the block, determines that GC will not be executed.
(FIG. 19: Step S15334)
The flash controller 220 determines the data transfer to be executable and terminates the present flow.
(FIG. 19: Step S15335)
The flash controller 220 starts execution of GC to the relevant block.
(FIG. 19: Step S15336)
The flash controller 220 determines the data transfer to be non-executable, and terminates the present flow.
FIG. 20 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 1. Hereinafter, each step of FIG. 20 will be described.
(FIG. 20: Step S3100)
It is assumed that the flash controller 220 is executing GC to the flash memory chip 231.
(FIG. 20: Steps S3101 to S3102)
The storage controller 100 starts the processing of querying the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. At this time, the storage controller 100 sends to the flash controller 220 the request command 1000 to query as to whether data write access to the flash memory chip 231 is allowed.
(FIG. 20: Steps S3103 to S3104)
During these steps, the flash controller 220 is executing GC to the flash memory chip 231. Thus, the flash controller 220 sends to the storage controller 100 the reply command 2000 indicating that data write access to the flash memory chip 231 is not allowed.
(FIG. 20: Step S3105)
The storage controller 100 suspends the data write processing to the flash memory chip 231 and executes another processing while waiting. The wait time herein can be either fixed or determined as needed.
(FIG. 20: Step S3106)
The flash controller 220, after the completion of the GC, enters a state ready to receive a request to write data to the flash memory chip 231.
(FIG. 20: Steps S3107 to S3108)
The storage controller 100 restarts the data write processing that has been suspended in step S3105, and executes the same processing as that in steps S3101 to S3102.
(FIG. 20: Steps S3109 to S3110)
The flash controller 220 sends the reply command 2000 to the storage controller 100 as in steps S3103 to S3104. Herein, as the GC has been completed, the flash controller 220 sends the reply command 2000 indicating that data write access to the flash memory chip 231 is allowed.
(FIG. 20: Steps S3111 to S3112)
The storage controller 100 starts the data write processing to the flash memory chip 231. At this time, the storage controller 100 sends to the flash controller 220 the request command 1000 to write data to the flash memory chip 231.
(FIG. 20: Steps S3113 to S3116)
The flash controller 220 executes data writing to the flash memory chip 231. Upon completion of the data writing, the flash controller 220 sends to the storage controller 100 the reply command 2000 indicating the completion of the data writing. The storage controller 100, upon receipt of the reply command 2000, terminates the data write processing.
According to the sequence of FIG. 20, the storage controller 100 and the flash controller 220 operate cooperatively, whereby the time required for the storage controller 100 to execute destaging can be minimized. Specifically, the time required for the storage controller 100 to execute destaging can be made substantially equal to the time in which data is actually written to the flash memory chip 231. Accordingly, the time in which the storage controller 100 can execute another processing (step S3105) can be increased with the result that the entire performance of the storage system 10000 can be improved.
As described above, according to Embodiment 1, the storage controller 100, before writing data to the flash memory chip 231, queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed using the request command 1000. Accordingly, the storage controller 100 can actually start data writing after the flash memory chip 231 becomes write-accessible, and thus the substantial data write time can be suppressed.
In addition, according to Embodiment 1, the storage controller 100, if it has received the reply command 2000 indicating that data write access to the flash memory chip 231 is not allowed, starts another processing. Accordingly, it becomes possible for the storage controller 100 to continue the processing without wasting the time in which the data write access to the flash memory chip 231 is not allowed, whereby the performance of the storage device 10 can be improved.
Further, according to Embodiment 1, the storage controller 100, having completed another processing after the receipt of the reply command 2000 indicating that data write access to the flash memory chip 231 is not allowed, again queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. Accordingly, even if data write access to the flash memory chip 231 is not allowed at a given moment in time, data writing can successfully be completed without being abandoned.
In Embodiment 1, description has been made of an example in which the storage controller 100 queries the flash controller 220 as to whether data write access to the flash memory chip 231 is allowed. However, it is also possible to configure the storage controller 100 to query as to whether data read access to the flash memory chip 231 is allowed. Accordingly, the storage controller 100, during execution of GC, can execute another processing without issuing a data read request, and can, after the completion of the GC, read the data from the flash memory chip 231.
In Embodiment 1, the flash controller 220 can be configured to execute GC not only when it has received an access request from the storage controller 100 but also at any given time as needed. This is also true of the following embodiments.
In Embodiment 1, when the flash controller 220 sends to the storage controller 100 the reply command 2000 indicating that the flash memory chip 231 is inaccessible, information on the wait time can be included in the reply command 2000. The storage controller 100, after having waited for the wait time described in the reply command 2000, retries data writing. The wait time herein depends on the time required for the flash controller 220 to execute GC. Thus, a value indicating an appropriate wait time can be stored in advance in a storage device such as the memory 224 in accordance with the performance of each device so that the flash controller 220 will read such value and inform the storage controller 100 of the value. Alternatively, a value indicating a wait time can be embedded in a program executed by the flash controller 220.
Embodiment 2
In Embodiment 2 of the present invention, an operation example will be described in which GC is executed in advance before the storage controller 100 issues a data write request to the flash controller 220. Configurations other than that are the same as those described in Embodiment 1.
FIG. 21 is an operation flow chart of an I/O processing program executed by the storage controller 100 in Embodiment 2. This flow differs from that in FIG. 15 in that, in addition to the flow described with reference to FIG. 15, step S240 is provided after step S130. Steps other than that are the same as those in FIG. 15.
FIG. 22 shows a detailed flow of step S240. Hereinafter, each step of FIG. 22 will be described.
(FIG. 22: Step S241)
The storage controller 100 determines whether the currently processed I/O request is a sequential write request. If the answer to step S241 is Yes, the flow proceeds to step S242, and if the answer to step S241 is No, the flow ends.
(FIG. 22: Step S241: Supplement 1)
The reason for determining whether the I/O request is a sequential write request in this step is that sequential write commands are typically write commands that are executed in writing large-size data.
(FIG. 22: Step S241: Supplement 2)
Among methods of determining if an I/O request is a sequential write request is a method of determining if addresses of data to be written are consecutive. Specifically, the storage controller 100, with reference to the memory management table 1423, determines whether data, which has an immediately preceding address of the address (LBA) contained in the I/O request received from the host computer 40, resides in the memory 140. If such data is determined to reside in the memory 140, the storage controller 100 determines the relevant I/O request to be a sequential write request.
(FIG. 22: Step S242)
The storage controller 100 creates the request command 1000 to execute GC, and sends it to the flash controller 220.
FIG. 23 is a diagram showing a schematic sequence in which the storage controller 100 accesses the flash memory chip 231 in Embodiment 2. Hereinafter, each step of FIG. 23 will be described.
(FIG. 23: Steps S3201 to S3205)
It is assumed that the storage controller 100 has received a sequential write request from the host computer 40. The storage controller 100 sends the request command 1000 to start execution of GC to the flash controller 220. Then, the storage controller 100 suspends the data write processing to the flash memory chip 231 and executes another processing while waiting.
(FIG. 23: Steps S3206 to S3208)
The flash controller 220 executes GC to the flash memory chip 231. Upon completion of the GC, the flash controller 220 enters a state ready to receive a request to write data to the flash memory chip 231.
(FIG. 23: Steps S3209 to S3218)
These steps are the same as steps S3107 to S3116 in FIG. 20
According to the sequence of FIG. 23, the storage controller 100, before starting destaging processing to the flash controller 220, requests the flash controller 220 to start GC. Accordingly, sufficient empty pages can be created in the flash memory chip 231 before data writing is actually started. Thus, even when processing of destaging large-size data is subsequently generated along with a sequential write request, the flash controller 220 can promptly store such data without executing GC to the flash memory chip 231.
In Embodiment 2, the storage controller 100 is configured to, upon receipt of a sequential write request from the host computer 40, determine that processing of writing large-size data will be generated. However, if it is possible to know in advance that the size of data to be written is large through the use of other means, the storage controller 100 need not necessarily determine the necessity of the execution of GC based on whether a received I/O request is a sequential write request or not. For example, if the storage controller 100 is informed of a data size together with a data write request from the host computer 40, it can determine the necessity of the execution of GC based on the data size.
In addition, in Embodiment 2, the storage controller 100 is configured to issue the request command 1000 to execute GC in advance to the flash controller 220. Alternatively, the storage controller 100 can be configured to send a data write request directly to the flash controller 220 so that the flash controller 220 determines the necessity of the execution of GC. In that case, the flash controller 220, in order to execute GC, sends to the storage controller 100 the reply command 2000 indicating that data write access is not allowed, and starts the GC. The subsequent operations are the same as those in Embodiment 2.
Embodiment 3
In Embodiment 2, description has been made of an example in which GC is executed in advance before the storage controller 100 writes data to the flash memory chip 231. In order to execute GC, there should be empty alternate blocks 237 available in order that data in the flash memory chip 231 can be once saved to the alternate blocks 237.
However, if all of the alternate blocks 237 have already been used for WL, GC, or the like for the other blocks 232, a shortage of the alternate blocks 237 could occur, and thus GC or WL cannot newly be executed. In that case, even if GC is attempted to be executed in advance before large-size data is written as described in Embodiment 2, it would be impossible, unless WL or GC for the other blocks 232 is completed and a new alternate block 237 is thus produced, to execute the GC to the relevant block.
Thus, in Embodiment 3 of the present invention, when an area of the flash memory chip 231 that has been used so far becomes no longer used, for example, at the timing when an LU is deleted, the storage controller 100 instructs the flash controller 220 to free the unused area. Accordingly, alternate blocks 237 that are available in the flash memory chip 231 can be secured.
In the conventional storage devices, a storage controller frees an LU that is a logical address space at the time of deleting the LU, but does not execute the processing of freeing a physical area like the format processing performed to HDDs. That is, a storage area appears to be freed when an LU is deleted, but in reality, data keeps on residing in the physical area.
In the storage device with the flash memory chips 231, in order to rewrite the page 233 to which data has once been written, it is necessary that the data written in the page 233 be physically deleted. Therefore, when the storage controller 100 deletes only an LU, a storage area is not actually freed. Thus, the number of available alternate blocks 237 would gradually run short, whereby the aforementioned problem could occur. Embodiment 3 aims to address such a problem.
FIG. 24 is a flow chart showing the processing in which the storage controller 100, after deleting an LU, requests the flash controller 220 to free the block 232. Hereinafter, each step of FIG. 24 will be described.
(FIG. 24: Step S300)
The storage controller 100 executes the present processing flow during or after the processing of deleting an LU.
(FIG. 24: Step S301)
The storage controller 100 issues the request command 1000 to request the flash controller 220 to free the block 232 that has been assigned to the deleted LU. For each parameter of the request command 1000, the start LBA, the last LBA, or the like of the area to be freed is set. Such a range to be freed can be identified by referring to the start LBA 14223 and the last LBA 14224 of the deleted LU on the LU management table 1422.
(FIG. 24: Step S302)
The flash controller 220 receives the request command 1000.
(FIG. 24: Step S303)
The flash controller 220 frees the requested block 232 in accordance with the content of the request command 1000. Specifically, the flash controller 220 identifies the block 232 to be freed with reference to the address management table 22422, and issues an erase command to the block 232. In addition, the flash controller 220 changes the attribute 224233 of the block in the block management table 22423 to “Alternate.”
(FIG. 24: Step S304)
The flash controller 220 creates the reply command 2000 to inform the storage controller 100 of the completion of the processing of freeing the block 232, and sends it to the storage controller 100.
(FIG. 24: Step S305)
The storage controller 100 receives the reply command 2000 from the flash controller 220, and terminates the present processing flow.
As described above, according to Embodiment 3, the processing of freeing a storage area of the flash memory chip 231 is executed at the point when the storage controller 100 has deleted an LU. Accordingly, it is possible to reduce the possibility that a shortage of the alternate blocks 237 may occur when the flash controller 220 attempts to execute GC, in which case the GC cannot be executed. In particular, the frequency of execution of GC will increase under the operating conditions that GC is executed in advance before large-size data is written as described in Embodiment 2. In such a case, the advantageous effect of the method described in Embodiment 3 can be maximized.
Embodiment 4
In Embodiment 4 of the present invention, a method of securing available alternate blocks 237 when the storage device 10 uses a volume virtualization technique will be described. The reason that the available alternate blocks 237 should be secured is the same as that described in Embodiment 3.
The volume virtualization technique is a technique of presenting a virtual volume, which is larger than the physical volume, of the storage device 10 to the host computer 40 in advance, and adding a storage capacity only after the virtual volume is actually needed. JP Patent Publication (Kokai) No. 2007-66259 A discloses an example of a storage device that uses a volume virtualization technique.
In Embodiment 4, a storage area that is assigned when data is written on the storage device 10 with a virtualized storage capacity is referred to as a “segment” in the present invention. JP Patent Publication (Kokai) No. 2007-66259 A discloses, as a means of improving the performance of a storage device that uses a volume virtualization technique, migrating a segment (the term “segment” is also used in this document) from one storage device to another.
When a segment is migrated from one storage device to another, data in the migration-source segment is not physically freed. Thus, freeing the block 232 that constitutes the migration-source segment as in Embodiment 3 will allow the number of the available alternate blocks 237 in the flash memory chip 231 to be increased.
FIG. 25 is a structure diagram of the memory 140 in accordance with Embodiment 4. This structure differs from that described with reference to FIG. 3 in Embodiment 1 in that a segment management table 1426, an LU management table 1427, and a statistical information management table 1428 are newly provided instead of the LU management table 1422. Structures other than that of the memory 140 are the same as those in Embodiments 1 to 3.
The segment management table 1426 manages the structures of segments. The LU management table 1427 manages the structures of LUs. The statistical information management table 1428 has recorded thereon the statistical information such as the number of commands issued to each segment.
FIG. 26 is a structure diagram of the segment management table 1426. The segment management table 1426 contains a segment No. 14261, flash box No. 14262, start LBA 14263, and last LBA 14264.
The segment No. 14261 is the identifier of a created segment. The flash box No. 14262 is the identification number of the flash box 210 in which the segment is created. The start LBA 14263 and the last LBA 14264 indicate the location of the segment's address space in the flash box 210, that is, each of the start address and the end address of the segment is indicated by an LBA.
FIG. 27 is a structure diagram of the LU management table 1427. The LU management table 1427 is provided in a number corresponding to the number of created LUs. The LU management table 1427 contains a LUN 14271, LU size 14272, and segment No. 14273.
The LUN 14271 is the identifier of an LU. The LU size 14272 indicates the size of the LU in units of GB (GigaByte). The segment No. 14273 indicates the identification number of a segment assigned to the LU, and is provided in a number corresponding to the LU size.
In the volume virtualization technique, a virtual logical unit that is a physical storage area is assigned only after data is actually stored. Therefore, in the segment No. 14273, “n/a” that indicates “not assigned” is set for an area that is not assigned a segment, whereas for an area that has been assigned a segment, the identification number of the segment is set.
FIG. 28 is a structure diagram of the statistical information management table 1428. The statistical information management table 1428 contains a segment No. 14281, the number of write commands 14282, and the number of read commands 14283.
The segment No. 14281 is the identifier of a segment. The number of write commands 14282 and the number of read commands 14283 indicate the number of write I/O requests and the number of read I/O requests, respectively, that have been issued to the segment from the host computer 40.
The storage controller 100, after migrating a segment or deleting an LU, requests the flash controller 220 to free the block 232 corresponding to the segment. This step is the same as that shown in FIG. 24, but differs in the method of identifying the target block 232 to be freed.
When migrating a segment, the storage controller 100 identifies from the segment management table 1426 the start LBA 14263 and the last LBA 14264 of the segment, based on the segment No. 14261 of the migration source as a key.
When deleting an LU, the storage controller 100 first refers to the LU management table 1427 based on the LUN 14271 as a key, and identifies the segment No. 14273 that constitutes the deleted LU. Next, the storage controller 100 refers to the segment management table 1426 based on the segment No. as a key, and identifies the start LBA 14263 and the last LBA 14264 of the segment.
As described above, according to Embodiment 4, the processing of freeing a storage area of the flash memory chip 231 is executed at the point when the storage controller 100 has migrated a segment or deleted an LU. Accordingly, even when the storage device 10 employs a volume virtualization technique, an advantageous effect similar to that in Embodiment 3 can be exerted.
It should be noted that in Embodiment 4, even when the storage controller 100 has deleted a segment, a storage area of the flash memory chip 231 can be freed in a similar way.
Embodiment 5
In Embodiment 5 of the present invention, a method of evenly distributing the frequency of write I/O access or the number of data erasure cycles across each of the plurality of flash boxes 210 will be described. Although the configuration of Embodiment 5 is based on the configuration described in Embodiment 4 that uses a volume virtualization technique, a similar method can also be used under the configuration described in any of Embodiments 1 to 3.
When only a particular flash box 210 is write I/O accessed with high frequency, the frequency of GC or WL executed to the flash box 210 would correspondingly increase, and thus the number of erasure cycles of such a flash memory chip 231 would also increase. However, as the flash memory chip 231 is a device with a finite number of erasure cycles, if the erasure cycles of a particular block 232 have exceeded a threshold limit, resulting in a crash, such a crashed block 232 is replaced with the alternate block 237. Thus, the number of the available alternate blocks 237 in the flash box 210 that is write I/O accessed with high frequency would easily decrease. The resulting shortage of the available alternate blocks 237 can lead to a decrease in the write processing performance.
Thus, in Embodiment 5, the storage controller 100, based on statistical information such as the frequency of write I/O access to or the number of erasure cycles of each flash box 210, migrates segments between the flash boxes 210. Accordingly, the frequency of write I/O access, the number of erasure cycles, and the like can be distributed evenly across each of the plurality of flash boxes 210, so that the system performance can be maximized and the lifetime of each flash box 210 can be extended.
To that end, in Embodiment 5, the storage controller 100 and the flash controller 220 exchange statistical information such as the number of data erasure cycles of each flash box 210.
FIG. 29 is a flow chart showing the processing in which the storage controller 100 makes equal the load on each flash box 210. Hereinafter, each step of FIG. 29 will be described.
(FIG. 29: Step S400)
The storage controller 100 creates the request command 100 to request the flash controller 220 to send statistical information on the flash boxes 210, and sends it to the flash controller 220.
(FIG. 29: Step S401)
The flash controller 220 receives the request command 1000.
(FIG. 29: Step S402)
The flash controller 220 acquires the statistical information on the flash boxes 210 based on the content of the request command 1000. Specifically, the flash controller 220 first acquires the status of each block 232 with reference to the block management table 22423. Next, the flash controller 220 acquires statistical information such as the sum of the numbers of erasure cycles of all of the blocks 232 from the statistical information management table 1428.
(FIG. 29: Step S403)
The flash controller 220 creates the reply command 2000 corresponding to the request content of the request command 1000, and sends it to the storage controller 100.
(FIG. 29: Step S403: Supplement)
As the statistical information sent from the flash controller 220 to the storage controller 100 in this step, the following information can be considered: the numbers of the assigned normal blocks 236 and alternate blocks 237, the number of the blocks 232 to which GC or WL is executed, the residual number of valid pages, the number of data erasure cycles, and the like. The flash controller 220 is configured to be capable of sending a variety of statistical information on the flash boxes 210 in response to a request from the storage controller 100.
(FIG. 29: Step S404)
The storage controller 100 receives the reply command 2000 from the flash controller 220.
(FIG. 29: Step S405)
The storage controller 100 determines the target segments to be exchanged. For example, the storage controller 100 compares the number of data erasure cycles of each flash box 210 acquired from the flash controller 220 with the number of write commands 14282 in the statistical information management table 1428, and determines a segment with the greatest number of write commands to be allocated to the flash box 210 with the smallest number of data erasure cycles.
(FIG. 29: Step S406)
The storage controller 100 exchanges the segments based on the result of step S405. Accordingly, the frequency of access to each segment can be made equal and I/O processing can be distributed evenly across each flash box 210.
As described above, according to Embodiment 5, the storage controller 100 can acquire statistical information on the flash boxes 210 from the flash controller 220, and thus can make equal the load on each flash box 210 based on the acquired information. Accordingly, the number of data write cycles or data erasure cycles can be distributed evenly not only within each flash memory chip 231 but also across each flash box 210.
REFERENCE SIGNS LIST
    • 10 Storage Device
    • 100 Storage Controller
    • 110 I/F Chip
    • 120 NIC
    • 130 CPU
    • 140 Memory
    • 141 Data Cache Area
    • 1411 User Data
    • 142 System Area
    • 1421 Program Group
    • 1422 LU Management Table
    • 14221 LUN
    • 14222 Flash Box No.
    • 14223 Start LBA
    • 14224 Last LBA
    • 1423 Memory Management Table
    • 14231 Cache Segment No.
    • 14232 LUN
    • 14233 LBA
    • 14234 Length
    • 1424 Transfer List
    • 14241 List No.
    • 14242 Transfer Source
    • 14243 Transfer Destination
    • 14244 Cache Segment No.
    • 14245 Length
    • 14246 LUN
    • 14247 LBA
    • 1425 Transfer Completion List
    • 14251 Transfer List No.
    • 1426 Segment Management Table
    • 14261 Segment No.
    • 14262 Flash Box No.
    • 14263 Start LBA
    • 14264 Last LBA
    • 1427 LU Management Table
    • 14271 LUN
    • 14272 LU Size
    • 14273 Segment No.
    • 1428 Statistical Information Management Table
    • 14281 Segment No.
    • 14282 Number of Write Commands
    • 14283 Number of Read Commands
    • 150 ASICs
    • 151 DMA Transfer Unit
    • 160 I/F Chip
    • 170 Connection Bus
    • 180 Dedicated Bus
    • 190 Connection Bus
    • 200 Flash Box Unit
    • 210 Flash Box
    • 220 Flash Controller
    • 221 I/F Chip
    • 222 CPU
    • 223 ASICs
    • 2231 DMA Transfer Unit
    • 2232 Memory Controller
    • 224 Memory
    • 2241 Data Cache Area
    • 22411 User Data
    • 2242 System Area
    • 22421 Program Group
    • 22422 Address Management Table
    • 224221 Start LBA
    • 224222 Chip No.
    • 224223 Block No.
    • 224224 Page No.
    • 22423 Block Management Table
    • 224231 Chip No.
    • 224232 Block No.
    • 224233 Attribute
    • 224234 Status
    • 224235 Residual Number of Valid Pages
    • 224236 Number of Erasures
    • 22424 Transfer List
    • 225 Dedicated Bus
    • 226 Dedicated Bus
    • 227 Dedicated Bus
    • 230 Memory Array
    • 231 Flash Memory Chip
    • 232 Block
    • 233 Page
    • 234 Base Area
    • 235 Update Area
    • 236 Normal Block
    • 237 Alternate Block
    • 20 SAN
    • 30 LAN
    • 40 Host Computer
    • 1000 Request Command
    • 1001 Command No.
    • 1002 Request Content
    • 1003 Start Address
    • 1004 Write Request Size
    • 2000 Reply Command
    • 10000 Storage System

Claims (10)

1. A storage device comprising:
a plurality of flash boxes, each of the plurality of flash boxes including a plurality of flash memory chips for storing data and a flash controller configured to control an operation of the plurality of flash memory chips; and
a storage controller configured to send a read/write command to the plurality of flash memory boxes,
wherein the storage controller is configured to send a querying command before sending the read/write command to a first flash box of the plurality of flash boxes,
wherein, in response to receiving the querying command, a first flash controller of the first flash box determines whether the first flash controller can process the read/write command during execution of an internal process or not so that the first flash controller determines whether or not the first flash controller can execute a subsequent read/write command, and sends a reply command indicating whether or not permission is granted to send the read/write command based on the determination, and
wherein the storage controller waits to send the read/write command to the first flash box until the reply command indicating permission to send the read/write command is received.
2. The storage device according to claim 1,
wherein the storage controller sends the read/write command to the first flash box upon receipt, from the first flash controller, of the reply command indicating permission to send the read/write command to the first flash box, and
wherein the storage controller executes another processing upon receipt of a reply from the first flash controller that does not indicate permission to send the read/write command to the flash box.
3. The storage device according to claim 2, wherein the storage controller, after executing the another processing upon receipt of the reply from the flash controller, queries the flash controller again as to whether the first flash controller can process the read/write command.
4. The storage device according to claim 1,
wherein the first flash controller executes garbage collection for securing a sufficient capacity in the flash memory chips to be written with data, and
wherein the first flash controller, upon receipt of a query, during execution of the garbage collection, from the storage controller as to whether the first flash controller can process the read/write command, sends a reply to the storage controller indicating that the first flash controller cannot process the read/write command.
5. The storage device according to claim 4,
wherein the storage controller, before issuing a request to write data with a size larger than a given size or to execute sequential writes to the flash memory chips, sends a request to the first flash controller to execute the garbage collection, and
wherein the first flash controller, upon receipt of the request, starts execution of the garbage collection to the flash memory chips.
6. The storage device according to claim 4, wherein the flash controller, upon receipt of a request from the storage controller to write data with a size larger than a given size or to execute sequential writes to the flash memory chips, executes the garbage collection before starting processing of writing the data with a size larger than the given size to the flash memory chips.
7. The storage device according to claim 1,
wherein the storage controller, when deleting a logical unit assigned to a storage area in the flash memory, sends an instruction to the first flash controller to free the storage area in the flash memory chips corresponding to the logical unit, and
wherein the first flash controller, upon receipt of the instruction from the storage controller, frees the storage area in the flash memory chips corresponding to the logical unit.
8. The storage device according to claim 1,
wherein the storage controller assigns a segment that is a virtual storage area to a storage area in the flash memory chips,
wherein the storage controller, when migrating or deleting the segment, sends an instruction to the first flash controller to free the storage area in the flash memory chips from which the segment has been migrated or deleted, and
wherein the first flash controller, upon receipt of the instruction from the storage controller, frees the storage area in the flash memory chips from which the segment has been migrated or deleted.
9. The storage device according to claim 1,
wherein the first flash controller informs the storage controller of the frequency of write access to each of the flash memory chips, and
wherein the storage controller evenly distributes the frequency of write access to the flash memory chips across each of the flash boxes.
10. The storage device according to claim 1,
wherein the flash controller informs the storage controller of the frequency of write access to each of the flash memory chips,
wherein the storage controller assigns a segment that is a virtual storage area to a storage area in the flash memory chips, and
wherein the storage controller assigns the segment such that the frequency of write access to the flash memory chips is distributed evenly across each of the flash boxes.
US12/742,783 2010-04-12 2010-04-12 Performance management of access to flash memory in a storage device Active 2030-12-15 US8392670B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/760,144 US8924659B2 (en) 2010-04-12 2013-02-06 Performance improvement in flash memory accesses

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2010/002639 WO2011128928A1 (en) 2010-04-12 2010-04-12 Storage device

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2010/002639 A-371-Of-International WO2011128928A1 (en) 2010-04-12 2010-04-12 Storage device

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/760,144 Continuation US8924659B2 (en) 2010-04-12 2013-02-06 Performance improvement in flash memory accesses

Publications (2)

Publication Number Publication Date
US20110283046A1 US20110283046A1 (en) 2011-11-17
US8392670B2 true US8392670B2 (en) 2013-03-05

Family

ID=43431936

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/742,783 Active 2030-12-15 US8392670B2 (en) 2010-04-12 2010-04-12 Performance management of access to flash memory in a storage device
US13/760,144 Expired - Fee Related US8924659B2 (en) 2010-04-12 2013-02-06 Performance improvement in flash memory accesses

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/760,144 Expired - Fee Related US8924659B2 (en) 2010-04-12 2013-02-06 Performance improvement in flash memory accesses

Country Status (3)

Country Link
US (2) US8392670B2 (en)
JP (1) JP5646633B2 (en)
WO (1) WO2011128928A1 (en)

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8392670B2 (en) 2010-04-12 2013-03-05 Hitachi, Ltd. Performance management of access to flash memory in a storage device
JP2012033002A (en) * 2010-07-30 2012-02-16 Toshiba Corp Memory management device and memory management method
US9015426B2 (en) * 2011-02-28 2015-04-21 Kove Corporation High performance data storage using observable client-side memory access
WO2012168962A1 (en) 2011-06-07 2012-12-13 Hitachi, Ltd. Storage system comprising flash memory, and storage control method
US10445229B1 (en) * 2013-01-28 2019-10-15 Radian Memory Systems, Inc. Memory controller with at least one address segment defined for which data is striped across flash memory dies, with a common address offset being used to obtain physical addresses for the data in each of the dies
US9652376B2 (en) 2013-01-28 2017-05-16 Radian Memory Systems, Inc. Cooperative flash memory control
WO2014141411A1 (en) 2013-03-13 2014-09-18 株式会社日立製作所 Storage system and method for controlling storage system
US9665286B2 (en) 2013-05-17 2017-05-30 Hitachi, Ltd. Storage device
TWI582594B (en) * 2014-01-22 2017-05-11 群聯電子股份有限公司 Method for preventing losing data, memory storage device and memory control circuit unit
US9632953B2 (en) * 2014-06-03 2017-04-25 Qualcomm Incorporated Providing input/output virtualization (IOV) by mapping transfer requests to shared transfer requests lists by IOV host controllers
US9690720B2 (en) 2014-06-03 2017-06-27 Qualcomm Incorporated Providing command trapping using a request filter circuit in an input/output virtualization (IOV) host controller (HC) (IOV-HC) of a flash-memory-based storage device
US9881680B2 (en) 2014-06-03 2018-01-30 Qualcomm Incorporated Multi-host power controller (MHPC) of a flash-memory-based storage device
US9947386B2 (en) * 2014-09-21 2018-04-17 Advanced Micro Devices, Inc. Thermal aware data placement and compute dispatch in a memory system
US10884924B2 (en) 2015-03-04 2021-01-05 Hitachi, Ltd. Storage system and data writing control method
US10678458B2 (en) * 2018-02-09 2020-06-09 Micron Technology, Inc. Data storage device idle time processing
US10489085B2 (en) 2018-02-28 2019-11-26 Micron Technology, Inc. Latency-based scheduling of command processing in data storage devices
JP2022120544A (en) * 2021-02-05 2022-08-18 東芝テック株式会社 Memory control device and program

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6449625B1 (en) * 1999-04-20 2002-09-10 Lucent Technologies Inc. Use of a two-way stack approach to optimize flash memory management for embedded database systems
US20050132103A1 (en) 2003-12-16 2005-06-16 Ikuya Yagisawa Disk array system and interface converter
JP2007066259A (en) 2005-09-02 2007-03-15 Hitachi Ltd Computer system, storage system and volume capacity expansion method
WO2007081638A2 (en) 2005-12-21 2007-07-19 Sandisk Corporation Non-volatile memories and methods with adaptive file handling in a directly mapped file storage system
US20070260837A1 (en) 2006-05-08 2007-11-08 Apacer Technology Inc. Method of dynamic memory management for a portable data storage device
US20080229003A1 (en) 2007-03-15 2008-09-18 Nagamasa Mizushima Storage system and method of preventing deterioration of write performance in storage system
US20090168525A1 (en) 2007-12-27 2009-07-02 Pliant Technology, Inc. Flash memory controller having reduced pinout
US20120150811A1 (en) * 2010-12-14 2012-06-14 International Business Machines Corporation Flash-copying within asynchronous mirroring environment

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS63147219A (en) * 1986-12-10 1988-06-20 Fujitsu Ltd Semiconductor disk device
US8027194B2 (en) * 1988-06-13 2011-09-27 Samsung Electronics Co., Ltd. Memory system and method of accessing a semiconductor memory device
US5603001A (en) 1994-05-09 1997-02-11 Kabushiki Kaisha Toshiba Semiconductor disk system having a plurality of flash memories
JPH07302176A (en) * 1994-05-09 1995-11-14 Toshiba Corp Semiconductor disk device
JP2000105734A (en) * 1998-07-29 2000-04-11 Fujitsu Ltd Method for controlling memory, memory device, and controller
JP2002366380A (en) * 2001-06-13 2002-12-20 Canon Inc Electronic device and its control method
GB0123416D0 (en) * 2001-09-28 2001-11-21 Memquest Ltd Non-volatile memory control
JP4280055B2 (en) * 2001-11-28 2009-06-17 株式会社Access Memory control method and apparatus
US6721820B2 (en) * 2002-05-15 2004-04-13 M-Systems Flash Disk Pioneers Ltd. Method for improving performance of a flash-based storage system using specialized flash controllers
EP1838443B1 (en) 2005-01-20 2012-05-16 3M Innovative Properties Company Structured polymer films and methods for forming the same
US7315917B2 (en) * 2005-01-20 2008-01-01 Sandisk Corporation Scheduling of housekeeping operations in flash memory systems
JP4863749B2 (en) * 2006-03-29 2012-01-25 株式会社日立製作所 Storage device using flash memory, erase number leveling method thereof, and erase number level program
JP5238235B2 (en) * 2007-12-07 2013-07-17 株式会社日立製作所 Management apparatus and management method
US7873619B1 (en) * 2008-03-31 2011-01-18 Emc Corporation Managing metadata
JP5489434B2 (en) * 2008-08-25 2014-05-14 株式会社日立製作所 Storage device with flash memory
US8392670B2 (en) 2010-04-12 2013-03-05 Hitachi, Ltd. Performance management of access to flash memory in a storage device

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6449625B1 (en) * 1999-04-20 2002-09-10 Lucent Technologies Inc. Use of a two-way stack approach to optimize flash memory management for embedded database systems
US20050132103A1 (en) 2003-12-16 2005-06-16 Ikuya Yagisawa Disk array system and interface converter
JP2005182104A (en) 2003-12-16 2005-07-07 Hitachi Ltd Disk array system and interface converting device
JP2007066259A (en) 2005-09-02 2007-03-15 Hitachi Ltd Computer system, storage system and volume capacity expansion method
WO2007081638A2 (en) 2005-12-21 2007-07-19 Sandisk Corporation Non-volatile memories and methods with adaptive file handling in a directly mapped file storage system
US20070260837A1 (en) 2006-05-08 2007-11-08 Apacer Technology Inc. Method of dynamic memory management for a portable data storage device
US20080229003A1 (en) 2007-03-15 2008-09-18 Nagamasa Mizushima Storage system and method of preventing deterioration of write performance in storage system
US20090168525A1 (en) 2007-12-27 2009-07-02 Pliant Technology, Inc. Flash memory controller having reduced pinout
US20120150811A1 (en) * 2010-12-14 2012-06-14 International Business Machines Corporation Flash-copying within asynchronous mirroring environment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
International Search Report in International Application No. PCT/JP2010/002639 mailed Jan. 31, 2011.
Written Opinion in International Application No. PCT/JP2010/002639 mailed Jan. 31, 2011.

Also Published As

Publication number Publication date
US8924659B2 (en) 2014-12-30
WO2011128928A1 (en) 2011-10-20
US20110283046A1 (en) 2011-11-17
US20130151762A1 (en) 2013-06-13
JP2013515979A (en) 2013-05-09
JP5646633B2 (en) 2014-12-24

Similar Documents

Publication Publication Date Title
US8392670B2 (en) Performance management of access to flash memory in a storage device
JP5075761B2 (en) Storage device using flash memory
US7574577B2 (en) Storage system, storage extent release method and storage apparatus
US9501231B2 (en) Storage system and storage control method
CN107908571B (en) Data writing method, flash memory device and storage equipment
US9116622B2 (en) Storage system having nonvolatile semiconductor storage device with nonvolatile semiconductor memory
US10795599B2 (en) Data migration method, host and solid state disk
US20150254186A1 (en) Information processing device having a plurality of types of memory caches with different characteristics
US20120110259A1 (en) Tiered data storage system with data management and method of operation thereof
US20130159651A1 (en) Storage system and storage migration method
JP2009043030A (en) Storage system
KR20150105323A (en) Method and system for data storage
US8694563B1 (en) Space recovery for thin-provisioned storage volumes
JP2015517697A (en) Storage system and storage control method using storage area based on secondary storage as cache area
KR20120050891A (en) Latency reduction associated with a response to a request in a storage system
US8862819B2 (en) Log structure array
US10049042B2 (en) Storage device, semiconductor memory device, and method for controlling same
US20190243758A1 (en) Storage control device and storage control method
US10664193B2 (en) Storage system for improved efficiency of parity generation and minimized processor load
US8799573B2 (en) Storage system and its logical unit management method
US8572347B2 (en) Storage apparatus and method of controlling storage apparatus
US11474750B2 (en) Storage control apparatus and storage medium
WO2014147786A1 (en) Storage system and data management method
WO2018142622A1 (en) Computer
JP5597266B2 (en) Storage system

Legal Events

Date Code Title Description
AS Assignment

Owner name: HITACHI, LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KOSEKI, HIDEYUKI;REEL/FRAME:024440/0664

Effective date: 20100401

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY