WO2008082584A2 - System and method for symmetric triple parity - Google Patents

System and method for symmetric triple parity Download PDF

Info

Publication number
WO2008082584A2
WO2008082584A2 PCT/US2007/026373 US2007026373W WO2008082584A2 WO 2008082584 A2 WO2008082584 A2 WO 2008082584A2 US 2007026373 W US2007026373 W US 2007026373W WO 2008082584 A2 WO2008082584 A2 WO 2008082584A2
Authority
WO
WIPO (PCT)
Prior art keywords
parity
diagonal
blocks
storage
devices
Prior art date
Application number
PCT/US2007/026373
Other languages
French (fr)
Other versions
WO2008082584A3 (en
Inventor
Atul Goel
Original Assignee
Network Appliance. Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Network Appliance. Inc. filed Critical Network Appliance. Inc.
Priority to EP07868052A priority Critical patent/EP2122467A2/en
Publication of WO2008082584A2 publication Critical patent/WO2008082584A2/en
Publication of WO2008082584A3 publication Critical patent/WO2008082584A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1076Parity data used in redundant arrays of independent storages, e.g. in RAID systems

Definitions

  • the present invention relates to arrays of storage systems and, more specifically, to a technique for symmetric parity computation and reconstruction to tolerate up to three failing storage devices of a storage array.
  • a storage system typically comprises one or more storage devices into which data may be entered, and from which data may be obtained, as desired.
  • the storage system may be implemented in accordance with a variety of storage architectures including, but not limited to, a network-attached storage environment, a storage area network and a disk assembly directly attached to a client or host computer.
  • the storage devices are typically disk drives, wherein the term “disk” commonly describes a self-contained rotating magnetic media storage device.
  • the term “disk” in this context is synonymous with hard disk drive (HDD) or direct access storage device (DASD).
  • HDD hard disk drive
  • DASD direct access storage device
  • the disks within a storage system are typically organized as one or more groups, wherein each group is operated as a Redundant Array of Independent (or Inexpensive) Disks (RAID).
  • RAID implementations enhance the reliability/integrity of data storage through the redundant writing of data "stripes" across a given number of physical disks in the RAID group, and the appropriate storing of redundant information with re- spect to the striped data.
  • the redundant information enables recovery of data lost when a storage device fails.
  • a goal of a high performance storage system is to make the mean time to data loss (MTTDL) as long as possible, preferably much longer than the expected service life of the system.
  • Data can be lost when one or more disks fail, making it impossible to recover data from the device.
  • Typical schemes to avoid loss of data include mirroring, backup and parity protection. Mirroring is an expensive solution in terms of consumption of storage resources, such as disks. Backup does not protect data modified since the backup was created. Parity schemes are common because they provide a redundant encoding of the data that allows for a single erasure (loss of one disk) with the addition of just one disk drive to the system.
  • Parity protection is used in computer systems to protect against loss of data on a storage device, such as a disk.
  • a parity value may be computed by summing (usually modulo 2) data of a particular word size (usually one bit) across a number of similar disks holding different data and then storing the results on an additional similar disk. That is, parity may be computed on vectors 1-bit wide, composed of bits in corresponding posi- tions on each of the disks. When computed on vectors 1-bit wide, the parity can be either the computed sum or its complement; these are referred to as even and odd parity respectively. Addition and subtraction on 1-bit vectors are both equivalent to exclusive-OR (XOR) logical operations.
  • XOR exclusive-OR
  • the data is then protected against the loss of any one of the disks, or of any portion of the data on any one of the disks. If the disk storing the parity is lost, the parity can be regenerated from the data. If one of the data disks is lost, the data can be regenerated by adding the contents of the surviving data disks together and then subtracting the result from the stored parity.
  • the disks are divided into parity groups, each of which comprises one or more data disks and a parity disk.
  • a parity set is a set of blocks, including several data blocks and one parity block, where the parity block is the XOR of all the data blocks.
  • a parity group is a set of disks from which one or more parity sets are selected.
  • the disk space is divided into stripes, with each stripe containing one block from each disk.
  • the blocks of a stripe are usually at the same locations on each disk in the parity group.
  • all but one block are blocks containing data (“data blocks") and one block is a block containing parity (“parity block”) computed by the XOR of all the data.
  • RAID-4 If the parity blocks are all stored on one disk, thereby providing a single disk that contains all (and only) parity information, a RAID-4 implementation is provided. If the parity blocks are contained within different disks in each stripe, usually in a rotating pattern, then the implementation is RAID-5.
  • RAID Redundant Arrays of Inexpensive Disks
  • the term "encoding” means the computation of a redundancy value over a predetermined subset of data blocks
  • the term “decoding” means the reconstruction of a data or parity block by using a subset of data blocks and redundancy values. If one disk fails in the parity group, the contents of that disk can be decoded (reconstructed) on a spare disk or disks by adding all the contents of the remaining data blocks and subtracting the result from the parity block. Since two's complement addition and subtraction over 1-bit fields are both equivalent to XOR operations, this reconstruc- tion consists of the XOR of all the surviving data and parity blocks. Similarly, if the parity disk is lost, it can be recomputed in the same way from the surviving data.
  • Parity schemes generally provide protection against a single disk failure within a parity group. These schemes can also protect against multiple disk failures as long as each failure occurs within a different parity group. However, if two disks fail concur- rently within a parity group, then an unrecoverable loss of data is suffered. Failure of two disks concurrently within a parity group is a fairly common occurrence, particularly because disks "wear out” and because of environmental factors with respect to the operation of the disks. In this context, the failure of two disks concurrently within a parity group is referred to as a "double failure". A double failure typically arises as a result of a failure of one disk and a subsequent failure of another disk while attempting to recover from the first failure.
  • the recovery or reconstruction time is dependent upon the level of activity of the storage system. That is, during reconstruction of a failed disk, it is possible that the storage system remains "online” and continues to serve requests (from clients or users) to access (i.e., read and/or write) data. If the storage system is busy serving requests, the elapsed time for reconstruction increases.
  • the reconstruction process time also increases as the size and number of disks in the storage system increases, as all of the surviving disks must be read to reconstruct the lost data.
  • the double disk failure rate is proportional to the square of the number of disks in a parity group. However, having small parity groups is expensive, as each parity group requires an entire disk devoted to redundant data.
  • Another failure mode of disks is media read errors, wherein a single block or sector of a disk cannot be read.
  • the unreadable data can be reconstructed if parity is maintained in the storage array.
  • a media read error on another disk in the array will result in lost data.
  • a known double failure correcting parity scheme is an EVENODD XOR-based technique that allows a serial reconstruction of lost (failed) disks.
  • the EVENODD technique is disclosed in an article of IEEE Transactions on Computers, Vol. 44, No. 2, titled EVENODD: An Efficient Scheme for Tolerating Double Disk Failures in RAID Architectures, by Blaum et al, Feb., 1995.
  • a variant of EVENODD is disclosed in U.S.
  • the above-mentioned article and patent are hereby incorporated by reference as though fully set forth herein.
  • asymmetric algorithms imply that each disk is not treated identically.
  • configuration management tasks must know and identify whether a disk is of a particular type, e.g., whether a disk is a parity disk and/or a data disk.
  • a reconstruction technique may involve a plurality of differing algorithms depending on the number of failed data and/or parity disks as well as the type of failed parity disks, e.g., row parity, diagonal parity, etc.
  • the asymmetric nature of these algorithms imposes additional computational complexity when implementing parity-based systems. This additional complexity may be especially noticeable when utiliz- ing embedded systems to implement parity-based computations.
  • a further noted disadvantage of asymmetric parity algorithms is that utilization of floating parity, i.e., parity stored on any of the storage devices within a parity group instead of on one or more dedicated parity storage devices, is not feasible when utilizing asymmetric parity algorithms.
  • floating parity relies on a scheme where some blocks on the newly added disk(s) are re-assigned as parity, converting their old locations within the parity set to data.
  • asymmetric algorithms cannot move/relocate parity blocks on these disks to newly added disks.
  • the present invention comprises a symmetric triple parity (TP) technique that utilizes an algorithm for parity computation is identical to that used for triple reconstruction.
  • the TP technique is preferably used in an array comprising a number/? of storage devices, such as disks, with/? being a prime number.
  • the disks are divided into blocks and the blocks are organized into stripes.
  • the blocks of rows selected to form a stripe are typically contiguous on each disk, although this is not a requirement of the invention.
  • the/? disks are organized as three symmetric parity disks and /7-3 data disks.
  • phantom diagonal and anti-diagonal parity disks are assumed to be present in the array.
  • the contents of the phantom di- agonal and/or anti-diagonal parity disk are assumed to be a predetermined value, for example, zero, thereby enabling parity encoding/decoding utilizing the phantom (anti-) diagonal disks.
  • the phantom (anti-)diagonal disks contain a predetermined value, missing values may be computed among the data, symmetric parity and/or row parity disks in accordance with a symmetric algorithm.
  • the technique further assumes that the phantom diagonal parity disk stores parity information computed along diagonal parity sets ("diagonals") of the array. The blocks in a stripe are organized into/?
  • the technique assumes that the phantom anti-diagonal parity disk stores parity information computed along anti-diagonal parity sets ("anti-diagonals") of the array.
  • anti-diagonals have a slope that is orthogonal to the diagonals.
  • the symmetric TP technique is invoked by computing a total of p 4-tuple sums along an intermediate (i.e., middle) disk of the failed disks via generation of a number of crosses.
  • the 4-tuple sums are thereafter reduced to a total of /7-1 pairwise sums on the middle disk.
  • the reduced pairwise sums form a solvable system of linear equations.
  • the data on the middle disk is first recovered by solving the system of equations. Once a first block on the middle disk has been resolved, its solution may be substituted into other equations until all blocks on the middle disk have been recovered. Once the middle disk has been recovered, the system performs a row-diagonal parity technique to recover the remaining two disks. In the case of an initial write operation, encoding of the initial parity information on the three parity disks utilizes the same algorithm as recovering three failed disks.
  • the three parity values are in- eluded within the computation of the phantom diagonal and phantom anti-diagonal pari- ties.
  • the three parity values may be computed using the same technique as used for a triple parity erasure.
  • a TP computation algorithm of the technique utilizes the "a priori" knowledge that the (anti-) diagonal parity values are set to a predetermined value, so that when thep-3 data blocks are written, the values of the three parity disks may be computed using TP reconstruction algorithm, thereby generating appropriate parity values to store on the parity disks.
  • data is reconstructed using the same algorithm as utilized for initial parity computation, i.e., in a symmetric fashion.
  • FIG. 1 is a schematic block diagram of a storage system environment including a storage system in accordance with an embodiment of the present invention
  • Fig. 2 is a schematic block diagram of an exemplary storage operating system for use on the storage system of Fig.1 in accordance with an embodiment of the present invention
  • Fig. 3 is a schematic block diagram of a disk array showing disk identifiers in accordance with an embodiment of the present invention.
  • Fig. 4 is a block diagram of a disk array organized in accordance with an embodiment of the present invention;
  • Fig. 5 is a schematic block diagram of a disk array showing diagonal parity stripes in accordance with an embodiment of the present invention
  • Fig. 6 is a schematic block diagram of a disk array showing anti-diagonal parity stripes in accordance with an embodiment of the present invention.
  • Fig. 7 is a flowchart detailing the steps of a procedure for performing symmetric triple parity computation in accordance with an embodiment of the present invention.
  • Fig. 1 is a schematic block diagram of an environment 100 including a storage system 120 that may be advantageously used with the present invention.
  • the inventive technique described herein may apply to any type of special-purpose (e.g., file server or filer) or general-purpose computer, including a standalone computer or portion thereof, embodied as or including a storage system 120.
  • the teachings of this invention can be adapted to a variety of storage system architectures including, but not limited to, a network-attached storage environment, a storage area network and a disk assembly di- rectly-attached to a client or host computer.
  • storage system should therefore be taken broadly to include such arrangements in addition to any subsystems configured to perform a storage function and associated with other equipment or systems.
  • the storage system 120 comprises a processor 122, a memory 124, a network adapter 125 and a storage adapter 128 interconnected by a sys- tern bus 132.
  • the memory 124 comprises storage locations that are addressable by the processor and adapters for storing software program code and data structures associated with the present invention.
  • the processor and adapters may, in turn, comprise processing elements and/or logic circuitry configured to execute the software code and manipulate the data structures.
  • a storage operating system 200 portions of which are typically resi- dent in memory and executed by the processing elements, functionally organizes the system 120 by, inter alia, invoking storage operations executed by the storage system.
  • the network adapter 125 comprises a plurality of ports adapted to couple the storage system 120 to one or more clients 110 over point-to-point links, wide area networks, virtual private networks implemented over a public network (Internet) or a shared local area network (hereinafter "network 105").
  • the network adapter 125 thus may comprise the mechanical, electrical and signaling circuitry needed to connect the storage system to the network 105, such as a computer network.
  • the network 105 may be embodied as an Ethernet network or a Fibre Channel (FC) network.
  • Each client 110 may communicate with the storage system 120 over network 105 by exchanging discrete frames or packets of data according to pre-defined protocols, such as TCP/IP.
  • the storage adapter 128 cooperates with the storage operating system 200 execut- 5 ing on the system 120 to access information requested by a user (or client).
  • the information may be stored on any type of attached array of writeable storage device media such as video tape, optical, DVD, magnetic tape, bubble memory, electronic random access memory, micro-electro mechanical and any other similar media adapted to store information, including data and parity information.
  • the information is preferably stored on disks 150, such as HDD and/or DASD, of array 140.
  • the storage adapter includes input/output (I/O) interface circuitry that couples to the disks over an I/O interconnect arrangement, such as a conventional high-performance, FC serial link topology.
  • Storage of information on array 140 is preferably implemented as one or mores storage 'Volumes" that comprise a cluster of physical storage disks 150, defining an overall logical arrangement of disk space.
  • Each volume is generally, although not necessarily, associated with its own file system.
  • the disks within a volume/file system are typically organized as one or more groups, wherein each group is operated as a Redundant Array of Independent (or Inexpensive) Disks (RAID).
  • RAID Redundant Array of Independent (or Inexpensive) Disks
  • stripes across a given number of physical disks in the RAID group, and the appropriate storing of parity information with respect to the striped data.
  • the present invention comprises a symmetric "triple" parity (TP) technique that provides triple failure parity correcting recovery using row parity, symmetric parity and5 diagonal and anti-diagonal parity in a disk array.
  • the inventive technique is preferably implemented by a disk storage layer (shown at 240 of Fig. 2) of the storage operating system 200 to provide a method and system for constructing parity in stripes across multiple storage devices, such as disks, in a storage system.
  • the data on the data disks is stored "in the clear", meaning it is not further encoded for storage.
  • the contents of the array can be reconstructed entirely, without loss of data, after any one, two or three concurrent disk failures.
  • the invention comprises/? storage devices where/? is a prime number.
  • the storage devices are divided into blocks, where the blocks are the same size.
  • the blocks are then organized into stripes that contain a same number of blocks in each device, wherein each stripe contains p-1 rows of blocks. That is, each stripe, one block is selected from each of all but one of the devices that are not the phantom diagonal or phantom anti-diagonal parity device in that stripe, with the further restriction that no two of the selected blocks belong to the same row. This is called a diagonal parity set or "di- agonal".
  • a diagonal may be formed, for example, by numbering the data, and parity devices from 0 to p-1, numbering the rows from 0 to p-2, and then assigning the block at device i, rowy to diagonal (/+/) mod(p).
  • the formation of diagonals continues (e.g., wraps around within the group of p-1 rows) until all blocks that are not on the phantom diagonal and anti-diagonal parity devices in the stripe are assigned to diagonals, with the further restriction that no two diagonals exclude all blocks from the same device. There are/? diagonals.
  • a similar technique is utilized to assign parity and blocks to anti- diagonal parity sets or ⁇ anti-diagonals.
  • anti-diagonals continue (e.g., wraps around within a group of p-1 rows) until all blocks that are not on the phantom anti-diagonal parity and diagonal devices in the stripe are assigned to anti-diagonals, with the further restriction that no two anti-diagonals excludes all blocks from the same device.
  • An anti-diagonal may be formed, for example, by numbering the data, and parity devices from 0 to p-1, numbering the rows from 0 to p-2, and then assigning the block at device /, rowy to diagonal (p- ⁇ -i+j) mod (/?).
  • a triple parity recovery technique can be made symmetric to enable the same procedure to be performed for both encoding and decoding.
  • Parity is generally calculated as an exclusive-OR (XOR) of data blocks to form a parity block.
  • the XOR operation is generally performed over the same 1-bit field in each input block to produce a single corresponding bit of output.
  • the XOR operation is equivalent to two's complement addition or subtraction of two 1-bit fields. Redundant parity information may also be computed as the sum of same-sized multi-bit fields (e.g., 8, 16, 32, 64, 128 bits) in all the inputs.
  • the equivalent of parity may be computed by adding data using two's complement addition on 32-bit fields to produce each 32 bits of redundant information. This is only the case assuming non-reliance on the fact that an XOR operation directed to the same input twice into a block produces the original content of the block, as the XOR of a block with itself produces zero.
  • a block for purposes of parity computation may or may not correspond to a file block, a database block, a disk sector or any other conveniently sized unit.
  • the block size used for parity computation have any relation to any other block sizes used in the system.
  • one or more integer number of parity blocks will fit into a unit defined to be one or more integer number of disk sectors.
  • some number of blocks will correspond to file system or database blocks, and will typically be of size 4k (4096) bytes or some higher power of two bytes (e.g., 8k, 16k, 32k, 64k, 128k, 256k).
  • the illustrative system described herein preferably performs full stripe write operations.
  • individual file blocks that are typically 4k or 8k bytes may be divided into smaller blocks used only for parity computation, so that full stripes of, e.g., 4k byte sized blocks can be written to the disks of the array.
  • all parity computations may be performed in memory before the results are written to disk, thus reducing the burden of computing and updating parity on disk.
  • the storage operating system 200 implements a write-anywhere file system that cooperates with virtualization modules to provide a function that "virtualizes" the storage space provided by disks.
  • the file system logically or- ganizes the information as a hierarchical structure of named directory and file objects (hereinafter "directories" and "files") on the disks.
  • directory names and file objects
  • Each "on-disk" file may be implemented as set of disk blocks configured to store information, such as data, whereas the directory may be implemented as a specially formatted file in which names and links to other files and directories are stored.
  • the virtualization system allows the file system to further logically organize information as a hierarchical structure of named virtual disks (vdisks) on the disks, thereby providing an integrated NAS and SAN system approach to storage by enabling file-based (NAS) access to the files and directories, while further enabling block-based (SAN) access to the vdisks on a file-based storage platform.
  • vdisks virtual disks
  • the storage operating system is preferably the NetApp® Data ONT AP® operating system available from Network Appliance, Inc., Sunnyvale, California that implements a Write Anywhere File Layout (WAFL®) file system.
  • WAFL® Write Anywhere File Layout
  • any appropriate storage operating system including a write in-place file system, may be enhanced for use in accordance with the inventive principles described herein.
  • Data ONTAP is employed, it should be taken broadly to refer to any storage operating system that is otherwise adaptable to the teachings of this invention.
  • the term "storage operating system” generally refers to the computer-executable code operable on a computer that manages data access and may, in the case of a storage system, implement data access semantics, such as the Data ONT AP® storage operating system, which is implemented as a microkernel.
  • the storage operating system can also be implemented as an application program operating over a general- purpose operating system, such as UNIX® or Windows XP®, or as a general-purpose operating system with configurable functionality, which is configured for storage applications as described herein.
  • inventive technique described herein may apply to any type of special-purpose (e.g., storage serving appliance) or general-purpose computer, including a standalone computer or portion thereof, embodied as or including a storage system.
  • teachings of this invention can be adapted to a variety of storage system architectures including, but not Hm- ited to, a network-attached storage environment, a storage area network and disk assembly directly-attached to a client or host computer.
  • storage system should therefore be taken broadly to include such arrangements in addition to any subsystems configured to perform a storage function and associated with other equipment or systems.
  • Fig. 2 is a schematic block diagram of the storage operating system 200 that may be advantageously used with the present invention.
  • the storage operating system com- prises a series of software layers organized to form an integrated network protocol stack or, more generally, a multi-protocol engine that provides data paths for clients to access information stored on the multi-protocol storage system using block and file access protocols.
  • the protocol stack includes a media access layer 210 of network drivers (e.g., gigabit Ethernet drivers) that interfaces to network protocol layers, such as the IP layer 212 and its supporting transport mechanisms, the TCP layer 214 and the User Datagram Protocol (UDP) layer 216.
  • network drivers e.g., gigabit Ethernet drivers
  • a file system protocol layer provides multi-protocol file access and, to that end, includes support for the DAFS protocol 218, the Network File System (NFS) protocol 220, the Common Internet File System (CIFS) protocol 222 and the Hypertext Transfer Protocol (HTTP) protocol 224.
  • a VI layer 226 implements the VI architecture to provide direct access transport (DAT) capabilities, such as RDMA, as required by the DAFS protocol 218.
  • DAT direct access transport
  • An iSCSI driver layer 228 provides block protocol access over the TCP/IP network protocol layers, while a FC driver layer 230 operates with the network adapter to receive and transmit block access requests and responses to and from the storage system.
  • the FC and iSCSI drivers provide FC-specific and iSCSI-specific access control to the luns (vdisks) and, thus, manage exports of vdisks to either iSCSI or FCP or, alternatively, to both iSCSI and FCP when accessing a single vdisk on the storage system.
  • the storage operating system includes a disk storage layer 240, such as a RAID system, that implements a disk storage protocol, such as a RAID protocol, and a disk driver layer 250 that implements a disk access protocol such as, e.g., a SCSI protocol.
  • a disk storage layer 240 such as a RAID system
  • a disk driver layer 250 that implements a disk access protocol such as, e.g., a SCSI protocol.
  • the disk storage layer (e.g., RAID system 240) implements the novel symmetric TP technique.
  • the RAID system 240 encodes data according to an encoding technique described below and, in response to the detection of one or more failures of storage devices, executes a novel TP reconstruction algorithm of the technique, also described further below.
  • the novel symmetric TP technique may be implemented by modules of the storage operating system other than the RAID system 240. As such, the description of the RAID system 240 im- plementing the novel symmetric TP technique should be taken as exemplary only.
  • a virtualization system 255 that is implemented by a file system 265 interacting with virtualization modules illustratively embodied as, e.g., vdisk module 270 and SCSI target module 260.
  • virtualization modules illustratively embodied as, e.g., vdisk module 270 and SCSI target module 260.
  • the vdisk module 270, the file system 265 and SCSI target module 260 can be implemented in software, hardware, firmware, or a combination thereof.
  • the vdisk module 270 interacts with the file system 265 to enable access by administrative interfaces in response to a system administrator issuing commands to the multi-protocol storage system 120.
  • the vdisk module 270 manages SAN deployments by, among other things, implementing a comprehensive set of vdisk (lun) commands issued through a user interface by a system administrator. These vdisk commands are converted to primitive file system operations ("primitives") that interact with the file system 265 and the SCSI target module 260 to implement the vdisks.
  • primary primitive file system operations
  • the SCSI target module 260 initiates emulation of a disk or lun by providing a mapping procedure that translates luns into the special vdisk file types.
  • the SCSI target module is illustratively disposed between the FC and iSCSI drivers 230, 228 and the file system 265 to thereby provide a translation layer of the virtualization system 255 between the SAN block (lun) space and the file system space, where luns are represented as vdisks.
  • the multi-protocol storage system reverses the approaches taken by prior systems to thereby provide a single unified storage platform for essentially all storage access protocols.
  • the file system 265 is illustratively a message-based system; as such, the SCSI target module 260 transposes a SCSI request into a message representing an operation directed to the file system.
  • the message generated by the SCSI target module may include a type of operation (e.g., read, write) along with a pathname (e.g., a path descriptor) and a filename (e.g., a special filename) of the vdisk object represented in the file system.
  • the SCSI target module 260 passes the message into the file system 265 as, e.g., a function call, where the operation is performed.
  • the file system 265 illustratively implements the WAFL file system having an on-disk format representation that is block-based using, e.g., 4 kilobyte (KB) blocks and using inodes to describe the files.
  • the WAFL file system uses files to store metadata de- scribing the layout of its file system; these metadata files include, among others, an inode file.
  • a file handle i.e., an identifier that includes an inode number, is used to retrieve an inode from disk.
  • a description of the structure of the file system, including on-disk ino- des and the inode file, is provided in the U.S. Patent No.
  • a request from the client 110 is forwarded as a packet over the computer network 105 and onto the storage system 120 where it is received at the net- work adapter 125.
  • a network driver processes the packet and, if appropriate, passes it on to a network protocol and file access layer for additional processing prior to forwarding to the write-anywhere file system 265.
  • the file system generates operations to load (retrieve) the requested data from disk 150 if it is not resident "in-core," i.e., in the memory 124. If the information is not in the memory, the file system 265 indexes into the inode file using the inode number to access an appropriate entry and retrieve a logical volume block number (vbn).
  • the file system then passes a message structure including the logical vbn to the RAID system 240; the logical vbn is mapped to a disk identifier and disk block number (disk,dbn) and sent to an appropriate driver (e.g., SCSI) of the disk driver system 250.
  • the disk driver accesses the dbn from the specified disk 150 and loads the requested data block(s) in memory for processing by the storage system.
  • the storage system (and operating system) returns a reply to the client 110 over the network 105.
  • a storage access request data path may be implemented as logic circuitry embodied within a field programmable gate array (FPGA) or an application specific integrated circuit (ASIC).
  • FPGA field programmable gate array
  • ASIC application specific integrated circuit
  • the processing elements of adapter 125, 128 may be configured to offload some or all of the packet processing and storage access operations, respectively, from processor 122, to thereby increase the performance of the storage service provided by the system. It is expressly contemplated that the various processes, architectures and procedures described herein can be implemented in hardware, firmware or software.
  • the term "storage operating system” generally refers to the computer-executable code operable to perform a storage function in a storage system, e.g., that manages data access and may implement file system semantics.
  • the Data ONTAP software is an example of such a storage operating system implemented as a microkernel and including a file system layer to implement file system semantics and manage data access.
  • the storage operating system can also be implemented as an application program operating over a general-purpose operating system, such as UNIX® or Windows XP®, or as a general-purpose operating system with configurable functionality, which is configured for storage applications as described herein.
  • inventive tech- nique described herein may apply to any type of special-purpose (e.g., file server, filer or storage system) or general-purpose computer, including a standalone computer or portion thereof, embodied as or including a storage system 120.
  • An example of a storage system that may be advantageously used with the present invention is described in U.S. Patent Application Serial No. 10/215,917 titled, MULTI-PROTOCOL STORAGE APPLIANCE THAT PROVIDES INTEGRATED SUPPORT FOR FILE AND BLOCK ACCESS
  • PROTOCOLS by Brian Pawlowski, et al., filed on August 8, 2002.
  • teachings of this invention can be adapted to a variety of storage system architectures including, but not limited to, a network-attached storage environment, a storage area network and disk assembly directly-attached to a client or host computer.
  • storage sys- tern should therefore be taken broadly to include such arrangements in addition to any subsystems configured to perform a storage function and associated with other equipment or systems.
  • the present invention comprises a symmetric triple parity (TP) technique that re- prises the overhead of computing parity for a storage array adapted to enable efficient re- covery from the concurrent failure of up to three storage devices in the array.
  • the TP technique is preferably used in an array comprising a number/? of storage devices, such as disks, with/? being a prime number.
  • the disks are divided into blocks and the blocks are organized into stripes.
  • the blocks of rows selected to form a stripe are typically con- tiguous on each disk, although this is not a requirement of the invention.
  • the p disks are organized as three parity disks and/>-3 data disks.
  • phantom diagonal and anti-diagonal parity disks are assumed to be present in the array.
  • the contents of the phantom diagonal and/or anti- diagonal parity disk are assumed to be a predetermined value, for example, zero, thereby enabling parity encoding/decoding utilizing the phantom (anti-) diagonal disks.
  • the phantom (anti-)diagonal disks contain a predetermined value, missing values may be computed among the data, and symmetric parity disks in accordance with a symmetric algorithm.
  • the technique further assumes that the phantom diagonal parity disk stores parity information computed along diagonal parity sets ("diagonals") of the array.
  • the blocks in a stripe are organized into p diagonals, each of which contains p- ⁇ blocks from the data and symmetric parity disks, and all but one of which stores its parity in a block on the diagonal parity disk.
  • the technique assumes that the anti-diagonal parity disk stores parity information computed along anti-diagonal parity sets ("anti-diagonals") of the array.
  • anti-diagonals have a slope that is orthogonal to the diagonals.
  • the symmetric TP technique is invoked by computing a total of p 4-tuple sums along an intermediate (i.e., middle) disk of the failed disks via generation of a number of crosses.
  • the 4-tuple sums are thereafter reduced to a total of /7-1 pairwise sums on the middle disk.
  • the reduced pairwise sums form a solvable system of linear equations.
  • the data on the middle disk is first recovered by solving the system of equations. Once a first block on the middle disk has been resolved, its solution may be substituted into other equations until all blocks on the middle disk have been recovered. Once the middle disk has been recovered, the system performs a row-diagonal parity technique to recover the remaining two disks.
  • the three parity disks utilizes the same algorithm as recovering three failed disks.
  • row parity and symmetric parity values are included within the computation of the diagonal and anti-diagonal parities.
  • the three symmetric parity values may be computed using the same technique as used for a triple parity erasure.
  • a TP computation algorithm of the technique utilizes the "a priori" knowledge that the (anti-) diagonal parity values are set to a predetermined value, so that when thep-3 data blocks are written, the values of the three symmetric parity disks may be computed using TP reconstruction algorithm, thereby generating appropriate parity values to store on the symmetric parity disks.
  • data is reconstructed using the same algorithm as utilized for initial parity computation, i.e., in a symmetric fashion.
  • the first p-3 disks (DO-I) hold data
  • disk p (RP) holds row parity
  • disks SP 1-2 hold symmetric parity information to enable symmetric encoding and recovery of the array. It should be noted that these three disks, i.e., disk RP and disk SPl and SP2, may be used interchangeably.
  • the three disks may be referred to generally as symmetric parity disks.
  • the final row of the array is a phantom row and is shown only for descriptive purposes, phantom disk DP holds diagonal parity and phantom disk ADP holds anti-diagonal parity.
  • Disks DP and ADP are assumed to store zero values in accordance with an illustrative embodiment of the invention. However, it should be noted that in alternate embodiments phantom disks ADP/DP may be assumed to store other prede- termined values. As such, the description of disks ADP/DP storing zero values should be taken as exemplary only.
  • Fig. 4 is a schematic diagram illustrating as- signment of blocks to parity groups in accordance with an embodiment of the present invention
  • the notation D a> b, c and P a b , c denotes the respective contributions of data (D) and parity (P) blocks to specific row (a), diagonal (b) and anti-diagonal (c) parity computations.
  • the notation D ⁇ c means that those data blocks belong to the row or diagonal used for purposes of computing row parity a, diagonal parity b and anti-diagonal parity c, while P a , b,c stores the parity for row parity set a and also contributes to diagonal par- ity set b and anti-diagonal parity c.
  • P a , b,c stores the parity for row parity set a and also contributes to diagonal par- ity set b and anti-diagonal parity c.
  • each of the diagonal parity blocks stored on the diagonal parity disk contains contributions from all but one of the other disks (including the row parity disk but excluding the anti-diagonal disk) of the array.
  • the diagonal parity block P 4 has contributions from DO (Do, 4, 9 ), SPl (SP 3 , 4, 1), SP2 (SP 2 , 4, 13) and RP (Pi, 4, 11), but no contribution from Dl.
  • Figs. 5 and 6 are schematic diagrams of an array showing the assignment of blocks to diagonals and anti-diagonals, respectively.
  • Fig. 5 shows an array 500 illustrating the assignment of blocks to diagonals, wherein each block is labeled with a number corresponding to the diagonal to which it belongs. Also shown (in phantom) is the dropped diagonal that is not computed when encoding the array.
  • Fig. 6 shows an array 600 illustrating the assignment of blocks to anti-diagonals, including illustration (in phantom) of the dropped anti-diagonal that is not computed on the array.
  • the assignment of blocks to (anti-)diagonals may be permuted as long as no single disk includes two blocks of an (anti-)diagonal thereon.
  • FIG. 7 is a flowchart detailing the steps of a procedure 700 for encoding/decoding symmetric triple parity in accordance with an embodiment of the present invention. It should be noted that in the illustrative embodiment, during an initial encoding operation, i.e., in response to an initial write operation, the three symmetric parity disks are assumed to have failed, thereby necessitating the computation of the parity that should be written in response to the write operation. It should be noted that for a write operation, only those parity blocks within the stripe where the write occurs need to be computed/reconstructed. This is in distinction to a disk failure where the disk(s) must be wholly reconstructed.
  • the term "failed disks" may include the three symmetric disks when performing an initial write operation.
  • the XOR sum of blocks along each diagonal and anti-diagonal is zero.
  • the procedure 700 begins in step 705 and continues to step 712 where the RAID system computes an algebraic operation, e.g., an XOR sum of the missing blocks on the three failed disks along each of the three dimensions - row, diagonal and anti-diagonal.
  • an algebraic operation e.g., an XOR sum of the missing blocks on the three failed disks along each of the three dimensions - row, diagonal and anti-diagonal.
  • com- puting this sum at this point helps to reduce the number of XOR operations required to reconstruct the remaining two disks once the middle disk has been recovered.
  • the anti-diagonal parity sum of missing blocks Doo, Di 1 and SP 33 (referencing Fig 6, these are blocks on anti-diagonal 4) can be computed as
  • the RAID system then identifies the failed disks as X, Y and Z having a value equal to the index of the three failed disks in the array and wherein the disks are labeled from 0-p- 1.
  • X O
  • other orderings of failed disks is possible.
  • the Y disk is the middle disk.
  • the row on which the (anti-)diagonal terminates is termed row q.
  • the right hand side of this equation is known, thereby leaving an equation with four unknowns on the middle disk.
  • p such sums are computed by starting crosses at different stripes. Since the array only contains p- 1 rows, the pth 4-tuple sum is formed by constructing a cross using the phantom row D 4 o, D 41 , SP 42 , SP 43 , and RP 4 and the dropped diagonal and anti-diagonal corresponding to Z and X respectively. More generally, duplicate terms in the top and bottom rows at each cross are canceled, thereby resulting in the sum of at most four blocks on the middle disk.
  • the first cross (based on row 0) results in a tuple comprising
  • the tuple corresponding to row k can be represented by [k, k+g, k+h, k+h+g] wherein all of the addition is performed modulo p.
  • the step of computing 4-tuple sums can be performed as- suming an arbitrary ordering of the failed disks.
  • the RAID system illustratively reduces the 4-tuple sums to pairwise sums on the middle disk in step 720.
  • the system selects a subset of the equations such that the common blocks can- eel out leaving only two blocks. Computing the sum for the subset results in a pairwise sum of blocks on the middle disk.
  • the 4-tuple sums represent a set of linear equations that can be solved in a number of ways. Reduction to pairwise sums is one illustrative embodiment. There may be multiple ways for choosing a subset of tuples to reduce them to a pairwise sum.
  • the subset of tuples is chosen by starting with a tuple corresponding to row k and chosing subsequent tuples at an offset g (or K) from the last. At each step, common blocks are cancelled and the process continues until only two unknown blocks are left remaining. This results in a pairwise sum. For example, starting with the 4-tuple sum corresponding to row 0, i.e., [0, g, h, g
  • step m results in the sum of blocks [0, (m*g) mod p, h, (m*g + h)mod
  • one of the equations results in a single unknown equaling a known quantity, which in this case is zero, which may then be substi- tuted into the previous equations to solve for all unknowns on the middle disk, thereby recovering the middle disk in step 725.
  • the pth block on disk Y is zero.
  • the value of the block that is pairwise XORed with the pth block is known at the completion of creation of the pairwise sums, i.e., an equation results with only one unknown.
  • the remaining blocks on the middle disk may be recovered.
  • the RAID system has reconstructed the middle disk leaving only two failed disks. As this problem has been solved previously using row diagonal parity, the system imple- ments such R-D parity to recover the missing two disks.
  • the RAID system utilizes R-D parity to recover the remaining two disks in step 730 before completing in step 735.
  • R-D parity is described in the above-incorporated U.S. Patent Application Serial No. 10/035,607, which was filed on December 28, 2001, by Peter F. Corbett for a ROW-DIAGONAL PARITY TECHNIQUE FOR ENABLING EFFICIENT RECOVERY FROM DOUBLE FAILURES IN A STORAGE ARRAY.
  • the symmetric TP technique of the present invention may also be utilized in recovering from triple failures in other applications that rely on redundant data streams.
  • the TP technique may be utilized in data communication applications, where additional data is transmitted to reduce the reliance on retransmissions to recover from up to three lost and/or corrupted packets.
  • other algebraic operations other than XOR operations, may be utilized.
  • any horizontal erasure code and technique may be modified to generate a symmetric variant in accordance with the principles of the present invention by introducing one or more symmetric parity disks and utilizing one or more phantom parity disks containing predetermined values to thereby enable parity encoding and decoding to occur utilizing identical algorithms.

Abstract

A symmetric triple parity (TP) technique in an array comprising a number p of storage devices, such as disks, with p being a prime number is provided. The p disks are organized as one row parity disk, two symmetric parity disks and p-3 data disks. Phantom diagonal and anti-diagonal parity disks assumed to be present are further assumed to contain a predetermined value, thereby enabling parity encoding/decoding utilizing the phantom (anti-) diagonal disks. Row parity and symmetric parity values are included within the computation of the diagonal and anti-diagonal parities; accordingly, the two symmetric parity and the row parity values may be computed using the same technique as used for a triple parity erasure, i.e., in a symmetric fashion.

Description

SYSTEM AND METHOD FOR SYMMETRIC TRIPLE PARITY
FIELD OF THE INVENTION
The present invention relates to arrays of storage systems and, more specifically, to a technique for symmetric parity computation and reconstruction to tolerate up to three failing storage devices of a storage array.
BACKGROUND OF THE INVENTION
A storage system typically comprises one or more storage devices into which data may be entered, and from which data may be obtained, as desired. The storage system may be implemented in accordance with a variety of storage architectures including, but not limited to, a network-attached storage environment, a storage area network and a disk assembly directly attached to a client or host computer. The storage devices are typically disk drives, wherein the term "disk" commonly describes a self-contained rotating magnetic media storage device. The term "disk" in this context is synonymous with hard disk drive (HDD) or direct access storage device (DASD). The disks within a storage system are typically organized as one or more groups, wherein each group is operated as a Redundant Array of Independent (or Inexpensive) Disks (RAID). Most RAID implementations enhance the reliability/integrity of data storage through the redundant writing of data "stripes" across a given number of physical disks in the RAID group, and the appropriate storing of redundant information with re- spect to the striped data. The redundant information enables recovery of data lost when a storage device fails.
In the operation of a disk array, it is anticipated that a disk can fail. A goal of a high performance storage system is to make the mean time to data loss (MTTDL) as long as possible, preferably much longer than the expected service life of the system. Data can be lost when one or more disks fail, making it impossible to recover data from the device. Typical schemes to avoid loss of data include mirroring, backup and parity protection. Mirroring is an expensive solution in terms of consumption of storage resources, such as disks. Backup does not protect data modified since the backup was created. Parity schemes are common because they provide a redundant encoding of the data that allows for a single erasure (loss of one disk) with the addition of just one disk drive to the system. Parity protection is used in computer systems to protect against loss of data on a storage device, such as a disk. A parity value may be computed by summing (usually modulo 2) data of a particular word size (usually one bit) across a number of similar disks holding different data and then storing the results on an additional similar disk. That is, parity may be computed on vectors 1-bit wide, composed of bits in corresponding posi- tions on each of the disks. When computed on vectors 1-bit wide, the parity can be either the computed sum or its complement; these are referred to as even and odd parity respectively. Addition and subtraction on 1-bit vectors are both equivalent to exclusive-OR (XOR) logical operations. The data is then protected against the loss of any one of the disks, or of any portion of the data on any one of the disks. If the disk storing the parity is lost, the parity can be regenerated from the data. If one of the data disks is lost, the data can be regenerated by adding the contents of the surviving data disks together and then subtracting the result from the stored parity.
Typically, the disks are divided into parity groups, each of which comprises one or more data disks and a parity disk. A parity set is a set of blocks, including several data blocks and one parity block, where the parity block is the XOR of all the data blocks. A parity group is a set of disks from which one or more parity sets are selected. The disk space is divided into stripes, with each stripe containing one block from each disk. The blocks of a stripe are usually at the same locations on each disk in the parity group. Within a stripe, all but one block are blocks containing data ("data blocks") and one block is a block containing parity ("parity block") computed by the XOR of all the data. If the parity blocks are all stored on one disk, thereby providing a single disk that contains all (and only) parity information, a RAID-4 implementation is provided. If the parity blocks are contained within different disks in each stripe, usually in a rotating pattern, then the implementation is RAID-5. The term "RAID" and its various implementations are well-known and disclosed in A Case for Redundant Arrays of Inexpensive Disks (RAID), by D. A. Patterson, G. A. Gibson and R. H. Katz, Proceedings of the International Conference on Management of Data (SIGMOD), June 1988.
As used herein, the term "encoding" means the computation of a redundancy value over a predetermined subset of data blocks, whereas the term "decoding" means the reconstruction of a data or parity block by using a subset of data blocks and redundancy values. If one disk fails in the parity group, the contents of that disk can be decoded (reconstructed) on a spare disk or disks by adding all the contents of the remaining data blocks and subtracting the result from the parity block. Since two's complement addition and subtraction over 1-bit fields are both equivalent to XOR operations, this reconstruc- tion consists of the XOR of all the surviving data and parity blocks. Similarly, if the parity disk is lost, it can be recomputed in the same way from the surviving data.
Parity schemes generally provide protection against a single disk failure within a parity group. These schemes can also protect against multiple disk failures as long as each failure occurs within a different parity group. However, if two disks fail concur- rently within a parity group, then an unrecoverable loss of data is suffered. Failure of two disks concurrently within a parity group is a fairly common occurrence, particularly because disks "wear out" and because of environmental factors with respect to the operation of the disks. In this context, the failure of two disks concurrently within a parity group is referred to as a "double failure". A double failure typically arises as a result of a failure of one disk and a subsequent failure of another disk while attempting to recover from the first failure. The recovery or reconstruction time is dependent upon the level of activity of the storage system. That is, during reconstruction of a failed disk, it is possible that the storage system remains "online" and continues to serve requests (from clients or users) to access (i.e., read and/or write) data. If the storage system is busy serving requests, the elapsed time for reconstruction increases. The reconstruction process time also increases as the size and number of disks in the storage system increases, as all of the surviving disks must be read to reconstruct the lost data. Moreover, the double disk failure rate is proportional to the square of the number of disks in a parity group. However, having small parity groups is expensive, as each parity group requires an entire disk devoted to redundant data. Another failure mode of disks is media read errors, wherein a single block or sector of a disk cannot be read. The unreadable data can be reconstructed if parity is maintained in the storage array. However, if one disk has already failed, then a media read error on another disk in the array will result in lost data. This is a second form of double failure. It can easily be shown that the minimum amount of redundant information required to correct a double failure is two units. Therefore, the minimum number of parity disks that can be added to the data disks is two. This is true whether the parity is distributed across the disks or concentrated on the two additional disks.
A known double failure correcting parity scheme is an EVENODD XOR-based technique that allows a serial reconstruction of lost (failed) disks. The EVENODD technique is disclosed in an article of IEEE Transactions on Computers, Vol. 44, No. 2, titled EVENODD: An Efficient Scheme for Tolerating Double Disk Failures in RAID Architectures, by Blaum et al, Feb., 1995. A variant of EVENODD is disclosed in U.S. Patent Number 5,579,475, titled METHOD AND MEANS FOR ENCODING AND REBUILDING THE DATA CONTENTS OF UP TO TWO UNAVAILABLE DASDS IN A DASD ARRAY USING SIMPLE NON-RECURSIVE DIAGONAL AND ROW PARITY, by Blaum, et al., issued on November 26, 1996. The above-mentioned article and patent are hereby incorporated by reference as though fully set forth herein.
In certain storage environments, it is common to utilize a significant number of lower quality disk drives, such as, e.g., in near line storage systems for use as short term storage before data is backed up to tape or other long-term archival systems. However, as the number of disks in an array increases, the probability that multiple failures will occur also increases. The probability is exacerbated by a lower mean time to failure (MTTF) of less expensive storage devices. Thus, it is possible to have storage systems experiencing triple failures, that is, the concurrent failures of three devices in the storage array. Furthermore, numerous storage protocols, such as Serial Attached SCSI (SAS), Fibre Channel (FC), etc., have resulted in increasingly complex architectures for disk shelves which have resulted in a concomitant increase in the number of failures experienced by the disk shelves, thereby resulting in loss of access to each disk connected to a failed disk shelf. One technique for correcting triple failures is an extension of the EVENODD technique termed the STAR technique, which is described in Efficient and Effective Schemes for Streaming Media Delivery, by Cheng Wang, dated August 2005, the contents of which is hereby incorporated by reference. A noted disadvantage of such EVENODD and/or STAR techniques is that they utilize asymmetric parity algorithms that require different computational steps when encoding and/or decoding parity. Furthermore, asymmetric algorithms imply that each disk is not treated identically. As a result, configuration management tasks must know and identify whether a disk is of a particular type, e.g., whether a disk is a parity disk and/or a data disk. For example, a reconstruction technique may involve a plurality of differing algorithms depending on the number of failed data and/or parity disks as well as the type of failed parity disks, e.g., row parity, diagonal parity, etc. The asymmetric nature of these algorithms imposes additional computational complexity when implementing parity-based systems. This additional complexity may be especially noticeable when utiliz- ing embedded systems to implement parity-based computations.
A further noted disadvantage of asymmetric parity algorithms is that utilization of floating parity, i.e., parity stored on any of the storage devices within a parity group instead of on one or more dedicated parity storage devices, is not feasible when utilizing asymmetric parity algorithms. This is because floating parity relies on a scheme where some blocks on the newly added disk(s) are re-assigned as parity, converting their old locations within the parity set to data. However, because of the special properties of some of the parity disks, e.g., diagonal/anti-diagonal, asymmetric algorithms cannot move/relocate parity blocks on these disks to newly added disks.
SUMMARY OF THE INVENTION The present invention comprises a symmetric triple parity (TP) technique that utilizes an algorithm for parity computation is identical to that used for triple reconstruction. The TP technique is preferably used in an array comprising a number/? of storage devices, such as disks, with/? being a prime number. The disks are divided into blocks and the blocks are organized into stripes. The blocks of rows selected to form a stripe are typically contiguous on each disk, although this is not a requirement of the invention. Illustratively, the/? disks are organized as three symmetric parity disks and /7-3 data disks. Furthermore, in the illustrative embodiment, phantom diagonal and anti-diagonal parity disks are assumed to be present in the array. Moreover, the contents of the phantom di- agonal and/or anti-diagonal parity disk are assumed to be a predetermined value, for example, zero, thereby enabling parity encoding/decoding utilizing the phantom (anti-) diagonal disks. By assuming that the phantom (anti-)diagonal disks contain a predetermined value, missing values may be computed among the data, symmetric parity and/or row parity disks in accordance with a symmetric algorithm. The technique further assumes that the phantom diagonal parity disk stores parity information computed along diagonal parity sets ("diagonals") of the array. The blocks in a stripe are organized into/? diagonals, each of which contains /7-1 blocks from the data, symmetric parity and row parity disks, and all but one of which stores its parity in a block on the diagonal parity disk. Similarly, the technique assumes that the phantom anti-diagonal parity disk stores parity information computed along anti-diagonal parity sets ("anti-diagonals") of the array. Notably, the anti-diagonals have a slope that is orthogonal to the diagonals.
In the event of a triple failure, the symmetric TP technique is invoked by computing a total of p 4-tuple sums along an intermediate (i.e., middle) disk of the failed disks via generation of a number of crosses. The 4-tuple sums are thereafter reduced to a total of /7-1 pairwise sums on the middle disk. The reduced pairwise sums form a solvable system of linear equations. Illustratively, the data on the middle disk is first recovered by solving the system of equations. Once a first block on the middle disk has been resolved, its solution may be substituted into other equations until all blocks on the middle disk have been recovered. Once the middle disk has been recovered, the system performs a row-diagonal parity technique to recover the remaining two disks. In the case of an initial write operation, encoding of the initial parity information on the three parity disks utilizes the same algorithm as recovering three failed disks.
In accordance with the illustrative embodiment, the three parity values are in- eluded within the computation of the phantom diagonal and phantom anti-diagonal pari- ties. The three parity values may be computed using the same technique as used for a triple parity erasure. During an initial write operation, a TP computation algorithm of the technique utilizes the "a priori" knowledge that the (anti-) diagonal parity values are set to a predetermined value, so that when thep-3 data blocks are written, the values of the three parity disks may be computed using TP reconstruction algorithm, thereby generating appropriate parity values to store on the parity disks. In the event of a failure of up to three disks of the array, data is reconstructed using the same algorithm as utilized for initial parity computation, i.e., in a symmetric fashion.
BRIEF DESCRIPTION OF THE DRAWINGS The above and further advantages of the invention may be better understood by referring to the following description in conjunction with the accompanying drawings in which like reference numerals indicate identical or functionally similar elements:
Fig. 1 is a schematic block diagram of a storage system environment including a storage system in accordance with an embodiment of the present invention; Fig. 2 is a schematic block diagram of an exemplary storage operating system for use on the storage system of Fig.1 in accordance with an embodiment of the present invention;
Fig. 3 is a schematic block diagram of a disk array showing disk identifiers in accordance with an embodiment of the present invention. Fig. 4 is a block diagram of a disk array organized in accordance with an embodiment of the present invention;
Fig. 5 is a schematic block diagram of a disk array showing diagonal parity stripes in accordance with an embodiment of the present invention;
Fig. 6 is a schematic block diagram of a disk array showing anti-diagonal parity stripes in accordance with an embodiment of the present invention; and
Fig. 7 is a flowchart detailing the steps of a procedure for performing symmetric triple parity computation in accordance with an embodiment of the present invention. DETAILED DESCRIPTION OF THE ILLUSTRATIVE EMBODIMENTS
A. Storage System Environment
Fig. 1 is a schematic block diagram of an environment 100 including a storage system 120 that may be advantageously used with the present invention. The inventive technique described herein may apply to any type of special-purpose (e.g., file server or filer) or general-purpose computer, including a standalone computer or portion thereof, embodied as or including a storage system 120. Moreover, the teachings of this invention can be adapted to a variety of storage system architectures including, but not limited to, a network-attached storage environment, a storage area network and a disk assembly di- rectly-attached to a client or host computer. The term "storage system" should therefore be taken broadly to include such arrangements in addition to any subsystems configured to perform a storage function and associated with other equipment or systems.
In the illustrative embodiment, the storage system 120 comprises a processor 122, a memory 124, a network adapter 125 and a storage adapter 128 interconnected by a sys- tern bus 132. The memory 124 comprises storage locations that are addressable by the processor and adapters for storing software program code and data structures associated with the present invention. The processor and adapters may, in turn, comprise processing elements and/or logic circuitry configured to execute the software code and manipulate the data structures. A storage operating system 200, portions of which are typically resi- dent in memory and executed by the processing elements, functionally organizes the system 120 by, inter alia, invoking storage operations executed by the storage system. It will be apparent to those skilled in the art that other processing and memory means, including various computer readable media, may be used for storing and executing program instructions pertaining to the inventive technique described herein. The network adapter 125 comprises a plurality of ports adapted to couple the storage system 120 to one or more clients 110 over point-to-point links, wide area networks, virtual private networks implemented over a public network (Internet) or a shared local area network (hereinafter "network 105"). The network adapter 125 thus may comprise the mechanical, electrical and signaling circuitry needed to connect the storage system to the network 105, such as a computer network. Illustratively, the network 105 may be embodied as an Ethernet network or a Fibre Channel (FC) network. Each client 110 may communicate with the storage system 120 over network 105 by exchanging discrete frames or packets of data according to pre-defined protocols, such as TCP/IP.
The storage adapter 128 cooperates with the storage operating system 200 execut- 5 ing on the system 120 to access information requested by a user (or client). The information may be stored on any type of attached array of writeable storage device media such as video tape, optical, DVD, magnetic tape, bubble memory, electronic random access memory, micro-electro mechanical and any other similar media adapted to store information, including data and parity information. However, as illustratively described herein,o the information is preferably stored on disks 150, such as HDD and/or DASD, of array 140. The storage adapter includes input/output (I/O) interface circuitry that couples to the disks over an I/O interconnect arrangement, such as a conventional high-performance, FC serial link topology.
Storage of information on array 140 is preferably implemented as one or mores storage 'Volumes" that comprise a cluster of physical storage disks 150, defining an overall logical arrangement of disk space. Each volume is generally, although not necessarily, associated with its own file system. The disks within a volume/file system are typically organized as one or more groups, wherein each group is operated as a Redundant Array of Independent (or Inexpensive) Disks (RAID). Most RAID implementationso enhance the reliability/integrity of data storage through the redundant writing of data
"stripes" across a given number of physical disks in the RAID group, and the appropriate storing of parity information with respect to the striped data.
The present invention comprises a symmetric "triple" parity (TP) technique that provides triple failure parity correcting recovery using row parity, symmetric parity and5 diagonal and anti-diagonal parity in a disk array. The inventive technique is preferably implemented by a disk storage layer (shown at 240 of Fig. 2) of the storage operating system 200 to provide a method and system for constructing parity in stripes across multiple storage devices, such as disks, in a storage system. The data on the data disks is stored "in the clear", meaning it is not further encoded for storage. The contents of the array can be reconstructed entirely, without loss of data, after any one, two or three concurrent disk failures.
Broadly stated, the invention comprises/? storage devices where/? is a prime number. The storage devices are divided into blocks, where the blocks are the same size. The blocks are then organized into stripes that contain a same number of blocks in each device, wherein each stripe contains p-1 rows of blocks. That is, each stripe, one block is selected from each of all but one of the devices that are not the phantom diagonal or phantom anti-diagonal parity device in that stripe, with the further restriction that no two of the selected blocks belong to the same row. This is called a diagonal parity set or "di- agonal". A diagonal may be formed, for example, by numbering the data, and parity devices from 0 to p-1, numbering the rows from 0 to p-2, and then assigning the block at device i, rowy to diagonal (/+/) mod(p). The formation of diagonals continues (e.g., wraps around within the group of p-1 rows) until all blocks that are not on the phantom diagonal and anti-diagonal parity devices in the stripe are assigned to diagonals, with the further restriction that no two diagonals exclude all blocks from the same device. There are/? diagonals. A similar technique is utilized to assign parity and blocks to anti- diagonal parity sets or ^anti-diagonals. " The formation of anti-diagonals continues (e.g., wraps around within a group of p-1 rows) until all blocks that are not on the phantom anti-diagonal parity and diagonal devices in the stripe are assigned to anti-diagonals, with the further restriction that no two anti-diagonals excludes all blocks from the same device. An anti-diagonal may be formed, for example, by numbering the data, and parity devices from 0 to p-1, numbering the rows from 0 to p-2, and then assigning the block at device /, rowy to diagonal (p-\-i+j) mod (/?).
By assuming that the phantom (anti-)diagonal parity disks have a zero (or other predetermined) value, a triple parity recovery technique can be made symmetric to enable the same procedure to be performed for both encoding and decoding.
Parity is generally calculated as an exclusive-OR (XOR) of data blocks to form a parity block. The XOR operation is generally performed over the same 1-bit field in each input block to produce a single corresponding bit of output. As noted, the XOR operation is equivalent to two's complement addition or subtraction of two 1-bit fields. Redundant parity information may also be computed as the sum of same-sized multi-bit fields (e.g., 8, 16, 32, 64, 128 bits) in all the inputs. For example, the equivalent of parity may be computed by adding data using two's complement addition on 32-bit fields to produce each 32 bits of redundant information. This is only the case assuming non-reliance on the fact that an XOR operation directed to the same input twice into a block produces the original content of the block, as the XOR of a block with itself produces zero.
It will be apparent to those skilled in the art that a block (for purposes of parity computation) may or may not correspond to a file block, a database block, a disk sector or any other conveniently sized unit. There is no requirement that the block size used for parity computation have any relation to any other block sizes used in the system. However, it is expected that one or more integer number of parity blocks will fit into a unit defined to be one or more integer number of disk sectors. In many cases, some number of blocks will correspond to file system or database blocks, and will typically be of size 4k (4096) bytes or some higher power of two bytes (e.g., 8k, 16k, 32k, 64k, 128k, 256k). The illustrative system described herein preferably performs full stripe write operations. In particular, individual file blocks that are typically 4k or 8k bytes may be divided into smaller blocks used only for parity computation, so that full stripes of, e.g., 4k byte sized blocks can be written to the disks of the array. When full stripes are written to disk, all parity computations may be performed in memory before the results are written to disk, thus reducing the burden of computing and updating parity on disk.
B. Storage Operating System
To facilitate access to the disks, the storage operating system 200 implements a write-anywhere file system that cooperates with virtualization modules to provide a function that "virtualizes" the storage space provided by disks. The file system logically or- ganizes the information as a hierarchical structure of named directory and file objects (hereinafter "directories" and "files") on the disks. Each "on-disk" file may be implemented as set of disk blocks configured to store information, such as data, whereas the directory may be implemented as a specially formatted file in which names and links to other files and directories are stored. The virtualization system allows the file system to further logically organize information as a hierarchical structure of named virtual disks (vdisks) on the disks, thereby providing an integrated NAS and SAN system approach to storage by enabling file-based (NAS) access to the files and directories, while further enabling block-based (SAN) access to the vdisks on a file-based storage platform.
In the illustrative embodiment, the storage operating system is preferably the NetApp® Data ONT AP® operating system available from Network Appliance, Inc., Sunnyvale, California that implements a Write Anywhere File Layout (WAFL®) file system. However, it is expressly contemplated that any appropriate storage operating system, including a write in-place file system, may be enhanced for use in accordance with the inventive principles described herein. As such, where the term "Data ONTAP" is employed, it should be taken broadly to refer to any storage operating system that is otherwise adaptable to the teachings of this invention.
As used herein, the term "storage operating system" generally refers to the computer-executable code operable on a computer that manages data access and may, in the case of a storage system, implement data access semantics, such as the Data ONT AP® storage operating system, which is implemented as a microkernel. The storage operating system can also be implemented as an application program operating over a general- purpose operating system, such as UNIX® or Windows XP®, or as a general-purpose operating system with configurable functionality, which is configured for storage applications as described herein. In addition, it will be understood to those skilled in the art that the inventive technique described herein may apply to any type of special-purpose (e.g., storage serving appliance) or general-purpose computer, including a standalone computer or portion thereof, embodied as or including a storage system. Moreover, the teachings of this invention can be adapted to a variety of storage system architectures including, but not Hm- ited to, a network-attached storage environment, a storage area network and disk assembly directly-attached to a client or host computer. The term "storage system" should therefore be taken broadly to include such arrangements in addition to any subsystems configured to perform a storage function and associated with other equipment or systems.
Fig. 2 is a schematic block diagram of the storage operating system 200 that may be advantageously used with the present invention. The storage operating system com- prises a series of software layers organized to form an integrated network protocol stack or, more generally, a multi-protocol engine that provides data paths for clients to access information stored on the multi-protocol storage system using block and file access protocols. The protocol stack includes a media access layer 210 of network drivers (e.g., gigabit Ethernet drivers) that interfaces to network protocol layers, such as the IP layer 212 and its supporting transport mechanisms, the TCP layer 214 and the User Datagram Protocol (UDP) layer 216. A file system protocol layer provides multi-protocol file access and, to that end, includes support for the DAFS protocol 218, the Network File System (NFS) protocol 220, the Common Internet File System (CIFS) protocol 222 and the Hypertext Transfer Protocol (HTTP) protocol 224. A VI layer 226 implements the VI architecture to provide direct access transport (DAT) capabilities, such as RDMA, as required by the DAFS protocol 218.
An iSCSI driver layer 228 provides block protocol access over the TCP/IP network protocol layers, while a FC driver layer 230 operates with the network adapter to receive and transmit block access requests and responses to and from the storage system. The FC and iSCSI drivers provide FC-specific and iSCSI-specific access control to the luns (vdisks) and, thus, manage exports of vdisks to either iSCSI or FCP or, alternatively, to both iSCSI and FCP when accessing a single vdisk on the storage system. In addition, the storage operating system includes a disk storage layer 240, such as a RAID system, that implements a disk storage protocol, such as a RAID protocol, and a disk driver layer 250 that implements a disk access protocol such as, e.g., a SCSI protocol.
In the illustrative embodiment of the present invention, the disk storage layer (e.g., RAID system 240) implements the novel symmetric TP technique. Illustratively, during write operations, the RAID system 240 encodes data according to an encoding technique described below and, in response to the detection of one or more failures of storage devices, executes a novel TP reconstruction algorithm of the technique, also described further below. It should be noted that in alternate embodiments, the novel symmetric TP technique may be implemented by modules of the storage operating system other than the RAID system 240. As such, the description of the RAID system 240 im- plementing the novel symmetric TP technique should be taken as exemplary only. Bridging the disk software layers with the integrated network protocol stack layers is a virtualization system 255 that is implemented by a file system 265 interacting with virtualization modules illustratively embodied as, e.g., vdisk module 270 and SCSI target module 260. It should be noted that the vdisk module 270, the file system 265 and SCSI target module 260 can be implemented in software, hardware, firmware, or a combination thereof. The vdisk module 270 interacts with the file system 265 to enable access by administrative interfaces in response to a system administrator issuing commands to the multi-protocol storage system 120. In essence, the vdisk module 270 manages SAN deployments by, among other things, implementing a comprehensive set of vdisk (lun) commands issued through a user interface by a system administrator. These vdisk commands are converted to primitive file system operations ("primitives") that interact with the file system 265 and the SCSI target module 260 to implement the vdisks.
The SCSI target module 260, in turn, initiates emulation of a disk or lun by providing a mapping procedure that translates luns into the special vdisk file types. The SCSI target module is illustratively disposed between the FC and iSCSI drivers 230, 228 and the file system 265 to thereby provide a translation layer of the virtualization system 255 between the SAN block (lun) space and the file system space, where luns are represented as vdisks. By "disposing" SAN virtualization over the file system 265, the multi-protocol storage system reverses the approaches taken by prior systems to thereby provide a single unified storage platform for essentially all storage access protocols.
The file system 265 is illustratively a message-based system; as such, the SCSI target module 260 transposes a SCSI request into a message representing an operation directed to the file system. For example, the message generated by the SCSI target module may include a type of operation (e.g., read, write) along with a pathname (e.g., a path descriptor) and a filename (e.g., a special filename) of the vdisk object represented in the file system. The SCSI target module 260 passes the message into the file system 265 as, e.g., a function call, where the operation is performed.
The file system 265 illustratively implements the WAFL file system having an on-disk format representation that is block-based using, e.g., 4 kilobyte (KB) blocks and using inodes to describe the files. The WAFL file system uses files to store metadata de- scribing the layout of its file system; these metadata files include, among others, an inode file. A file handle, i.e., an identifier that includes an inode number, is used to retrieve an inode from disk. A description of the structure of the file system, including on-disk ino- des and the inode file, is provided in the U.S. Patent No. 5,819,292 entitled METHOD FOR MAINTAINING CONSISTENT STATES OF A FILE SYSTEM AND FOR CREATING USER-ACCESSIBLE READ-ONLY COPIES OF A FILE SYSTEM, by David Hitz, et al.
Operationally, a request from the client 110 is forwarded as a packet over the computer network 105 and onto the storage system 120 where it is received at the net- work adapter 125. A network driver processes the packet and, if appropriate, passes it on to a network protocol and file access layer for additional processing prior to forwarding to the write-anywhere file system 265. Here, the file system generates operations to load (retrieve) the requested data from disk 150 if it is not resident "in-core," i.e., in the memory 124. If the information is not in the memory, the file system 265 indexes into the inode file using the inode number to access an appropriate entry and retrieve a logical volume block number (vbn). The file system then passes a message structure including the logical vbn to the RAID system 240; the logical vbn is mapped to a disk identifier and disk block number (disk,dbn) and sent to an appropriate driver (e.g., SCSI) of the disk driver system 250. The disk driver accesses the dbn from the specified disk 150 and loads the requested data block(s) in memory for processing by the storage system. Upon completion of the request, the storage system (and operating system) returns a reply to the client 110 over the network 105.
It should be noted that the software "path" through the storage operating system layers described above needed to perform data storage access for the client request re- ceived at the storage system may alternatively be implemented in hardware. That is, in an alternate embodiment of the invention, a storage access request data path may be implemented as logic circuitry embodied within a field programmable gate array (FPGA) or an application specific integrated circuit (ASIC). This type of hardware implementation increases the performance of the storage service provided by storage system 120 in re- sponse to a request issued by client 110. Moreover, in another alternate embodiment of the invention, the processing elements of adapter 125, 128 may be configured to offload some or all of the packet processing and storage access operations, respectively, from processor 122, to thereby increase the performance of the storage service provided by the system. It is expressly contemplated that the various processes, architectures and procedures described herein can be implemented in hardware, firmware or software. As used herein, the term "storage operating system" generally refers to the computer-executable code operable to perform a storage function in a storage system, e.g., that manages data access and may implement file system semantics. In this sense, the Data ONTAP software is an example of such a storage operating system implemented as a microkernel and including a file system layer to implement file system semantics and manage data access. The storage operating system can also be implemented as an application program operating over a general-purpose operating system, such as UNIX® or Windows XP®, or as a general-purpose operating system with configurable functionality, which is configured for storage applications as described herein.
In addition, it will be understood to those skilled in the art that the inventive tech- nique described herein may apply to any type of special-purpose (e.g., file server, filer or storage system) or general-purpose computer, including a standalone computer or portion thereof, embodied as or including a storage system 120. An example of a storage system that may be advantageously used with the present invention is described in U.S. Patent Application Serial No. 10/215,917 titled, MULTI-PROTOCOL STORAGE APPLIANCE THAT PROVIDES INTEGRATED SUPPORT FOR FILE AND BLOCK ACCESS
PROTOCOLS, by Brian Pawlowski, et al., filed on August 8, 2002. Moreover, the teachings of this invention can be adapted to a variety of storage system architectures including, but not limited to, a network-attached storage environment, a storage area network and disk assembly directly-attached to a client or host computer. The term "storage sys- tern" should therefore be taken broadly to include such arrangements in addition to any subsystems configured to perform a storage function and associated with other equipment or systems.
C. Array Configuration
The present invention comprises a symmetric triple parity (TP) technique that re- duces the overhead of computing parity for a storage array adapted to enable efficient re- covery from the concurrent failure of up to three storage devices in the array. The TP technique is preferably used in an array comprising a number/? of storage devices, such as disks, with/? being a prime number. The disks are divided into blocks and the blocks are organized into stripes. The blocks of rows selected to form a stripe are typically con- tiguous on each disk, although this is not a requirement of the invention. Illustratively, the p disks are organized as three parity disks and/>-3 data disks. Furthermore, in the illustrative embodiment, phantom diagonal and anti-diagonal parity disks are assumed to be present in the array. Moreover, the contents of the phantom diagonal and/or anti- diagonal parity disk are assumed to be a predetermined value, for example, zero, thereby enabling parity encoding/decoding utilizing the phantom (anti-) diagonal disks. By assuming that the phantom (anti-)diagonal disks contain a predetermined value, missing values may be computed among the data, and symmetric parity disks in accordance with a symmetric algorithm.
The technique further assumes that the phantom diagonal parity disk stores parity information computed along diagonal parity sets ("diagonals") of the array. The blocks in a stripe are organized into p diagonals, each of which contains p-\ blocks from the data and symmetric parity disks, and all but one of which stores its parity in a block on the diagonal parity disk. Similarly, the technique assumes that the anti-diagonal parity disk stores parity information computed along anti-diagonal parity sets ("anti-diagonals") of the array. Notably, the anti-diagonals have a slope that is orthogonal to the diagonals.
In the event of a triple failure, the symmetric TP technique is invoked by computing a total of p 4-tuple sums along an intermediate (i.e., middle) disk of the failed disks via generation of a number of crosses. The 4-tuple sums are thereafter reduced to a total of /7-1 pairwise sums on the middle disk. The reduced pairwise sums form a solvable system of linear equations. Illustratively, the data on the middle disk is first recovered by solving the system of equations. Once a first block on the middle disk has been resolved, its solution may be substituted into other equations until all blocks on the middle disk have been recovered. Once the middle disk has been recovered, the system performs a row-diagonal parity technique to recover the remaining two disks. In the case of an initial write operation, encoding of the initial parity information the three parity disks utilizes the same algorithm as recovering three failed disks. In accordance with the illustrative embodiment, row parity and symmetric parity values are included within the computation of the diagonal and anti-diagonal parities. The three symmetric parity values may be computed using the same technique as used for a triple parity erasure. During an initial write operation, a TP computation algorithm of the technique utilizes the "a priori" knowledge that the (anti-) diagonal parity values are set to a predetermined value, so that when thep-3 data blocks are written, the values of the three symmetric parity disks may be computed using TP reconstruction algorithm, thereby generating appropriate parity values to store on the symmetric parity disks. In the event of a failure of up to three disks of the array, data is reconstructed using the same algorithm as utilized for initial parity computation, i.e., in a symmetric fashion.
Fig. 3 is a block diagram of a disk array 300 organized in accordance with the novel symmetric TP technique of the present invention. Assume/? equals the number of disks in the array, where/? =5. It should be noted that p=5 is used for exemplary purposes only and that any prime number may be utilized. As such, it should be noted that the principles of the present invention apply to any p. The first p-3 disks (DO-I) hold data, while disk p (RP) holds row parity, and disks SP 1-2 hold symmetric parity information to enable symmetric encoding and recovery of the array. It should be noted that these three disks, i.e., disk RP and disk SPl and SP2, may be used interchangeably. As such, the three disks may be referred to generally as symmetric parity disks. Furthermore, it should be noted that the final row of the array is a phantom row and is shown only for descriptive purposes, phantom disk DP holds diagonal parity and phantom disk ADP holds anti-diagonal parity. Disks DP and ADP are assumed to store zero values in accordance with an illustrative embodiment of the invention. However, it should be noted that in alternate embodiments phantom disks ADP/DP may be assumed to store other prede- termined values. As such, the description of disks ADP/DP storing zero values should be taken as exemplary only.
The data blocks, symmetric parity blocks and the row parity blocks are numbered such that each block belongs to a diagonal parity set and, within each row, each block belongs to a different diagonal parity set. Fig. 4 is a schematic diagram illustrating as- signment of blocks to parity groups in accordance with an embodiment of the present invention The notation Da>b,c and Pa b,c denotes the respective contributions of data (D) and parity (P) blocks to specific row (a), diagonal (b) and anti-diagonal (c) parity computations. That is, the notation D^c means that those data blocks belong to the row or diagonal used for purposes of computing row parity a, diagonal parity b and anti-diagonal parity c, while Pa,b,c stores the parity for row parity set a and also contributes to diagonal par- ity set b and anti-diagonal parity c. For example, Po, 8, io - Do, 4, 9 θ Do, 5, 13 θ SP0, 6, 12 Φ SPo, 7, 11, wherein θ represent an XOR operation. The notation also includes the parity block used for purposes of computing the diagonal parity for a particular diagonal, e.g., P4= Do, 4, 9 θ SP3,4,io θ SP2,4, i3 θ Pi, 4, 11. Note that each of the diagonal parity blocks stored on the diagonal parity disk contains contributions from all but one of the other disks (including the row parity disk but excluding the anti-diagonal disk) of the array. For example, the diagonal parity block P4 has contributions from DO (Do, 4, 9), SPl (SP3, 4, 1), SP2 (SP2, 4, 13) and RP (Pi, 4, 11), but no contribution from Dl. Note also that the diagonal parity for diagonal 8 (P8) is neither computed nor is it stored on the diagonal parity disk DP. Figs. 5 and 6 are schematic diagrams of an array showing the assignment of blocks to diagonals and anti-diagonals, respectively. Fig. 5 shows an array 500 illustrating the assignment of blocks to diagonals, wherein each block is labeled with a number corresponding to the diagonal to which it belongs. Also shown (in phantom) is the dropped diagonal that is not computed when encoding the array. Similarly, Fig. 6 shows an array 600 illustrating the assignment of blocks to anti-diagonals, including illustration (in phantom) of the dropped anti-diagonal that is not computed on the array. As noted above, the assignment of blocks to (anti-)diagonals may be permuted as long as no single disk includes two blocks of an (anti-)diagonal thereon.
D. Parity Encoding / Decoding Fig. 7 is a flowchart detailing the steps of a procedure 700 for encoding/decoding symmetric triple parity in accordance with an embodiment of the present invention. It should be noted that in the illustrative embodiment, during an initial encoding operation, i.e., in response to an initial write operation, the three symmetric parity disks are assumed to have failed, thereby necessitating the computation of the parity that should be written in response to the write operation. It should be noted that for a write operation, only those parity blocks within the stripe where the write occurs need to be computed/reconstructed. This is in distinction to a disk failure where the disk(s) must be wholly reconstructed. Thus, as used herein, the term "failed disks" may include the three symmetric disks when performing an initial write operation. Before initiating the encoding the parity, it should be noted that for each of the p disks, the XOR sum of blocks along each diagonal and anti-diagonal is zero. The procedure 700 begins in step 705 and continues to step 712 where the RAID system computes an algebraic operation, e.g., an XOR sum of the missing blocks on the three failed disks along each of the three dimensions - row, diagonal and anti-diagonal. Illustratively, com- puting this sum at this point helps to reduce the number of XOR operations required to reconstruct the remaining two disks once the middle disk has been recovered.
For the purposes of description of procedure 700, assume that disks DO, Dl and SP2 have failed. For example, the row parity sum of missing blocks Doo, DOi and SP03 (blocks on the first row of Fig. 3) can be computed as D00 ΘDoi ΘSP03 = RP0 ΘSP02
Similarly, the row parity sum of missing blocks D3o, D31, and SP33 (blocks on the fourth row of Fig. 3) can be computed as
D30 ΘD31 ΘSP33 = RP3 ΘSP32
The anti-diagonal parity sum of missing blocks Doo, Di 1 and SP33 (referencing Fig 6, these are blocks on anti-diagonal 4) can be computed as
D00 ΘDi i ΦSP33 = ADP00SP22
The diagonal parity sum of missing blocks D30, D21 and SP03 (referencing Fig 5, these are blocks on diagonal 3) can be computed as
D30 ΘD21 ΘSP03 = DP3 ΘSP12 The RAID system then computes a total of/? 4-tuple sums on the failed middle disk (steps 714-718). In step 713, the disks are ordered by defining a distance between the columns of the failed disks as: g=Y-X h=Z-Y
The RAID system then identifies the failed disks as X, Y and Z having a value equal to the index of the three failed disks in the array and wherein the disks are labeled from 0-p- 1. Thus, for example, if disks DO, Dl and SP2 from array 500 fail, then X=O, Y=I and Z=3. It should be noted that other orderings of failed disks is possible. For a write operation, the failed disks are X=2, Y=3 and Z=4, ie., to compute the parity for disks SPl, SP2 and RP. However, using the example of procedure 700 wherein X=O, Y=I and Z=3, then g = 1-0=1 and h=3-l=2. By this definition, the Y disk is the middle disk.
In step 714, the RAID system selects a row k. For example, assume k=0. The system then retrieves the sum of the row parity of blocks on the failing (missing) disks corresponding to the selected row (step 715). In this example, the sum for row k = 0 was previously computed as Doo®Doi ΘSPo3 = RPoΘSPo2. In step 716, the system retrieves the diagonal for the block within row k that is on disk Z. Illustratively, assuming k=0, the sum of the missing blocks on this diagonal is D3oθD2i ΘSP03 = DP3 ΘSPi2. The RAID system then retrieves the anti-diagonal for the block on row k on disk X in step 717. Illustratively, this is equal to DOoθDu ΘSP33 = ADP0ΘSP22.
As used herein, the row on which the (anti-)diagonal terminates is termed row q. In step 718, the RAID system retrieves the row parity sum of blocks on the missing disks corresponding to row q, which is illustratively D3oθD31 ΘSP33 = RP3 ΘSP32. In the illustrative embodiment, at each step 715-718, the retrieved sum is XORed together with the previous sum. For the example of row k=0, the total is:
DooθDoi ΘSP03 0D30 ΘD31 ΘSP33 ΘDoo ΘDπ ΘSP33 ΘD30 ΘD2] ®SP03 =
RP0 SSP02 ΘRP3 ΘSP32 ΘADPo ΘSP22 SDP3 θSPi2 which reduces to: Doi 0D11 ΘD21ΘD31 = RP0 ΘSP02 ΘRP3 ΘSP32 ΘADPo 0SP22 θDP3 ΦSPi2
The right hand side of this equation is known, thereby leaving an equation with four unknowns on the middle disk. To solve for the missing data, p such sums are computed by starting crosses at different stripes. Since the array only contains p- 1 rows, the pth 4-tuple sum is formed by constructing a cross using the phantom row D4o, D41, SP42, SP43, and RP4 and the dropped diagonal and anti-diagonal corresponding to Z and X respectively. More generally, duplicate terms in the top and bottom rows at each cross are canceled, thereby resulting in the sum of at most four blocks on the middle disk. In this example, the first cross (based on row 0) results in a tuple comprising
[0,1,2,3] which represents the index of the blocks on the middle disk that are summed. By generating p crosses, the tuple corresponding to row k can be represented by [k, k+g, k+h, k+h+g] wherein all of the addition is performed modulo p.
It should be noted that the step of computing 4-tuple sums can be performed as- suming an arbitrary ordering of the failed disks. In the above example the order chosen is X=O, Y=I and Z=3. Alternately, a different ordering X=O, Y=3 and Z=I could have been chosen. In this alternate case the middle disk would be Y=2 and the values are g=Y-X = 3 and h=Z-Y = -2. It is thus apparent that for 3 disk failures there is a total of 6 possible orderings. Since each ordering results in a different set of values for g and h, the number of XOR operations required to solve the set of linear equations represented by the 4-tuple sums in order to recover the middle disk is different. For example, if the middle disk is recovered by first reducing the set of 4-tuple sums to pairwise sums, then the number of 4-tuples required would be different. Hence, to minimize the number of XOR operations required, the ordering that results in the fewest set of tuples required to reduce to a pair- wise sum is chosen.
Also, the distance between erasure columns is notable in this step. Clearly, if g = h, then two additional blocks on the middle disk cancel out, thereby converting the 4- tuple to a 2-tuple. Hence the next step is not required for equidistant failures. Also, representing the tuple corresponding to row '0' as [0, g, h, h + g], one can generalize equidis- tant failures to the condition g = h mod (p) or [(g - h) mod p = O]. This condition implies that the second and the third blocks within the 4-tuple are identical and hence cancel out.
Once the 4-tuple sums have been constructed, the RAID system illustratively reduces the 4-tuple sums to pairwise sums on the middle disk in step 720. To form a pair- wise sum, the system selects a subset of the equations such that the common blocks can- eel out leaving only two blocks. Computing the sum for the subset results in a pairwise sum of blocks on the middle disk. It should be noted that the 4-tuple sums represent a set of linear equations that can be solved in a number of ways. Reduction to pairwise sums is one illustrative embodiment. There may be multiple ways for choosing a subset of tuples to reduce them to a pairwise sum. In one embodiment the subset of tuples is chosen by starting with a tuple corresponding to row k and chosing subsequent tuples at an offset g (or K) from the last. At each step, common blocks are cancelled and the process continues until only two unknown blocks are left remaining. This results in a pairwise sum. For example, starting with the 4-tuple sum corresponding to row 0, i.e., [0, g, h, g
+ h], it is apparent that choosing another tuple at an offset g, helps cancel two blocks while adding two new blocks, thus keeping the total number of unknown blocks the same. This is because the tuple corresponding to row g is [g, 2g, h + g,-2g + h] which helps cancel common blocks g and h+g since they are present in both tuples, (all addi- tions and multiplications are here assumed to be modulo p). Hence, starting with the 4- tuple corresponding to row 0 (treating this as the first step), and selecting consecutive tuples at an offset g, step m results in the sum of blocks [0, (m*g) mod p, h, (m*g + h)mod
P]-
Given that/> is a prime and the condition {g, h < p} holds, one can always find a m (0 < m < p) such that [ (m*g + h) mod p ] = 0 is true. Similarly, one can always find a m such that [ (m*g - h) mod p] = 0 is true. Hence, by choosing a m such that [(m*g + h) mod p] = 0, the first and the fourth blocks in the result [0, (m*g) mod p, h, (m*g + h)mod p ] can be cancelled after the mth step. Alternatively, by choosing a m such that'[ (m*g - h) mod p ] = 0, the second and the third blocks can be cancelled after the /wth step. Since only two unknown blocks are left, the process of selecting tuples can be terminated at this step. Repeating the above step by starting with 4-tuple sums at each of thep-\ rows results in /7-1 pairwise sums
By performing algebraic operations, one of the equations results in a single unknown equaling a known quantity, which in this case is zero, which may then be substi- tuted into the previous equations to solve for all unknowns on the middle disk, thereby recovering the middle disk in step 725.
Specifically, as the construction of the array uses only/>-l rows, the pth block on disk Y is zero. Thus, the value of the block that is pairwise XORed with the pth block is known at the completion of creation of the pairwise sums, i.e., an equation results with only one unknown. By substituting the solution to the recovered block, and using the other pairwise sums, the remaining blocks on the middle disk may be recovered. At this point, the RAID system has reconstructed the middle disk leaving only two failed disks. As this problem has been solved previously using row diagonal parity, the system imple- ments such R-D parity to recover the missing two disks. Accordingly, once the middle disk has been recovered, the RAID system utilizes R-D parity to recover the remaining two disks in step 730 before completing in step 735. R-D parity is described in the above-incorporated U.S. Patent Application Serial No. 10/035,607, which was filed on December 28, 2001, by Peter F. Corbett for a ROW-DIAGONAL PARITY TECHNIQUE FOR ENABLING EFFICIENT RECOVERY FROM DOUBLE FAILURES IN A STORAGE ARRAY.
It should be noted that the symmetric TP technique of the present invention may also be utilized in recovering from triple failures in other applications that rely on redundant data streams. For example, the TP technique may be utilized in data communication applications, where additional data is transmitted to reduce the reliance on retransmissions to recover from up to three lost and/or corrupted packets. Additionally, it should be noted that in alternate embodiments, other algebraic operations, other than XOR operations, may be utilized.
It should be understood that the principles of the present invention may be utilized with any asymmetrical horizontal erasure code to generate a symmetric variant thereof. As such, the above description relating to symmetric triple erasure code algorithms is to be taken as exemplary only. More generally, any horizontal erasure code and technique may be modified to generate a symmetric variant in accordance with the principles of the present invention by introducing one or more symmetric parity disks and utilizing one or more phantom parity disks containing predetermined values to thereby enable parity encoding and decoding to occur utilizing identical algorithms.
The foregoing description has been directed to specific embodiments of this invention. Additionally, the procedures, processes and/or modules described herein may be implemented in hardware, software, embodied as a computer-readable medium having program instructions, firmware, or a combination thereof. It will be apparent, however, that other variations and modifications may be made to the described embodiments, with the attainment of some or all of their advantages. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the invention.
What is claimed is:

Claims

1. A method for enabling recovery from three or fewer concurrent failures of storage devices in a storage array, the method comprising the steps of: providing the array with a predetermined number of storage devices, including a plurality of first devices configured to store data, and symmetric parity, wherein the pre- determined number of storage devices is p and wherein p is a prime number; dividing each device into blocks; organizing the blocks into stripes that contain a same number of blocks in each device, wherein each stripe comprises p-\ rows of blocks; defining the diagonal parity along diagonal parity sets that span the first devices, wherein the diagonal parity sets wrap around within a group ofp-1 rows so that all blocks belonging to diagonal parity sets of a stripe are stored in the stripe; defining the anti-diagonal parity along anti-diagonal parity sets that span the first devices, wherein the anti-diagonal parity set wraps around within a group ofp-1 rows so that all blocks belonging to the anti-diagonal parity sets of a stripe are stored in the stripe; assigning a predetermined value to the diagonal parity and anti-diagonal parity; and computing parity for the plurality of devices configured to store three parity values using values written to the plurality of devices configured to store data and the predetermined value assigned to the diagonal parity and anti-diagonal parity sets.
2. The method of claim 1 wherein row parity blocks in a stripe are all stored on a single device.
3. The method of claim 1 wherein the step of computing parity comprises computing a total of p 4-tuple sums on one of the first devices configured to store symmetric parity.
4. The method of claim 3 wherein the step of computing further comprises generat- ing the computed 4-tuple sums using a number of crosses.
5. The method of claim 1 wherein the first devices include a plurality of data devices, and wherein the data devices are not all present and absent data devices are treated as containing zero-valued data.
6. The method of claim 1 wherein locations of parity blocks shift from device to device within different stripes.
7. The method of claim 1 wherein the stripes are a power of two number of bits in size.
8. A system configured to enable recovery from three or fewer concurrent failures of storage devices, the system comprising: an array having a predetermined number of storage devices, including a plurality of first devices configured to store data, and symmetric parity, wherein the predetermined number of storage devices isp and wherein/? is a prime number; a storage operating system including a device storage layer configured to implement a symmetric triple parity (TP) technique that (i) computes the diagonal parity along diagonal parity sets that span the first devices,(ii) computes the anti-diagonal parity along anti-diagonal parity sets that span the first devices, wherein the diagonal and anti- diagonal parity are assigned a predetermined value, (iii) computes a total of p 4-tuple sums along an intermediate storage device of the failed storage devices; and (iv) generates the computed 4-tuple sums using a number of crosses; and a processing element configured to execute the storage operating system to thereby invoke storage access operations to and from the array in accordance with the TP parity technique.
9. The system of claim 8 wherein row parity blocks in a stripe are all stored on a single device.
10. The system of claim 8 wherein the device storage layer is a RAID system and wherein the storage devices are disks.
11. The system of claim 8 wherein the RAID system further divides each disk into blocks and organizes the blocks into stripes.
12. The system of claim 11 wherein each stripe comprises p-\ rows of blocks, wherein each row contains one block from each disk.
13. The system of claim 10 wherein the RAID system is implemented in logic cir- cuitry.
14. The system of claim 8 wherein the devices are one of video tape, magnetic tape, optical, DVD, bubble memory, magnetic disk, electronic random access memory and micro-electro mechanical storage devices.
15. The system of claim 8 wherein the first devices include a plurality of data devices, and wherein the data devices are not all present and absent data devices are treated as containing zero-valued data.
16. A computer readable medium containing executable program instructions for enabling recovery from three or fewer concurrent failures of storage devices in a storage array, the executable program instructions comprising program instructions for: providing the array with a predetermined number of storage devices, including a plurality of first devices configured to store data, and symmetric parity, wherein the predetermined number of storage devices \sp and wherein/? is a prime number; dividing each device into blocks; organizing the blocks into stripes that contain a same number of blocks in each device, wherein each stripe comprises p-\ rows of blocks; defining the diagonal parity along diagonal parity sets that span the first devices, wherein the diagonal parity sets wrap around within a group ofp-1 rows so that all blocks belonging to diagonal parity sets of a stripe are stored in the stripe; defining the anti-diagonal parity along anti-diagonal parity sets that span the first devices, wherein the anti-diagonal parity set wraps around within a group ofp-1 rows so that all blocks belonging to the anti-diagonal parity sets of a stripe are stored in the stripe; assigning a predetermined value to the diagonal parity and anti-diagonal parity; and computing parity for the plurality of devices configured to store row parity and symmetric parity using values written to the plurality of devices configured to store data and the predetermined value assigned to the diagonal parity and anti-diagonal parity.
17. A method for enabling recovery from three concurrent failures of storage devices in a storage array, the method comprising the steps of: computing an algebraic operation on missing blocks on each of a set of failed storage devices along a row, a diagonal and an anti-diagonal, wherein a diagonal parity value and an anti-diagonal parity value are set to a predetermined value; and computing a set of 4-tuple sums on a middle failed storage device.
18. The method of claim 17 wherein the step of computing a set of 4-tuple sums on a middle storage device further comprises the steps of: ordering the failed storage devices; for each row of the storage devices:
(a) selecting a row of the storage array; (b) retrieving row parity sum of blocks on missing storage devices corresponding to the selected row;
(c) retrieving diagonal parity sums of blocks on missing storage devices corresponding to the diagonal on the selected row for the third storage device;
(d) retrieving anti-diagonal parity parity sums of blocks on missing storage devices corresponding to the anti-diagonal on the selected row for the first storage device;
(e) retrieving row parity sum of blocks on missing storage devices corresponding to a row where the anti-diagonal and diagonal end; performing an exclusive or (XOR) operation on the results of steps (b), (c), (d) and (e) to form the 4-tuple sum.
19. The method of claim 17 further comprising the steps of: reducing the 4-tuple sums to a set of pairwise sums on the middle failed storage device; recovering the middle storage device.
20. The method of claim 19 further comprising the step of performing a row-diagonal parity recovery technique to recover the other failed storage devices.
21. The method of claim 17 wherein the storage devices are disks.
22. The method of claim 17 wherein the algebraic operation comprises an exclusive or (XOR) sum.
23. The method of claim 17 further comprising the step of solving a set of linear equations represented by the 4-tuple sums of blocks on the middle disk to recover the middle disk.
PCT/US2007/026373 2006-12-29 2007-12-27 System and method for symmetric triple parity WO2008082584A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP07868052A EP2122467A2 (en) 2006-12-29 2007-12-27 System and method for symmetric triple parity

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/648,709 US7613984B2 (en) 2001-12-28 2006-12-29 System and method for symmetric triple parity for failing storage devices
US11/648,709 2006-12-29

Publications (2)

Publication Number Publication Date
WO2008082584A2 true WO2008082584A2 (en) 2008-07-10
WO2008082584A3 WO2008082584A3 (en) 2009-11-05

Family

ID=39555648

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2007/026373 WO2008082584A2 (en) 2006-12-29 2007-12-27 System and method for symmetric triple parity

Country Status (3)

Country Link
US (2) US7613984B2 (en)
EP (1) EP2122467A2 (en)
WO (1) WO2008082584A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8010874B2 (en) 2001-12-28 2011-08-30 Netapp, Inc. Triple parity technique for enabling efficient recovery from triple failures in a storage array
US8402346B2 (en) 2001-12-28 2013-03-19 Netapp, Inc. N-way parity technique for enabling recovery from up to N storage device failures

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7613984B2 (en) 2001-12-28 2009-11-03 Netapp, Inc. System and method for symmetric triple parity for failing storage devices
US8230316B2 (en) 2008-01-25 2012-07-24 Nevion Usa, Inc. Forward error correction for burst and random packet loss for real-time multi-media communication
US8335966B1 (en) * 2008-08-01 2012-12-18 Dell Products L.P. Dual parity RAID wherein no more than N+1 data symbols contribute to any parity symbol
ATE540356T1 (en) 2008-08-12 2012-01-15 Strato Ag DATA STORAGE SYSTEM AND METHOD FOR OPERATING THE SAME
WO2010049928A1 (en) * 2008-10-27 2010-05-06 Kaminario Tehnologies Ltd. System and methods for raid writing and asynchronous parity computation
US8930622B2 (en) 2009-08-11 2015-01-06 International Business Machines Corporation Multi-level data protection for flash memory system
US8176360B2 (en) * 2009-08-11 2012-05-08 Texas Memory Systems, Inc. Method and apparatus for addressing actual or predicted failures in a FLASH-based storage system
US7941696B2 (en) * 2009-08-11 2011-05-10 Texas Memory Systems, Inc. Flash-based memory system with static or variable length page stripes including data protection information and auxiliary protection stripes
US7818525B1 (en) 2009-08-12 2010-10-19 Texas Memory Systems, Inc. Efficient reduction of read disturb errors in NAND FLASH memory
US8189379B2 (en) 2009-08-12 2012-05-29 Texas Memory Systems, Inc. Reduction of read disturb errors in NAND FLASH memory
US8645799B2 (en) 2010-12-31 2014-02-04 Microsoft Corporation Storage codes for data recovery
US8645751B2 (en) 2011-11-02 2014-02-04 Empire Technology Development Llc Triple parity encoding to facilitate data recovery
US9529675B2 (en) 2013-07-26 2016-12-27 Huawei Technologies Co., Ltd. Data recovery method, data recovery device and distributed storage system
CN103412799B (en) * 2013-07-26 2016-08-24 华为技术有限公司 Data reconstruction method, data recovery apparatus and distributed memory system
JP6206160B2 (en) * 2013-12-18 2017-10-04 富士通株式会社 Storage control device, storage control program, and storage control method
WO2016058289A1 (en) * 2015-01-20 2016-04-21 北京大学深圳研究生院 Mds erasure code capable of repairing multiple node failures
US20160328396A1 (en) * 2015-05-06 2016-11-10 Blackfire Research Corporation Collaborative intelligent media playlists
US10146618B2 (en) * 2016-01-04 2018-12-04 Western Digital Technologies, Inc. Distributed data storage with reduced storage overhead using reduced-dependency erasure codes
US10031701B2 (en) 2016-02-09 2018-07-24 International Business Machines Corporation Hierarchical processing for extended product codes
US9940196B2 (en) * 2016-04-21 2018-04-10 Netapp, Inc. Methods and systems for offloading RAID parity reconstruction
CN111124738B (en) * 2018-10-31 2023-08-18 伊姆西Ip控股有限责任公司 Data management method, apparatus and computer program product for redundant array of independent disks
CN111078460B (en) * 2019-11-18 2020-08-25 北京中电兴发科技有限公司 Fast erasure code calculation method

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2011138A (en) 1977-12-23 1979-07-04 Ibm Multi-track tape units

Family Cites Families (86)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3876978A (en) * 1973-06-04 1975-04-08 Ibm Archival data protection
US4092732A (en) * 1977-05-31 1978-05-30 International Business Machines Corporation System for recovering data stored in failed memory unit
US4205324A (en) * 1977-12-23 1980-05-27 International Business Machines Corporation Methods and means for simultaneously correcting several channels in error in a parallel multi channel data system using continuously modifiable syndromes and selective generation of internal channel pointers
JPS5637883A (en) * 1979-09-04 1981-04-11 Fanuc Ltd Information rewrite system
US4467421A (en) * 1979-10-18 1984-08-21 Storage Technology Corporation Virtual storage system and method
GB2061575B (en) * 1979-10-24 1984-09-19 Matsushita Electric Ind Co Ltd Method and apparatus for encoding low redundancy check words from source data
US4435807A (en) * 1980-06-26 1984-03-06 Scott Edward W Orchard error correction system
US4547882A (en) * 1983-03-01 1985-10-15 The Board Of Trustees Of The Leland Stanford Jr. University Error detecting and correcting memories
US4825403A (en) * 1983-05-16 1989-04-25 Data General Corporation Apparatus guaranteeing that a controller in a disk drive system receives at least some data from an invalid track sector
JPS60142418A (en) * 1983-12-28 1985-07-27 Hitachi Ltd Input/output error recovery system
FR2561428B1 (en) * 1984-03-16 1986-09-12 Bull Sa DISC MEMORY RECORDING METHOD AND DISC MEMORY SYSTEM
US4667326A (en) * 1984-12-20 1987-05-19 Advanced Micro Devices, Inc. Method and apparatus for error detection and correction in systems comprising floppy and/or hard disk drives
US5202979A (en) * 1985-05-08 1993-04-13 Thinking Machines Corporation Storage system using multiple independently mechanically-driven storage units
US4722085A (en) * 1986-02-03 1988-01-26 Unisys Corp. High capacity disk storage system having unusually high fault tolerance level and bandpass
JPH0675329B2 (en) * 1986-02-18 1994-09-21 ソニー株式会社 Disc player
US4761785B1 (en) * 1986-06-12 1996-03-12 Ibm Parity spreading to enhance storage access
USRE34100E (en) * 1987-01-12 1992-10-13 Seagate Technology, Inc. Data error correction system
US4775978A (en) * 1987-01-12 1988-10-04 Magnetic Peripherals Inc. Data error correction system
US4796260A (en) * 1987-03-30 1989-01-03 Scs Telecom, Inc. Schilling-Manela forward error correction and detection code method and apparatus
US5257367A (en) * 1987-06-02 1993-10-26 Cab-Tek, Inc. Data storage system with asynchronous host operating system communication link
US4849976A (en) * 1987-08-03 1989-07-18 Scs Telecom, Inc. PASM and TASM forward error correction and detection code method and apparatus
US4849974A (en) * 1987-08-03 1989-07-18 Scs Telecom, Inc. PASM and TASM forward error correction and detection code method and apparatus
US4837680A (en) * 1987-08-28 1989-06-06 International Business Machines Corporation Controlling asynchronously operating peripherals
US4870643A (en) * 1987-11-06 1989-09-26 Micropolis Corporation Parallel drive array storage system
US4847842A (en) * 1987-11-19 1989-07-11 Scs Telecom, Inc. SM codec method and apparatus
US4899342A (en) * 1988-02-01 1990-02-06 Thinking Machines Corporation Method and apparatus for operating multi-unit array of memories
US5077736A (en) * 1988-06-28 1991-12-31 Storage Technology Corporation Disk drive memory
US4989205A (en) * 1988-06-28 1991-01-29 Storage Technology Corporation Disk drive memory
US4989206A (en) * 1988-06-28 1991-01-29 Storage Technology Corporation Disk drive memory
US5128810A (en) * 1988-08-02 1992-07-07 Cray Research, Inc. Single disk emulation interface for an array of synchronous spindle disk drives
US5218689A (en) * 1988-08-16 1993-06-08 Cray Research, Inc. Single disk emulation interface for an array of asynchronously operating disk drives
US5148432A (en) * 1988-11-14 1992-09-15 Array Technology Corporation Arrayed disk drive system and method
US5163131A (en) * 1989-09-08 1992-11-10 Auspex Systems, Inc. Parallel i/o network file server architecture
US5101492A (en) * 1989-11-03 1992-03-31 Compaq Computer Corporation Data redundancy and recovery protection
US5233618A (en) * 1990-03-02 1993-08-03 Micro Technology, Inc. Data correcting applicable to redundant arrays of independent disks
US5088081A (en) * 1990-03-28 1992-02-11 Prime Computer, Inc. Method and apparatus for improved disk access
US5166936A (en) * 1990-07-20 1992-11-24 Compaq Computer Corporation Automatic hard disk bad sector remapping
US5210860A (en) * 1990-07-20 1993-05-11 Compaq Computer Corporation Intelligent disk array controller
US5375128A (en) * 1990-10-18 1994-12-20 Ibm Corporation (International Business Machines Corporation) Fast updating of DASD arrays using selective shadow writing of parity and data blocks, tracks, or cylinders
US5208813A (en) * 1990-10-23 1993-05-04 Array Technology Corporation On-line reconstruction of a failed redundant array system
US5235601A (en) * 1990-12-21 1993-08-10 Array Technology Corporation On-line restoration of redundancy information in a redundant array system
US5274799A (en) * 1991-01-04 1993-12-28 Array Technology Corporation Storage device array architecture with copyback cache
US5579475A (en) * 1991-02-11 1996-11-26 International Business Machines Corporation Method and means for encoding and rebuilding the data contents of up to two unavailable DASDS in a DASD array using simple non-recursive diagonal and row parity
US5271012A (en) * 1991-02-11 1993-12-14 International Business Machines Corporation Method and means for encoding and rebuilding data contents of up to two unavailable DASDs in an array of DASDs
US5179704A (en) * 1991-03-13 1993-01-12 Ncr Corporation Method and apparatus for generating disk array interrupt signals
EP0519669A3 (en) * 1991-06-21 1994-07-06 Ibm Encoding and rebuilding data for a dasd array
US5237658A (en) * 1991-10-01 1993-08-17 Tandem Computers Incorporated Linear and orthogonal expansion of array storage in multiprocessor computing systems
US5305326A (en) * 1992-03-06 1994-04-19 Data General Corporation High availability disk arrays
US5410667A (en) * 1992-04-17 1995-04-25 Storage Technology Corporation Data record copy system for a disk drive array data storage subsystem
US5513192A (en) * 1992-08-28 1996-04-30 Sun Microsystems, Inc. Fault tolerant disk drive system with error detection and correction
DE69431186T2 (en) * 1993-06-03 2003-05-08 Network Appliance Inc Method and file system for assigning file blocks to storage space in a RAID disk system
US5963962A (en) * 1995-05-31 1999-10-05 Network Appliance, Inc. Write anywhere file-system layout
ATE409907T1 (en) * 1993-06-03 2008-10-15 Network Appliance Inc METHOD AND DEVICE FOR DESCRIBING ANY AREAS OF A FILE SYSTEM
US6138126A (en) * 1995-05-31 2000-10-24 Network Appliance, Inc. Method for allocating files in a file system integrated with a raid disk sub-system
JPH08511368A (en) * 1993-06-04 1996-11-26 ネットワーク・アプリアンス・コーポレーション Method for forming parity in RAID subsystem using non-volatile memory
US5537567A (en) * 1994-03-14 1996-07-16 International Business Machines Corporation Parity block configuration in an array of storage devices
US5623595A (en) * 1994-09-26 1997-04-22 Oracle Corporation Method and apparatus for transparent, real time reconstruction of corrupted data in a redundant array data storage system
US5657468A (en) * 1995-08-17 1997-08-12 Ambex Technologies, Inc. Method and apparatus for improving performance in a reduntant array of independent disks
US5812753A (en) * 1995-10-13 1998-09-22 Eccs, Inc. Method for initializing or reconstructing data consistency within an array of storage elements
US5862158A (en) * 1995-11-08 1999-01-19 International Business Machines Corporation Efficient method for providing fault tolerance against double device failures in multiple device systems
US5884098A (en) * 1996-04-18 1999-03-16 Emc Corporation RAID controller system utilizing front end and back end caching systems including communication path connecting two caching systems and synchronizing allocation of blocks in caching systems
US5805788A (en) * 1996-05-20 1998-09-08 Cray Research, Inc. Raid-5 parity generation and data reconstruction
US5950225A (en) * 1997-02-28 1999-09-07 Network Appliance, Inc. Fly-by XOR for generating parity for data gleaned from a bus
KR100267366B1 (en) * 1997-07-15 2000-10-16 Samsung Electronics Co Ltd Method for recoding parity and restoring data of failed disks in an external storage subsystem and apparatus therefor
US6092215A (en) * 1997-09-29 2000-07-18 International Business Machines Corporation System and method for reconstructing data in a storage array system
JP3616487B2 (en) * 1997-11-21 2005-02-02 アルプス電気株式会社 Disk array device
US6138125A (en) * 1998-03-31 2000-10-24 Lsi Logic Corporation Block coding method and system for failure recovery in disk arrays
US6138201A (en) * 1998-04-15 2000-10-24 Sony Corporation Redundant array of inexpensive tape drives using data compression and data allocation ratios
US6247157B1 (en) * 1998-05-13 2001-06-12 Intel Corporation Method of encoding data signals for storage
US6198413B1 (en) * 1999-07-01 2001-03-06 International Business Machines Corporation Partitioned DC balanced (0,6) 16B/18B transmission code with error correction
US6557123B1 (en) * 1999-08-02 2003-04-29 Inostor Corporation Data redundancy methods and apparatus
US6742137B1 (en) 1999-08-17 2004-05-25 Adaptec, Inc. Object oriented fault tolerance
US6532548B1 (en) * 1999-09-21 2003-03-11 Storage Technology Corporation System and method for handling temporary errors on a redundant array of independent tapes (RAIT)
US6581185B1 (en) * 2000-01-24 2003-06-17 Storage Technology Corporation Apparatus and method for reconstructing data using cross-parity stripes on storage media
US6779095B2 (en) * 2000-06-19 2004-08-17 Storage Technology Corporation Apparatus and method for instant copy of data using pointers to new and original data in a data location
US7072916B1 (en) * 2000-08-18 2006-07-04 Network Appliance, Inc. Instant snapshot
US6671772B1 (en) * 2000-09-20 2003-12-30 Robert E. Cousins Hierarchical file system structure for enhancing disk transfer efficiency
US6745284B1 (en) 2000-10-02 2004-06-01 Sun Microsystems, Inc. Data storage subsystem including a storage disk array employing dynamic data striping
US20020124137A1 (en) * 2001-01-29 2002-09-05 Ulrich Thomas R. Enhancing disk array performance via variable parity based load balancing
US6571326B2 (en) * 2001-03-08 2003-05-27 Intel Corporation Space allocation for data in a nonvolatile memory
US7613984B2 (en) 2001-12-28 2009-11-03 Netapp, Inc. System and method for symmetric triple parity for failing storage devices
US7640484B2 (en) * 2001-12-28 2009-12-29 Netapp, Inc. Triple parity technique for enabling efficient recovery from triple failures in a storage array
US7073115B2 (en) * 2001-12-28 2006-07-04 Network Appliance, Inc. Correcting multiple block data loss in a storage array using a combination of a single diagonal parity group and multiple row parity groups
US6993701B2 (en) * 2001-12-28 2006-01-31 Network Appliance, Inc. Row-diagonal parity technique for enabling efficient recovery from double failures in a storage array
US7873700B2 (en) * 2002-08-09 2011-01-18 Netapp, Inc. Multi-protocol storage appliance that provides integrated support for file and block access protocols
US7328305B2 (en) * 2003-11-03 2008-02-05 Network Appliance, Inc. Dynamic parity distribution technique

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2011138A (en) 1977-12-23 1979-07-04 Ibm Multi-track tape units

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHENG WANG, EFFICIENT AND EFFECTIVE SCHEMES FOR STREAMING MEDIA DELIVERY, August 2005 (2005-08-01)
M. BLAUM: "A Family of MDS Array Codes with minimal Number of Encoding Operations", IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, 9 July 2006 (2006-07-09), pages 2784 - 2788, XP031032727

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8010874B2 (en) 2001-12-28 2011-08-30 Netapp, Inc. Triple parity technique for enabling efficient recovery from triple failures in a storage array
US8015472B1 (en) 2001-12-28 2011-09-06 Netapp, Inc. Triple parity technique for enabling efficient recovery from triple failures in a storage array
US8181090B1 (en) 2001-12-28 2012-05-15 Netapp, Inc. Triple parity technique for enabling efficient recovery from triple failures in a storage array
US8402346B2 (en) 2001-12-28 2013-03-19 Netapp, Inc. N-way parity technique for enabling recovery from up to N storage device failures
US8516342B2 (en) 2001-12-28 2013-08-20 Netapp, Inc. Triple parity technique for enabling efficient recovery from triple failures in a storage array

Also Published As

Publication number Publication date
US7979779B1 (en) 2011-07-12
EP2122467A2 (en) 2009-11-25
WO2008082584A3 (en) 2009-11-05
US20080016435A1 (en) 2008-01-17
US7613984B2 (en) 2009-11-03

Similar Documents

Publication Publication Date Title
US7979779B1 (en) System and method for symmetric triple parity for failing storage devices
US8516342B2 (en) Triple parity technique for enabling efficient recovery from triple failures in a storage array
US8402346B2 (en) N-way parity technique for enabling recovery from up to N storage device failures
US9280457B2 (en) System and method for volume block number to disk block number mapping
US7627715B1 (en) Concentrated parity technique for handling double failures and enabling storage of more than one parity block per stripe on a storage device of a storage array
US8495417B2 (en) System and method for redundancy-protected aggregates
US7386757B2 (en) Method and apparatus for enabling high-reliability storage of distributed data on a plurality of independent storage devices
US20090055682A1 (en) Data storage systems and methods having block group error correction for repairing unrecoverable read errors
EP1324200A2 (en) Row-diagonal parity technique for enabling efficient recovery from double failures in a storage array
JP2004030577A (en) Method to correct a plurality of failures of storage device in storage array
US20070180303A1 (en) Uniform and symmetric double failure correcting technique for protecting against two disk failures in a disk array
US6871317B1 (en) Technique for efficiently organizing and distributing parity blocks among storage devices of a storage array
US7398460B1 (en) Technique for efficiently organizing and distributing parity blocks among storage devices of a storage array

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 2007868052

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 07868052

Country of ref document: EP

Kind code of ref document: A2