US5239640A - Data storage system and method including data and checksum write staging storage - Google Patents

Data storage system and method including data and checksum write staging storage Download PDF

Info

Publication number
US5239640A
US5239640A US07/648,944 US64894491A US5239640A US 5239640 A US5239640 A US 5239640A US 64894491 A US64894491 A US 64894491A US 5239640 A US5239640 A US 5239640A
Authority
US
United States
Prior art keywords
data
checksum
write
staging
storage
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
US07/648,944
Inventor
James W. Froemke
David A. Styczinski
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US07/648,944 priority Critical patent/US5239640A/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION, A CORP. OF NEW YORK reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION, A CORP. OF NEW YORK ASSIGNMENT OF ASSIGNORS INTEREST. Assignors: FROEMKE, JAMES W., STYCZINSKI, DAVID A.
Priority to JP3318794A priority patent/JPH087702B2/en
Priority to EP91403305A priority patent/EP0497067A1/en
Application granted granted Critical
Publication of US5239640A publication Critical patent/US5239640A/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1076Parity data used in redundant arrays of independent storages, e.g. in RAID systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/08Error detection or correction by redundancy in data representation, e.g. by using checking codes
    • G06F11/10Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's
    • G06F11/1008Adding special bits or symbols to the coded information, e.g. parity check, casting out 9's or 11's in individual solid state devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/0671In-line storage system
    • G06F3/0673Single storage device

Definitions

  • This invention relates to data processing and more particularly to reliable, economical and fast storage of data in direct access storage devices.
  • Data processing systems are commonly provided with data storage systems in which data is stored for use by the data processing system.
  • a data storage system can include one or more direct access storage devices (DASD's).
  • DASD's direct access storage devices
  • One commonly used type of DASD is a fixed disk drive assembly in which data is written to and read from magnetic disks.
  • a number of DASD's may be employed to provide sufficient capacity to meet the data storage requirements of a data processing system.
  • a variation of the periodic back up approach is to mirror or shadow all stored data by storing it in duplicate on two different DASD's as it is written. If one set of the data is lost due to DASD failure, the redundant set of data can be used in its place.
  • a disadvantage of this method is the expense of the DASD devices required for duplicate data storage.
  • Checksumming data recovery has been suggested as a way to provide acceptable data security without the expense of mirroring all stored data.
  • bits of data at corresponding locations in three or more data DASD's are used in an exclusive -OR calculation and the result is a string of transverse parity bits called a checksum.
  • the checksum data is stored in a checksum DASD so that if a set of the source data is lost, it can be recovered and restored by an exclusive -OR calculation using the remaining data and the checksum data. In this way, the data in three or more DASD's can be protected using one DASD rather than the three or more DASD's that would be required with a data mirroring system.
  • checksum data storage creates a serious bottleneck that further slows system operation.
  • system data read and write commands may be distributed among several data DASD's, the checksum data storage is more concentrated into one DASD or into fewer DASD's. Every time that data is written or updated on any of several data DASD's, the checksum data must be updated. The result is that the flow of read and write system commands is interrupted while the checksum data is maintained. The resulting speed penalty has prevented checksum techniques from fulfilling the goal of complete data protection with both speed and economy.
  • a data storage system in which the cost savings of checksum data protection methods are realized without a penalty in system speed; to provide a system in which the updating of stored checksum data is streamlined and checksum data storage does not result in a bottleneck decreasing the system speed; to provide a system in which the availability of the storage system for system read commands is not blocked by checksum operations; to provide a system in which write command response time is not tied to write command and checksum data update processing; to provide a system in which system availability can be maintained during replacement of a failed data storage device; and to provide a data storage system overcoming disadvantages of those used in the past.
  • a data storage system for a data processing system that provides data read and write commands.
  • the data storage system includes a source data storage facility including a plurality of devices in which source data from the data processing system is written.
  • a calculating unit calculates checksum data from the source data.
  • the system includes a dedicated checksum storage facility in which calculated checksum data is written and a write staging storage area in which calculated checksum data is stored prior to being written in the dedicated checksum storage facility.
  • the present invention provides a method for storing data in a data processing system including the steps of staging write commands including source data and address information by placing them into a RAM staging area and removing the staged write commands from the staging area and writing the source data therefrom onto address locations on a plurality of data DASD's.
  • Checksum data recovery information is calculated form data contained in write commands and is stored in the staging area. The stored checksum information is written onto a dedicated checksum DASD.
  • FIG. 1 is a simplified partial schematic block diagram of a data processing system having a data storage system constructed in accordance with the present invention
  • FIG. 2 is a schematic block diagram showing the flow of data in the data storage system of the present invention.
  • FIG. 3 is a schematic block diagram of functional elements of the data storage system of the present invention.
  • FIG. 4 is a flow chart showing a routine for handling data read commands in the data storage system in accordance with the present invention
  • FIG. 5 is a flow chart showing a routine for handling data write commands in the data storage system in accordance with the present invention.
  • FIG. 6 is a flow chart showing a routine for checksum information in the data storage system in accordance with the present invention.
  • Data processing system 10 includes a central processing unit (CPU) 14 that communicates through a channel adapter 16 with a high speed system channel or bus 18.
  • the storage system 12 of the present invention is coupled to the channel 18 for exchange of data and commands with the system CPU.
  • Other input/output devices such as printers, keyboards, displays and the like may also communicate with the CPU through the system channel 18. Data that is supplied by or under the control of the CPU 14 is transmitted through the channel 18 for storage in the data storage system 12.
  • a write command including data to be stored is applied to the storage system 12 from the CPU.
  • data can be supplied from the storage system 12 for use by the CPU or by other components of the data processing system. This function is accomplished by a read command coupled to the system 12 through the channel 18.
  • the data storage system 12 is implemented outboard and independent of the CPU 14 and includes a channel interface controller 20 governing communications between the system 12 and the channel 18 and a data storage facility 22 in which data received via the channel 18 is written and from which stored data is read.
  • a dedicated checksum data storage facility 24 that is accessed independently of the storage facility 22 is provided for the storage of checksum parity bit information calculated from the data stored in data storage facility 22. The checksum information permits the recovery and restoration of lost data in the event of device failure in the data storage facility 22.
  • An important feature of this invention is the inclusion within the outboard system 12 of a write staging storage area 26, the use and operation of which is regulated and optimized by an intelligent controller 28.
  • the data storage system 12 permits the flow of written and read data indicated schematically in FIG. 2, where the flow of written data is shown in solid lines and the flow of read data is shown in broken lines.
  • Staging storage area 26 is preferably nonvolatile RAM, safe from power interruptions. Speed of the staging storage area 26 is substantially faster than that of the storage facilities 22 and 24, and the capacity of area 26 is relatively large.
  • the intelligent controller 28 preferably operates parallel to and substantially independently of the operation of the CPU 14 so that internal management of the data storage system 12 can proceed without external CPU control.
  • a write command channeled to the data storage system 12 by the CPU 14 including source data for storage is initially placed in the write staging storage area 26 and an acknowledgement of completion of the write command is signalled without waiting for the data to be written to the data storage facility 22. Due to the speed of the write staging storage area 26, this transfer is accomplished quickly and the rapid acknowledgement frees the system resources promptly for other activities while the data storage procedures continue independently within the system 12. (Blocks 514 and 516 of FIG. 5). As a result, the data storage system 12 achieves a fast response time to write commands.
  • the data storage facility 22 includes an array of DASD assemblies 30, 32, 34 and 36, although more or fewer DASD assemblies may be employed depending upon system requirements.
  • the checksum data storage facility 24 is a dedicated, separate DASD assembly 38. Because the checksum data is segregated into a device 38 different from the devices 30-36 in which source data is stored, the reading and writing of data in the checksum device provides minimal interference with access to the data storage devices.
  • Data initially stored in the write staging storage area 26 is subsequently written to the DASD assemblies 30-36 and checksum information is calculated and written to the checksum DASD assembly 38.
  • These data calculation and writing functions are controlled as described below independently of the operation of the CPU 14 and without substantially reducing the accessibility of the storage system 12 to further write and read commands.
  • the intelligent controller 28 includes a read command processor 40 and a write command processor 42 both connected to the channel interface 20 for receiving read commands and write commands transmitted through the channel 18.
  • a control unit 44 for access to data stored in the write staging storage area 26 and a control unit 46 for queuing of the data stored in the write staging storage area 26.
  • Each DASD assembly 30-38 includes a DASD device preferably in the form of a magnetic disk drive assembly and a small, dedicated track buffer for holding a track of data to be written to the corresponding device.
  • the command processors 40 and 42 are connected to communicate with one another and with the data DASD assemblies 30-36.
  • the write command processor 42 also communicates with the write staging storage area 26, with the checksum DASD assembly 38 and with a checksum calculation unit 48.
  • a typical checksum data recovery method may be employed. As an example, assume that the following data bits are stored at the same physical addresses at the four DASD assemblies 30-36. The four sets of data bits are identified as data 1-4 in the following table. The checksum bits are obtained by performing exclusive -OR calculations on the columns of bits to determine the transverse parity of the columns of bits as indicated in the checksum bits row. The checksum data is stored for use in data recovery.
  • checksum data When data is replaced by overwriting, the checksum data is also updated. Assume that data 2 of the above table 1 is to be replaced with new data 2. It is possible to calculate the new checksum bits by determining the transverse parity bits of all the source data:
  • this calculation requires reading data from all of the data DASD assemblies, it is preferred to calculate the new checksum data from the new and old data 2 and from the existing checksum data. First, an exclusive -OR calculation is made with the new and the replaced data.
  • the use of the dedicated DASD assembly 38 for checksum data storage permits the data DASD assemblies 30-36 to be accessed for execution of read commands during writes to the DASD assembly 38.
  • the data DASD assemblies 30-36 need not all have the same capacity.
  • the DASD assembly 38 should be as large as the largest data DASD or as large as the largest quantity of data on any one data DASD assembly that is to be protected. If desired, more than one checksum DASD assembly may be used.
  • all write commands coupled through the channel interface 20 are directed by the write command processor 42 of the controller 28 to the write staging storage area 26.
  • the controller 28 signals completed execution of the write command to the CPU through interface 20 and channel 18 and write excecution delays are avoided. (Blocks 514 and 516 of FIG. 5).
  • the data to be replaced by the contents of the newly stored write command is read from the target DASD and is supplied to the checksum calculation unit 48 along with the new data.
  • the resulting exclusive or bits are stored in the write staging storage area 26.
  • the write staging storage area 26 has a relatively large capacity to permit the use of data handling techniques that streamline data throughput and minimize the duration of access to the checksum DASD assembly 38, thereby to avoid the checksum bottleneck problem experienced with past systems.
  • the capacity of the storage area 26 is large enough to hold hundreds or thousands of write commands along with the associated checksum information.
  • advantages of the invention can be realized with a write staging storage area of about 1 Mb minimum size, with the range of 2 Mb to 16 Mb and a specific value of about 8 Mb being preferred.
  • Source data and checksum data is written to the DASD assemblies 30-38 by transducer heads that sweep along paths over the surfaces of the magnetic media of the DASD's.
  • Data storage sectors in tracks defined on the media have addresses that are used to locate the heads in registration with specific regions of the media for writing or reading data, and these addresses have a sequence related to the head sweep path. For example, in typical DASD assemblies, the heads sweep from the lowest physical data address to the highest physical data address.
  • One reason for using a relatively large staging storage area is to permit the use of data queuing techniques for enhancing performance of the system by coordinating the writing of data to the DASD assembly 38 with the physical data addresses along the sweep path to enhance the speed of operation of the DASD assembly 38.
  • the write command queuing control unit 46 includes data blocking and sweep ordering functions. Normally the staging storage area 26 is filled with numerous write commands ready for writing to one of the data DASD's 30-36. Each write command is associated with a corresponding exclusive or checksum calculation result provided by the checksum calculation unit. This exclusive or checksum data is used for calculating new checksum data to be written on the DASD assembly 38.
  • the write command source data and the checksum data have the same physical addresses but on different DASD assemblies: the checksum data on DASD assembly 38 and the source data on one of the DASD assemblies 30-36.
  • the DASD assembly 38 When data is removed from the write staging storage area 26 for writing to the DASD assemblies, the DASD assembly 38 is read to obtain the existing checksum data.
  • the existing checksum data together with the checksum calculation data from the area 26 are supplied to the checksum calculation unit 48 where an exclusive or calculation is made to obtain the new checksum data that is then written to the DASD assembly 38.
  • the unit 48 is fast enough to permit the old data to be read and then to permit the new data to be written in consecutive or close passes of the magnetic media under the transducer head.
  • Another alternative is to read the existing checksum data and make the exclusive or calculation and store that result, namely the new checksum data, in the write staging storage area 26.
  • the checksum data is written from the area 26 onto the DASD assembly 38 without the need for further calculations.
  • the corresponding new source data is written to the same physical address on one of the data DASD assemblies 30-36.
  • the data blocking function of the control unit 46 is carried out by identifying data that has adjacent and consecutive addresses. While it resides in area 26, this data is combined into a block of data contained in a single write command before being written onto the DASD assemblies. (Blocks 608 and 610 of FIG. 6). This speeds the write operation because it is faster to execute one longer command than a number of shorter commands.
  • the sweep function of the control unit 46 is carried out by ordering the stored data with its physical addresses in the same order as the head sweep path of the checksum DASD 38. Data queued in this order flows to the DASD assembly as the head sweeps the storage media, and data is written with no unnecessary head movement. (Blocks 612 and 614 of FIG. 6). The time lost to head seeking movement is minimized.
  • the storage area is large enough so that in typical operation, checksum data is written on the average in about every third or fourth track as the head sweeps the data storage medium.
  • the staged data access control unit 44 permits access to and use of data stored in the staging storage area 26 prior to the writing of that data to one of the DASD assemblies.
  • the control unit 44 includes a fast read function. When a read command is received by the read command processor 40, the control unit 44 is interrogated to determine whether any of the requested data is present in the write staging storage 26. If so, then that data is provided to the read command processor and the read command is executed without the need to access any of the DASD assemblies. 404 and 406). This provides a speed increase because RAM speed is substantially faster than DASD speed.
  • a typical DASD may have an access time with an order of magnitude in the millisecond range, while typical RAM storage may provide data access times in the nanosecond range.
  • Requested data may also be present in the track buffers of the DASD assemblies 30-36 and this provides an additional source of fast reads.
  • Staged data access control unit 44 also has a write preempt function. Write commands that are received by the write command processor 42 are checked to determine if they are overwrites of data that is held in the staging storage area 26. If so, the data to be replaced is removed from the data writing queue and is provided along with the new write command to the checksum calculation unit 48 for a new exclusive or calculation. The new data and the corresponding checksum data are then added to the write staging storage area 26. (Block 510 of FIG. 5 and block 606 of FIG. 6). This speeds the operation of the system 12 by avoiding the writing of overwritten data to the DASD assemblies.
  • the write command can be stored in the area 26 without reading of the preexisting data to be replaced.
  • the data DASD and the checksum DASD would both be read and the exclusive or calculations would be made when the write command reaches the head of the queue.
  • the advantage is that a write preempt would eliminate not only the final checksum calculation and data and checksum writes but also would avoid the need for the initial data read and checksum calculation.
  • the checksum calculation unit may be incorporated into the controller 28 or into the write staging storage area 26 rather than being a discrete unit.
  • the staging storage area 26 may include DRAM's or SRAM's provided with battery back up power to be immune from power interruptions.
  • volatile RAM may be used in conjunction with a DASD write journal for recovery of RAM contents after a power loss.
  • Other non-volatile RAM such as charge wells or magnetic particles may be used.

Abstract

An outboard data storage system for a data processing system including a plurality of data DASD's and a dedicated checksum DASD. Write commands and checksum data calculated from the write commands are placed in a staging area under the control of an intelligent controller, operating in parallel with the data processing system CPU, that acknowledges execution of write commands as they are staged. The controller blocks data with contiguous addresses into single write commands and queues the write commands in an order corresponding to physical DASD addresses. Fast reads are achieved by executing read commands directly from the staging storage area and DASD writes are reduced by preempting write commands in the staging storage area with new write commands having the same address.

Description

FIELD OF THE INVENTION
This invention relates to data processing and more particularly to reliable, economical and fast storage of data in direct access storage devices.
DESCRIPTION OF THE PRIOR ART
Data processing systems are commonly provided with data storage systems in which data is stored for use by the data processing system. A data storage system can include one or more direct access storage devices (DASD's). One commonly used type of DASD is a fixed disk drive assembly in which data is written to and read from magnetic disks. A number of DASD's may be employed to provide sufficient capacity to meet the data storage requirements of a data processing system.
One serious concern with data storage systems is the integrity of stored data. DASD's are subject to physical and mechanical failure and under such circumstances it can be difficult or impossible to recover data stored on the DASD. Various approaches have been used to decrease the possibility that data may be lost due to DASD failure. A common method is to periodically back up stored data by copying it to another storage system or device. But this does not avoid loss of data written to the storage system more recently than the last periodic back up.
A variation of the periodic back up approach is to mirror or shadow all stored data by storing it in duplicate on two different DASD's as it is written. If one set of the data is lost due to DASD failure, the redundant set of data can be used in its place. A disadvantage of this method is the expense of the DASD devices required for duplicate data storage.
Checksumming data recovery has been suggested as a way to provide acceptable data security without the expense of mirroring all stored data. In a checksum system, bits of data at corresponding locations in three or more data DASD's are used in an exclusive -OR calculation and the result is a string of transverse parity bits called a checksum. The checksum data is stored in a checksum DASD so that if a set of the source data is lost, it can be recovered and restored by an exclusive -OR calculation using the remaining data and the checksum data. In this way, the data in three or more DASD's can be protected using one DASD rather than the three or more DASD's that would be required with a data mirroring system.
Another important concern relating to data storage is speed of data storage and of data access. Data checksumming has not been an acceptable solution to the data integrity problem because of decreased system speed. In a checksum system, when stored data is revised and overwritten with new data, the checksum data must also be changed. In order to update the checksum data, the exclusive -OR of the old and new source data is calculated, and then the exclusive -OR of the result and the original checksum data is calculated to arrive at the new checksum data. System speed is impaired because for each data update, the data DASD must first be read and then rewritten, and the checksum DASD must also be read and then rewritten. The four read and write commands require time and, in addition, while the data DASD is in use, it is unavailable for read commands channeled from the data processing system.
Another problem with checksum systems as implemented in the past is that the checksum data storage creates a serious bottleneck that further slows system operation. Although system data read and write commands may be distributed among several data DASD's, the checksum data storage is more concentrated into one DASD or into fewer DASD's. Every time that data is written or updated on any of several data DASD's, the checksum data must be updated. The result is that the flow of read and write system commands is interrupted while the checksum data is maintained. The resulting speed penalty has prevented checksum techniques from fulfilling the goal of complete data protection with both speed and economy.
SUMMARY OF THE INVENTION
Among the important objects of the present invention are to provide a data storage system in which the cost savings of checksum data protection methods are realized without a penalty in system speed; to provide a system in which the updating of stored checksum data is streamlined and checksum data storage does not result in a bottleneck decreasing the system speed; to provide a system in which the availability of the storage system for system read commands is not blocked by checksum operations; to provide a system in which write command response time is not tied to write command and checksum data update processing; to provide a system in which system availability can be maintained during replacement of a failed data storage device; and to provide a data storage system overcoming disadvantages of those used in the past.
In brief, in accordance with the present invention there is provided a data storage system for a data processing system that provides data read and write commands. The data storage system includes a source data storage facility including a plurality of devices in which source data from the data processing system is written. A calculating unit calculates checksum data from the source data. The system includes a dedicated checksum storage facility in which calculated checksum data is written and a write staging storage area in which calculated checksum data is stored prior to being written in the dedicated checksum storage facility.
In brief, the present invention provides a method for storing data in a data processing system including the steps of staging write commands including source data and address information by placing them into a RAM staging area and removing the staged write commands from the staging area and writing the source data therefrom onto address locations on a plurality of data DASD's. Checksum data recovery information is calculated form data contained in write commands and is stored in the staging area. The stored checksum information is written onto a dedicated checksum DASD.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention together with the above and other objects and advantages may be best understood from the following detailed description of the embodiments of the invention shown in the accompanying drawings, wherein:
FIG. 1 is a simplified partial schematic block diagram of a data processing system having a data storage system constructed in accordance with the present invention;
FIG. 2 is a schematic block diagram showing the flow of data in the data storage system of the present invention; and
FIG. 3 is a schematic block diagram of functional elements of the data storage system of the present invention.
FIG. 4 is a flow chart showing a routine for handling data read commands in the data storage system in accordance with the present invention;
FIG. 5 is a flow chart showing a routine for handling data write commands in the data storage system in accordance with the present invention; and
FIG. 6 is a flow chart showing a routine for checksum information in the data storage system in accordance with the present invention.
DETAILED DESCRIPTION OF THE INVENTION
With reference to the drawings and first to FIG. 1, there is shown part of a data processing or computer system generally designated as 10 having a data storage system generally designated as 12 and constructed in accordance with the principles of the present invention. Data processing system 10 includes a central processing unit (CPU) 14 that communicates through a channel adapter 16 with a high speed system channel or bus 18. The storage system 12 of the present invention is coupled to the channel 18 for exchange of data and commands with the system CPU. Other input/output devices (not shown) such as printers, keyboards, displays and the like may also communicate with the CPU through the system channel 18. Data that is supplied by or under the control of the CPU 14 is transmitted through the channel 18 for storage in the data storage system 12. To accomplish this, a write command including data to be stored is applied to the storage system 12 from the CPU. Conversely, data can be supplied from the storage system 12 for use by the CPU or by other components of the data processing system. This function is accomplished by a read command coupled to the system 12 through the channel 18.
In general, the data storage system 12 is implemented outboard and independent of the CPU 14 and includes a channel interface controller 20 governing communications between the system 12 and the channel 18 and a data storage facility 22 in which data received via the channel 18 is written and from which stored data is read. A dedicated checksum data storage facility 24 that is accessed independently of the storage facility 22 is provided for the storage of checksum parity bit information calculated from the data stored in data storage facility 22. The checksum information permits the recovery and restoration of lost data in the event of device failure in the data storage facility 22.
An important feature of this invention is the inclusion within the outboard system 12 of a write staging storage area 26, the use and operation of which is regulated and optimized by an intelligent controller 28. The data storage system 12 permits the flow of written and read data indicated schematically in FIG. 2, where the flow of written data is shown in solid lines and the flow of read data is shown in broken lines. Staging storage area 26 is preferably nonvolatile RAM, safe from power interruptions. Speed of the staging storage area 26 is substantially faster than that of the storage facilities 22 and 24, and the capacity of area 26 is relatively large. The intelligent controller 28 preferably operates parallel to and substantially independently of the operation of the CPU 14 so that internal management of the data storage system 12 can proceed without external CPU control.
Referring to FIG. 2, a write command channeled to the data storage system 12 by the CPU 14 including source data for storage is initially placed in the write staging storage area 26 and an acknowledgement of completion of the write command is signalled without waiting for the data to be written to the data storage facility 22. Due to the speed of the write staging storage area 26, this transfer is accomplished quickly and the rapid acknowledgement frees the system resources promptly for other activities while the data storage procedures continue independently within the system 12. ( Blocks 514 and 516 of FIG. 5). As a result, the data storage system 12 achieves a fast response time to write commands.
In the illustrated embodiment of the invention, the data storage facility 22 includes an array of DASD assemblies 30, 32, 34 and 36, although more or fewer DASD assemblies may be employed depending upon system requirements. The checksum data storage facility 24 is a dedicated, separate DASD assembly 38. Because the checksum data is segregated into a device 38 different from the devices 30-36 in which source data is stored, the reading and writing of data in the checksum device provides minimal interference with access to the data storage devices.
Data initially stored in the write staging storage area 26 is subsequently written to the DASD assemblies 30-36 and checksum information is calculated and written to the checksum DASD assembly 38. These data calculation and writing functions are controlled as described below independently of the operation of the CPU 14 and without substantially reducing the accessibility of the storage system 12 to further write and read commands.
Because data to be written is stored in the write staging storage area 26, certain data handling advantages are possible. If new data is written to the same address as data held in area 26, the first write is preempted by the second and the first data in area 26 is replaced by the second data without writing the first data to a DASD assembly. Blocks 504 and 510 of FIG. 5). If a command is received to read data that is present in the write staging storage area 26, the data is read from the area 26 quickly and without the necessity for accessing a DASD assembly. ( Blocks 404 and 406 of FIG. 4). Both of these operations increase data handling speed because they eliminate a relatively slow DASD access. Commands to read data that is not held in the staging storage area 26 are handled in the conventional manner by reading data from the DASD assemblies 30-36.
Functional elements of the data storage system 12 are seen in more detail in FIG. 3. The intelligent controller 28 includes a read command processor 40 and a write command processor 42 both connected to the channel interface 20 for receiving read commands and write commands transmitted through the channel 18. Associated with the write command processor 42 is a control unit 44 for access to data stored in the write staging storage area 26 and a control unit 46 for queuing of the data stored in the write staging storage area 26. Each DASD assembly 30-38 includes a DASD device preferably in the form of a magnetic disk drive assembly and a small, dedicated track buffer for holding a track of data to be written to the corresponding device. The command processors 40 and 42 are connected to communicate with one another and with the data DASD assemblies 30-36. The write command processor 42 also communicates with the write staging storage area 26, with the checksum DASD assembly 38 and with a checksum calculation unit 48.
A typical checksum data recovery method may be employed. As an example, assume that the following data bits are stored at the same physical addresses at the four DASD assemblies 30-36. The four sets of data bits are identified as data 1-4 in the following table. The checksum bits are obtained by performing exclusive -OR calculations on the columns of bits to determine the transverse parity of the columns of bits as indicated in the checksum bits row. The checksum data is stored for use in data recovery.
              TABLE 1                                                     
______________________________________                                    
1 0 0 1 0 1 0 0     Data 1                                                
1 1 1 1 0 1 0 1     Data 2                                                
0 0 0 1 1 0 1 1     Data 3                                                
1 1 1 1 0 0 1 1     Data 4                                                
1 0 0 0 1 0 0 1     checksum bits.                                        
______________________________________                                    
If data on one of the DASD assemblies 30-36 is lost, it can be recovered using the checksum information. For example, if Data 3 from table 1 is lost, an exclusive -OR calculation is made on the remaining source data and on the checksum data and the transverse parity bits are the same as the lost data bits:
              TABLE 2                                                     
______________________________________                                    
1 0 0 0 1 0 0 1    Checksum bits.                                         
1 0 0 1 0 1 0 0    Data 1                                                 
1 1 1 1 0 1 0 1    Data 2                                                 
1 1 1 1 0 0 1 1    Data 4                                                 
0 0 0 1 1 0 1 1    Recovered data 3                                       
______________________________________                                    
When data is replaced by overwriting, the checksum data is also updated. Assume that data 2 of the above table 1 is to be replaced with new data 2. It is possible to calculate the new checksum bits by determining the transverse parity bits of all the source data:
              TABLE 3                                                     
______________________________________                                    
1 0 0 1 0 1 0 0   Data 1                                                  
1 0 1 1 1 0 0 0   New data 2                                              
0 0 0 1 1 0 1 1   Data 3                                                  
1 1 1 1 0 0 1 1   Data 4                                                  
1 1 0 0 0 1 0 0   New checksum bits                                       
______________________________________                                    
Because this calculation requires reading data from all of the data DASD assemblies, it is preferred to calculate the new checksum data from the new and old data 2 and from the existing checksum data. First, an exclusive -OR calculation is made with the new and the replaced data.
              TABLE 4                                                     
______________________________________                                    
1 1 1 1 0 1 0 1  Old data 2                                               
1 0 1 1 1 0 0 0  New data 2                                               
0 1 0 0 1 1 0 1  Data 2 exclusive-OR bits                                 
______________________________________                                    
Then, the resulting data 2 exclusive -OR bits are compared in an exclusive -OR calculation with the existing checksum bits to be replaced in order to obtain the new checksum bits.
              TABLE 5                                                     
______________________________________                                    
0 1 0 0 1 1 0 1  Data 2 exclusive-OR bits                                 
1 0 0 0 1 0 0 1  Old checksum bits                                        
1 1 0 0 0 1 0 0  New checksum bits                                        
______________________________________                                    
This type of checksum data recovery approach has introduced undesirable delay in past implementations. Every write command requires that the existing data be read, that an exclusive or calculation be made, and that the new data be written. In addition, the existing checksum data must be read, another exclusive or calculation must be made, and the new checksum data must be written. These reads and writes interfere with other read commands and this interference contributes to degraded performance of checksum systems. This interference is reduced in the system 12 by avoiding unnecessary write commands to the DASD's 30-36 as described below. Although the source data reads and writes are distributed among several DASD's, the checksum data reads and writes are concentrated into one DASD or a smaller number of DASD's. Every write command requires checksum DASD access and the resulting bottleneck increase the response time of the storage system and slows the data processing system.
As indicated previously, the use of the dedicated DASD assembly 38 for checksum data storage permits the data DASD assemblies 30-36 to be accessed for execution of read commands during writes to the DASD assembly 38. The data DASD assemblies 30-36 need not all have the same capacity. The DASD assembly 38 should be as large as the largest data DASD or as large as the largest quantity of data on any one data DASD assembly that is to be protected. If desired, more than one checksum DASD assembly may be used.
In the system 12 of the present invention, all write commands coupled through the channel interface 20 are directed by the write command processor 42 of the controller 28 to the write staging storage area 26. When a write command is stored in area 26, the controller 28 signals completed execution of the write command to the CPU through interface 20 and channel 18 and write excecution delays are avoided. ( Blocks 514 and 516 of FIG. 5). Independently of continued operation of the data processing system 10, the data to be replaced by the contents of the newly stored write command is read from the target DASD and is supplied to the checksum calculation unit 48 along with the new data. The resulting exclusive or bits are stored in the write staging storage area 26.
Unlike typical disk buffer and cache arrangements, the write staging storage area 26 has a relatively large capacity to permit the use of data handling techniques that streamline data throughput and minimize the duration of access to the checksum DASD assembly 38, thereby to avoid the checksum bottleneck problem experienced with past systems. Preferably the capacity of the storage area 26 is large enough to hold hundreds or thousands of write commands along with the associated checksum information. For example, in a data processing system with DASD devices having capacities of about 500 megabytes (Mb) or more, average write command lengths of about 4 kilobytes (Kb) and a maximum write command length of 256 Kb, advantages of the invention can be realized with a write staging storage area of about 1 Mb minimum size, with the range of 2 Mb to 16 Mb and a specific value of about 8 Mb being preferred.
Source data and checksum data is written to the DASD assemblies 30-38 by transducer heads that sweep along paths over the surfaces of the magnetic media of the DASD's. Data storage sectors in tracks defined on the media have addresses that are used to locate the heads in registration with specific regions of the media for writing or reading data, and these addresses have a sequence related to the head sweep path. For example, in typical DASD assemblies, the heads sweep from the lowest physical data address to the highest physical data address. One reason for using a relatively large staging storage area is to permit the use of data queuing techniques for enhancing performance of the system by coordinating the writing of data to the DASD assembly 38 with the physical data addresses along the sweep path to enhance the speed of operation of the DASD assembly 38.
As indicated in FIG. 3, the write command queuing control unit 46 includes data blocking and sweep ordering functions. Normally the staging storage area 26 is filled with numerous write commands ready for writing to one of the data DASD's 30-36. Each write command is associated with a corresponding exclusive or checksum calculation result provided by the checksum calculation unit. This exclusive or checksum data is used for calculating new checksum data to be written on the DASD assembly 38. Preferably, the write command source data and the checksum data have the same physical addresses but on different DASD assemblies: the checksum data on DASD assembly 38 and the source data on one of the DASD assemblies 30-36.
When data is removed from the write staging storage area 26 for writing to the DASD assemblies, the DASD assembly 38 is read to obtain the existing checksum data. The existing checksum data together with the checksum calculation data from the area 26 are supplied to the checksum calculation unit 48 where an exclusive or calculation is made to obtain the new checksum data that is then written to the DASD assembly 38. Preferably the unit 48 is fast enough to permit the old data to be read and then to permit the new data to be written in consecutive or close passes of the magnetic media under the transducer head.
Another alternative is to read the existing checksum data and make the exclusive or calculation and store that result, namely the new checksum data, in the write staging storage area 26. With this approach, the checksum data is written from the area 26 onto the DASD assembly 38 without the need for further calculations. At or prior to the time that the new checksum data is written to the DASD assembly 38, the corresponding new source data is written to the same physical address on one of the data DASD assemblies 30-36. In order to maximize the benefit of the write preempt function described below, it is preferred that the the new source data be held in the write staging storage area 26 and that the writing of that data to the DASD's 30-36 be delayed until writing of the corresponding checksum data is to be accomplished.
The data blocking function of the control unit 46 is carried out by identifying data that has adjacent and consecutive addresses. While it resides in area 26, this data is combined into a block of data contained in a single write command before being written onto the DASD assemblies. (Blocks 608 and 610 of FIG. 6). This speeds the write operation because it is faster to execute one longer command than a number of shorter commands. The sweep function of the control unit 46 is carried out by ordering the stored data with its physical addresses in the same order as the head sweep path of the checksum DASD 38. Data queued in this order flows to the DASD assembly as the head sweeps the storage media, and data is written with no unnecessary head movement. ( Blocks 612 and 614 of FIG. 6). The time lost to head seeking movement is minimized. In a preferred embodiment of the invention, the storage area is large enough so that in typical operation, checksum data is written on the average in about every third or fourth track as the head sweeps the data storage medium.
The staged data access control unit 44 permits access to and use of data stored in the staging storage area 26 prior to the writing of that data to one of the DASD assemblies. The control unit 44 includes a fast read function. When a read command is received by the read command processor 40, the control unit 44 is interrogated to determine whether any of the requested data is present in the write staging storage 26. If so, then that data is provided to the read command processor and the read command is executed without the need to access any of the DASD assemblies. 404 and 406). This provides a speed increase because RAM speed is substantially faster than DASD speed. A typical DASD may have an access time with an order of magnitude in the millisecond range, while typical RAM storage may provide data access times in the nanosecond range. Requested data may also be present in the track buffers of the DASD assemblies 30-36 and this provides an additional source of fast reads.
Staged data access control unit 44 also has a write preempt function. Write commands that are received by the write command processor 42 are checked to determine if they are overwrites of data that is held in the staging storage area 26. If so, the data to be replaced is removed from the data writing queue and is provided along with the new write command to the checksum calculation unit 48 for a new exclusive or calculation. The new data and the corresponding checksum data are then added to the write staging storage area 26. (Block 510 of FIG. 5 and block 606 of FIG. 6). This speeds the operation of the system 12 by avoiding the writing of overwritten data to the DASD assemblies.
As an alternative to reading the existing data and performing an exclusive or calculation following receipt of a write command, the write command can be stored in the area 26 without reading of the preexisting data to be replaced. In this approach, the data DASD and the checksum DASD would both be read and the exclusive or calculations would be made when the write command reaches the head of the queue. The advantage is that a write preempt would eliminate not only the final checksum calculation and data and checksum writes but also would avoid the need for the initial data read and checksum calculation.
The checksum calculation unit may be incorporated into the controller 28 or into the write staging storage area 26 rather than being a discrete unit. The staging storage area 26 may include DRAM's or SRAM's provided with battery back up power to be immune from power interruptions. Alternatively, volatile RAM may be used in conjunction with a DASD write journal for recovery of RAM contents after a power loss. Other non-volatile RAM such as charge wells or magnetic particles may be used.
In the event of power interruption, data storage status is preserved and there is no loss of data not yet written to the DASD assemblies 30-38. In addition, the availability of the storage system 12 can be maintained if one of the data DASD assemblies fails and must be replaced. The other DASD assemblies are available for execution of read commands and the data formerly stored in the failed data DASD assembly is available for access and recovery purposes using the checksum data and other stored source data. In addition, if one data DASD is disabled, write commands can continue to be executed.

Claims (17)

We claim:
1. A data storage system for a data processing system that provides data read and write commands, said data storage system comprising in combination:
a data storage facility including a plurality of devices in which data from the data processing system is written;
calculating means for calculating checksum data from said data;
a dedicated checksum storage facility in which calculated checksum data is written; and
a write staging storage area shared by said plurality of devices and by said dedicated checksum storage facility in which calculated checksum data and data from the data processing system are stored prior to writing checksum data in said dedicated checksum storage facility.
2. A data storage system as claimed in claim 1 in which said data storage facility includes a plurality of DASD's and said dedicated checksum storage facility includes another DASD.
3. A data storage system as claimed in claim 2 wherein said write staging storage area comprises RAM.
4. A data storage system as claimed in claim 2 wherein said write staging storage area comprises non-volatile RAM.
5. A data storage system as claimed in claim 1 further comprising a controller for controlling the data storage system in response to read and write commands, said controller operating in parallel with the data processing system.
6. A data storage system for a data processing system that provides data read and write commands, said data storage system comprising in combination:
a data storage facility including a plurality of devices in which data from the data processing system is written;
calculating means for calculating checksum data from said data;
a dedicated checksum storage facility in which calculated checksum data is written;
a write staging storage area in which calculated checksum data is stored prior to writing checksum data in said dedicated checksum storage facility; and
a controller for controlling the data storage system in response to read and write commands, said controller operating in parallel with the data processing system;
said controller including queuing means for ordering checksum data in said write staging storage area in an order corresponding to addresses in said dedicated checksum storage facility.
7. A data storage system for a data processing system that provides data read and write commands, said data storage system comprising in combination:
a data storage facility including a plurality of devices in which data from the data processing system is written;
calculating means for calculating checksum data from said data;
a dedicated checksum storage facility in which calculated checksum data is written;
a write staging storage area in which calculated checksum data is stored prior to writing checksum data in said dedicated checksum storage facility; and
a controller for controlling the data storage system in response to read and write commands, said controller operating in parallel with the data processing system;
said controller including blocking means for combining checksum data having adjacent addresses into single write commands.
8. A data storage system for a data processing system that provides data read and write commands, said data storage system comprising in combination:
a data storage facility including a plurality of devices in which data from the data processing system is written;
calculating means for calculating checksum data from said data;
a dedicated checksum storage facility in which calculated checksum data is written;
a write staging storage area in which calculated checksum data is stored prior to writing checksum data in said dedicated checksum storage facility; and
a controller for controlling the data storage system in response to read and write commands, said controller operating in parallel with the data processing system;
said controller including write preempt means for replacing write commands in said write staging storage area with newer write commands having the same address.
9. A data storage system for a data processing system that provides data read and write commands, said data storage system comprising in combination:
a data storage facility including a plurality of devices in which data from the data processing system is written;
calculating means for calculating checksum data from said data;
a dedicated checksum storage facility in which calculated checksum data is written;
a write staging storage area in which calculated checksum data is stored prior to writing checksum data in said dedicated checksum storage facility; and
a controller for controlling the data storage system in response to read and write commands, said controller operating in parallel with the data processing system;
said controller including read command processing means for executing read commands by reading data from said write staging storage area.
10. A method of storing data in a data processing system comprising the steps of:
staging write commands including data and address information by placing them into a common RAM staging area;
removing staged write commands from the staging area and writing the data therefrom onto address locations on a plurality of data DASD's;
calculating checksum data recovery information from data contained in write commands;
storing in the common staging area checksum information corresponding to staged write commands; and
writing checksum information onto a dedicated checksum DASD.
11. The method of claim 10 further comprising identifying checksum information stored in the staging area having contiguous addresses and combining the located checksum information into single blocks.
12. The method of claim 11 further comprising queuing checksum information stored in the staging area into an address sequence corresponding to physical addresses in the checksum DASD.
13. The method of claim 10 further comprising executing read commands for staged data by reading the staged data from the staging area.
14. The method of claim 10 further comprising preempting staged write commands by replacing them in the staging area with new write commands having the same address.
15. A data storage system as claimed in claim 1 wherein said write staging storage area has a capacity of more than about one megabyte.
16. A data storage system as claimed in claim 15 wherein said write staging storage area has a capacity in the range of about two to about sixteen megabytes.
17. A data storage system as claimed in claim 16 wherein said write staging storage area has a capacity of about two to about eight megabytes.
US07/648,944 1991-02-01 1991-02-01 Data storage system and method including data and checksum write staging storage Expired - Fee Related US5239640A (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US07/648,944 US5239640A (en) 1991-02-01 1991-02-01 Data storage system and method including data and checksum write staging storage
JP3318794A JPH087702B2 (en) 1991-02-01 1991-12-03 Data storage system and method
EP91403305A EP0497067A1 (en) 1991-02-01 1991-12-06 High performance data storage system and method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US07/648,944 US5239640A (en) 1991-02-01 1991-02-01 Data storage system and method including data and checksum write staging storage

Publications (1)

Publication Number Publication Date
US5239640A true US5239640A (en) 1993-08-24

Family

ID=24602855

Family Applications (1)

Application Number Title Priority Date Filing Date
US07/648,944 Expired - Fee Related US5239640A (en) 1991-02-01 1991-02-01 Data storage system and method including data and checksum write staging storage

Country Status (3)

Country Link
US (1) US5239640A (en)
EP (1) EP0497067A1 (en)
JP (1) JPH087702B2 (en)

Cited By (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5442752A (en) * 1992-01-24 1995-08-15 International Business Machines Corporation Data storage method for DASD arrays using striping based on file length
US5469566A (en) * 1992-03-12 1995-11-21 Emc Corporation Flexible parity generation circuit for intermittently generating a parity for a plurality of data channels in a redundant array of storage units
US5572659A (en) * 1992-05-12 1996-11-05 International Business Machines Corporation Adapter for constructing a redundant disk storage system
US5572729A (en) * 1994-01-11 1996-11-05 Sun Microsystems, Inc. Method for stateless rename propagation between hierarchical file name spaces
US5673415A (en) * 1993-12-03 1997-09-30 Unisys Corporation High speed two-port interface unit where read commands suspend partially executed write commands
US5680570A (en) * 1991-06-12 1997-10-21 Quantum Corporation Memory system with dynamically allocatable non-volatile storage capability
US5682499A (en) * 1995-06-06 1997-10-28 International Business Machines Corporation Directory rebuild method and apparatus for maintaining and rebuilding directory information for compressed data on direct access storage device (DASD)
US5701304A (en) * 1996-09-16 1997-12-23 Cirrus Logic, Inc. Method for correcting unrecoverable sectors using track level redundancy in a disc drive storage system
US5726922A (en) * 1994-01-03 1998-03-10 International Business Machines Corp. Assembly for removably connecting data storage devices
US5751733A (en) * 1996-09-16 1998-05-12 Cirrus Logic, Inc. Interleaved redundancy sector for correcting an unrecoverable sector in a disc storage device
US5794056A (en) * 1993-10-04 1998-08-11 International Business Machines Corporation System for automatic buffering of commands for DASD units
WO1998052288A1 (en) * 1997-05-16 1998-11-19 Motorola Inc. Method and apparatus for estimating capacity in a communication system
US5844919A (en) * 1996-09-16 1998-12-01 Cirrus Logic, Inc. Sector and track level error correction system for disc storage systems
US5867640A (en) * 1993-06-01 1999-02-02 Mti Technology Corp. Apparatus and method for improving write-throughput in a redundant array of mass storage devices
US5872800A (en) * 1996-09-16 1999-02-16 Cirrus Logic, Inc. Write verify method for correcting unrecoverable sectors in a disc storage system using track level redundancy
US5875455A (en) * 1994-06-10 1999-02-23 Matsushita Electric Industrial Co., Ltd. Information recording and reproducing apparatus merging sequential recording requests into a single recording request, and method of data caching for such apparatus
US5931920A (en) * 1997-08-05 1999-08-03 Adaptec, Inc. Command interpreter system in an I/O controller
US5996107A (en) * 1996-11-14 1999-11-30 Sanyo Electric Co., Ltd. Error correction decoder including an address generation circuit
US6105075A (en) * 1997-08-05 2000-08-15 Adaptec, Inc. Scatter gather memory system for a hardware accelerated command interpreter engine
US6269374B1 (en) 1998-05-26 2001-07-31 International Business Machines Corporation Method and apparatus for updating checksums of data structures
US6397348B1 (en) 1994-10-12 2002-05-28 International Business Machines Corporation Redundant array of disk drives with asymmetric mirroring and asymmetric mirroring data processing method
US20030023933A1 (en) * 2001-07-27 2003-01-30 Sun Microsystems, Inc. End-to-end disk data checksumming
US20030191707A1 (en) * 1992-05-29 2003-10-09 Shepherd Ian Kenneth Methods and apparatus relating to the formulation and trading of risk management contracts
US20040205317A1 (en) * 2003-04-08 2004-10-14 International Business Machines Corporation Method, apparatus and program storage device for providing data integrity using check data and other metadata on a formatted storage medium
DE102004006645A1 (en) * 2004-02-11 2005-09-22 Giesecke & Devrient Gmbh Method for the reliable calculation of a checksum
US6952797B1 (en) 2000-10-25 2005-10-04 Andy Kahn Block-appended checksums
US6970890B1 (en) * 2000-12-20 2005-11-29 Bitmicro Networks, Inc. Method and apparatus for data recovery
US6988219B2 (en) 1993-06-04 2006-01-17 Network Appliance, Inc. Providing parity in a RAID sub-system using non-volatile memory
US20060075281A1 (en) * 2004-09-27 2006-04-06 Kimmel Jeffrey S Use of application-level context information to detect corrupted data in a storage system
US20070162838A1 (en) * 2005-09-22 2007-07-12 Atsushi Esumi Data writing apparatus and a storage system
US20080162594A1 (en) * 2000-06-06 2008-07-03 Quantum Corporation Data storage system and process
US9465821B1 (en) 2012-08-08 2016-10-11 Amazon Technologies, Inc. Data storage integrity validation
US9563681B1 (en) 2012-08-08 2017-02-07 Amazon Technologies, Inc. Archival data flow management
US9652487B1 (en) * 2012-08-08 2017-05-16 Amazon Technologies, Inc. Programmable checksum calculations on data storage devices
US9767129B2 (en) 2012-08-08 2017-09-19 Amazon Technologies, Inc. Data storage inventory indexing
US9767098B2 (en) 2012-08-08 2017-09-19 Amazon Technologies, Inc. Archival data storage system
US9779035B1 (en) 2012-08-08 2017-10-03 Amazon Technologies, Inc. Log-based data storage on sequentially written media
US9830111B1 (en) 2012-08-08 2017-11-28 Amazon Technologies, Inc. Data storage space management
US9904788B2 (en) 2012-08-08 2018-02-27 Amazon Technologies, Inc. Redundant key management
US10120579B1 (en) 2012-08-08 2018-11-06 Amazon Technologies, Inc. Data storage management for sequentially written media
US10558581B1 (en) 2013-02-19 2020-02-11 Amazon Technologies, Inc. Systems and techniques for data recovery in a keymapless data storage system
US10698880B2 (en) 2012-08-08 2020-06-30 Amazon Technologies, Inc. Data storage application programming interface
CN112000524A (en) * 2020-08-27 2020-11-27 北京浪潮数据技术有限公司 Backup and reading method, device, equipment and medium of system data in hard disk
US11386060B1 (en) 2015-09-23 2022-07-12 Amazon Technologies, Inc. Techniques for verifiably processing data in distributed computing systems
US11397647B2 (en) * 2018-08-30 2022-07-26 Apollo Intelligent Driving Technology (Beijing) Co., Ltd. Hot backup system, hot backup method, and computer device

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2743606B2 (en) * 1991-04-11 1998-04-22 三菱電機株式会社 Array type recording device
US5708668A (en) * 1992-05-06 1998-01-13 International Business Machines Corporation Method and apparatus for operating an array of storage devices
US5315602A (en) * 1992-08-12 1994-05-24 Digital Equipment Corporation Optimized stripe detection for redundant arrays of disk drives
JP3183719B2 (en) * 1992-08-26 2001-07-09 三菱電機株式会社 Array type recording device
EP0612015A1 (en) * 1993-02-16 1994-08-24 International Business Machines Corporation Improved disk array system having special parity groups for data blocks with high update activity
US6138126A (en) 1995-05-31 2000-10-24 Network Appliance, Inc. Method for allocating files in a file system integrated with a raid disk sub-system
JP3862274B2 (en) * 1993-06-03 2006-12-27 ネットワーク・アプライアンス・インコーポレイテッド File allocation method of file system integrated with RAID disk subsystem
US6604118B2 (en) 1998-07-31 2003-08-05 Network Appliance, Inc. File system image transfer
US5530948A (en) * 1993-12-30 1996-06-25 International Business Machines Corporation System and method for command queuing on raid levels 4 and 5 parity drives
US6457130B2 (en) 1998-03-03 2002-09-24 Network Appliance, Inc. File access control in a multi-protocol file server
US6317844B1 (en) 1998-03-10 2001-11-13 Network Appliance, Inc. File server storage arrangement
US6343984B1 (en) 1998-11-30 2002-02-05 Network Appliance, Inc. Laminar flow duct cooling system
US6728922B1 (en) 2000-08-18 2004-04-27 Network Appliance, Inc. Dynamic data space
US6636879B1 (en) 2000-08-18 2003-10-21 Network Appliance, Inc. Space allocation in a write anywhere file system
US7072916B1 (en) 2000-08-18 2006-07-04 Network Appliance, Inc. Instant snapshot
US6668264B1 (en) 2001-04-03 2003-12-23 Network Appliance, Inc. Resynchronization of a target volume with a source volume
TWI320139B (en) 2005-08-01 2010-02-01 Method for improving writing data efficiency and storage subsystem and system implementing the same
FR2916065B1 (en) * 2007-05-07 2011-08-19 Siemens Vdo Automotive METHOD OF CONTROLLING DATA INTEGRITY AND DEVICE USING THE METHOD.

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4092732A (en) * 1977-05-31 1978-05-30 International Business Machines Corporation System for recovering data stored in failed memory unit
US4761785A (en) * 1986-06-12 1988-08-02 International Business Machines Corporation Parity spreading to enhance storage access
US4849978A (en) * 1987-07-02 1989-07-18 International Business Machines Corporation Memory unit backup using checksum
US4870643A (en) * 1987-11-06 1989-09-26 Micropolis Corporation Parallel drive array storage system
US4989206A (en) * 1988-06-28 1991-01-29 Storage Technology Corporation Disk drive memory
US4993030A (en) * 1988-04-22 1991-02-12 Amdahl Corporation File system for a plurality of storage classes
US5124987A (en) * 1990-04-16 1992-06-23 Storage Technology Corporation Logical track write scheduling system for a parallel disk drive array data storage subsystem

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2561428B1 (en) * 1984-03-16 1986-09-12 Bull Sa DISC MEMORY RECORDING METHOD AND DISC MEMORY SYSTEM
CA1296103C (en) * 1987-06-02 1992-02-18 Theodore Jay Goodlander High-speed, high capacity, fault-tolerant, error-correcting storage system
AU630635B2 (en) * 1988-11-14 1992-11-05 Emc Corporation Arrayed disk drive system and method
JPH02297777A (en) * 1989-05-12 1990-12-10 Mitsubishi Electric Corp Optical recording system
JPH02291011A (en) * 1989-04-28 1990-11-30 Fujitsu Ltd Memory device

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4092732A (en) * 1977-05-31 1978-05-30 International Business Machines Corporation System for recovering data stored in failed memory unit
US4761785A (en) * 1986-06-12 1988-08-02 International Business Machines Corporation Parity spreading to enhance storage access
US4761785B1 (en) * 1986-06-12 1996-03-12 Ibm Parity spreading to enhance storage access
US4849978A (en) * 1987-07-02 1989-07-18 International Business Machines Corporation Memory unit backup using checksum
US4870643A (en) * 1987-11-06 1989-09-26 Micropolis Corporation Parallel drive array storage system
US4993030A (en) * 1988-04-22 1991-02-12 Amdahl Corporation File system for a plurality of storage classes
US4989206A (en) * 1988-06-28 1991-01-29 Storage Technology Corporation Disk drive memory
US5124987A (en) * 1990-04-16 1992-06-23 Storage Technology Corporation Logical track write scheduling system for a parallel disk drive array data storage subsystem

Cited By (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5680570A (en) * 1991-06-12 1997-10-21 Quantum Corporation Memory system with dynamically allocatable non-volatile storage capability
US5442752A (en) * 1992-01-24 1995-08-15 International Business Machines Corporation Data storage method for DASD arrays using striping based on file length
US5675726A (en) * 1992-03-12 1997-10-07 Emc Corporation Flexible parity generation circuit
US5831393A (en) * 1992-03-12 1998-11-03 Emc Corporation Flexible parity generation circuit
US5469566A (en) * 1992-03-12 1995-11-21 Emc Corporation Flexible parity generation circuit for intermittently generating a parity for a plurality of data channels in a redundant array of storage units
US5572659A (en) * 1992-05-12 1996-11-05 International Business Machines Corporation Adapter for constructing a redundant disk storage system
US8401954B2 (en) 1992-05-29 2013-03-19 Alice Corporation Pty Ltd. Systems relating to the formulation of risk management contracts
US20030191707A1 (en) * 1992-05-29 2003-10-09 Shepherd Ian Kenneth Methods and apparatus relating to the formulation and trading of risk management contracts
US20080120222A1 (en) * 1992-05-29 2008-05-22 Alice Corporation Pty Ltd. Methods and apparatus relating to the formulation and trading of risk management contracts
US20110093377A1 (en) * 1992-05-29 2011-04-21 Alice Corporation Pty Ltd. Systems and Computer Program Products for Exchanging an Obligation
US7725375B2 (en) 1992-05-29 2010-05-25 Alice Corporation Pty Ltd Systems and computer program products for exchanging an obligation
US7822676B2 (en) 1992-05-29 2010-10-26 Alice Corporation Pty Ltd Systems and methods relating to the formulation of a multi-party contract
US20050246197A1 (en) * 1992-05-29 2005-11-03 Alice Corporation Pty Ltd. Methods and apparatus relating to the formulation and trading of risk management contracts
US7149720B2 (en) * 1992-05-29 2006-12-12 Alice Corporation Pty Ltd Systems for exchanging an obligation
US5867640A (en) * 1993-06-01 1999-02-02 Mti Technology Corp. Apparatus and method for improving write-throughput in a redundant array of mass storage devices
US6988219B2 (en) 1993-06-04 2006-01-17 Network Appliance, Inc. Providing parity in a RAID sub-system using non-volatile memory
US5794056A (en) * 1993-10-04 1998-08-11 International Business Machines Corporation System for automatic buffering of commands for DASD units
US5673415A (en) * 1993-12-03 1997-09-30 Unisys Corporation High speed two-port interface unit where read commands suspend partially executed write commands
US5726922A (en) * 1994-01-03 1998-03-10 International Business Machines Corp. Assembly for removably connecting data storage devices
US5572729A (en) * 1994-01-11 1996-11-05 Sun Microsystems, Inc. Method for stateless rename propagation between hierarchical file name spaces
US5875455A (en) * 1994-06-10 1999-02-23 Matsushita Electric Industrial Co., Ltd. Information recording and reproducing apparatus merging sequential recording requests into a single recording request, and method of data caching for such apparatus
US5983319A (en) * 1994-06-10 1999-11-09 Matsushita Electric Industrial Co., Ltd. Information recording and reproduction apparatus and a method of data caching including read-ahead capability
US6397348B1 (en) 1994-10-12 2002-05-28 International Business Machines Corporation Redundant array of disk drives with asymmetric mirroring and asymmetric mirroring data processing method
US5682499A (en) * 1995-06-06 1997-10-28 International Business Machines Corporation Directory rebuild method and apparatus for maintaining and rebuilding directory information for compressed data on direct access storage device (DASD)
US5701304A (en) * 1996-09-16 1997-12-23 Cirrus Logic, Inc. Method for correcting unrecoverable sectors using track level redundancy in a disc drive storage system
US5872800A (en) * 1996-09-16 1999-02-16 Cirrus Logic, Inc. Write verify method for correcting unrecoverable sectors in a disc storage system using track level redundancy
US5844919A (en) * 1996-09-16 1998-12-01 Cirrus Logic, Inc. Sector and track level error correction system for disc storage systems
US5751733A (en) * 1996-09-16 1998-05-12 Cirrus Logic, Inc. Interleaved redundancy sector for correcting an unrecoverable sector in a disc storage device
US5996107A (en) * 1996-11-14 1999-11-30 Sanyo Electric Co., Ltd. Error correction decoder including an address generation circuit
WO1998052288A1 (en) * 1997-05-16 1998-11-19 Motorola Inc. Method and apparatus for estimating capacity in a communication system
US6105075A (en) * 1997-08-05 2000-08-15 Adaptec, Inc. Scatter gather memory system for a hardware accelerated command interpreter engine
US5931920A (en) * 1997-08-05 1999-08-03 Adaptec, Inc. Command interpreter system in an I/O controller
US6269374B1 (en) 1998-05-26 2001-07-31 International Business Machines Corporation Method and apparatus for updating checksums of data structures
US20080162594A1 (en) * 2000-06-06 2008-07-03 Quantum Corporation Data storage system and process
US6952797B1 (en) 2000-10-25 2005-10-04 Andy Kahn Block-appended checksums
US6970890B1 (en) * 2000-12-20 2005-11-29 Bitmicro Networks, Inc. Method and apparatus for data recovery
US20030023933A1 (en) * 2001-07-27 2003-01-30 Sun Microsystems, Inc. End-to-end disk data checksumming
US20040205317A1 (en) * 2003-04-08 2004-10-14 International Business Machines Corporation Method, apparatus and program storage device for providing data integrity using check data and other metadata on a formatted storage medium
DE102004006645A1 (en) * 2004-02-11 2005-09-22 Giesecke & Devrient Gmbh Method for the reliable calculation of a checksum
US20060075281A1 (en) * 2004-09-27 2006-04-06 Kimmel Jeffrey S Use of application-level context information to detect corrupted data in a storage system
US20070162838A1 (en) * 2005-09-22 2007-07-12 Atsushi Esumi Data writing apparatus and a storage system
US7783950B2 (en) * 2005-09-22 2010-08-24 Rohm Co., Ltd. Data writing apparatus and a storage system
US9767129B2 (en) 2012-08-08 2017-09-19 Amazon Technologies, Inc. Data storage inventory indexing
US9904788B2 (en) 2012-08-08 2018-02-27 Amazon Technologies, Inc. Redundant key management
US9652487B1 (en) * 2012-08-08 2017-05-16 Amazon Technologies, Inc. Programmable checksum calculations on data storage devices
US9465821B1 (en) 2012-08-08 2016-10-11 Amazon Technologies, Inc. Data storage integrity validation
US9767098B2 (en) 2012-08-08 2017-09-19 Amazon Technologies, Inc. Archival data storage system
US9779035B1 (en) 2012-08-08 2017-10-03 Amazon Technologies, Inc. Log-based data storage on sequentially written media
US9830111B1 (en) 2012-08-08 2017-11-28 Amazon Technologies, Inc. Data storage space management
US9563681B1 (en) 2012-08-08 2017-02-07 Amazon Technologies, Inc. Archival data flow management
US10120579B1 (en) 2012-08-08 2018-11-06 Amazon Technologies, Inc. Data storage management for sequentially written media
US10157199B2 (en) 2012-08-08 2018-12-18 Amazon Technologies, Inc. Data storage integrity validation
US10936729B2 (en) 2012-08-08 2021-03-02 Amazon Technologies, Inc. Redundant key management
US10698880B2 (en) 2012-08-08 2020-06-30 Amazon Technologies, Inc. Data storage application programming interface
US10558581B1 (en) 2013-02-19 2020-02-11 Amazon Technologies, Inc. Systems and techniques for data recovery in a keymapless data storage system
US11386060B1 (en) 2015-09-23 2022-07-12 Amazon Technologies, Inc. Techniques for verifiably processing data in distributed computing systems
US11397647B2 (en) * 2018-08-30 2022-07-26 Apollo Intelligent Driving Technology (Beijing) Co., Ltd. Hot backup system, hot backup method, and computer device
CN112000524A (en) * 2020-08-27 2020-11-27 北京浪潮数据技术有限公司 Backup and reading method, device, equipment and medium of system data in hard disk

Also Published As

Publication number Publication date
JPH04278641A (en) 1992-10-05
JPH087702B2 (en) 1996-01-29
EP0497067A1 (en) 1992-08-05

Similar Documents

Publication Publication Date Title
US5239640A (en) Data storage system and method including data and checksum write staging storage
US4755928A (en) Outboard back-up and recovery system with transfer of randomly accessible data sets between cache and host and cache and tape simultaneously
EP1160654B1 (en) Method for on-line, real time, data migration
US6009481A (en) Mass storage system using internal system-level mirroring
US5487160A (en) Concurrent image backup for disk storage system
US6148368A (en) Method for accelerating disk array write operations using segmented cache memory and data logging
US5959860A (en) Method and apparatus for operating an array of storage devices
US7293196B2 (en) Method, apparatus, and system for preserving cache data of redundant storage controllers
US5826002A (en) Parity storage unit, in a disk array system for generating updated parity data from received data records
US5457791A (en) Storage system and method of control
US7228381B2 (en) Storage system using fast storage device for storing redundant data
US6397348B1 (en) Redundant array of disk drives with asymmetric mirroring and asymmetric mirroring data processing method
US5420983A (en) Method for merging memory blocks, fetching associated disk chunk, merging memory blocks with the disk chunk, and writing the merged data
US5715424A (en) Apparatus and method for writing data onto rewritable optical media
EP0701208B1 (en) Disk array subsystem and data generation method therefor
US5694570A (en) Method and system of buffering data written to direct access storage devices in data processing systems
JPH10105344A (en) Data recorder system and data write method therefor
JPH07225715A (en) Method and system for accessing data in cache
EP0156179A2 (en) Method for protecting volatile primary store in a staged storage system
US6615284B2 (en) Method and apparatus for performing input/output operations between a requesting device and responding device
JPH0863394A (en) Storage device system and controlling mathod for storage device
JP3250859B2 (en) Disk array device, computer system and data storage device
GB2293912A (en) Disk storage device for disk array
KR100423812B1 (en) RAID Controller and Parity Operator Method having Disk Cash Memory Controller and Parity Operator Device
JP3202550B2 (en) Disk array subsystem

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, A CO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST.;ASSIGNORS:FROEMKE, JAMES W.;STYCZINSKI, DAVID A.;REEL/FRAME:005592/0374

Effective date: 19910201

FPAY Fee payment

Year of fee payment: 4

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
FP Lapsed due to failure to pay maintenance fee

Effective date: 20010824

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362