WO2010135136A2 - Block-level single instancing - Google Patents

Block-level single instancing Download PDF

Info

Publication number
WO2010135136A2
WO2010135136A2 PCT/US2010/034676 US2010034676W WO2010135136A2 WO 2010135136 A2 WO2010135136 A2 WO 2010135136A2 US 2010034676 W US2010034676 W US 2010034676W WO 2010135136 A2 WO2010135136 A2 WO 2010135136A2
Authority
WO
WIPO (PCT)
Prior art keywords
data
blocks
block
file
storage
Prior art date
Application number
PCT/US2010/034676
Other languages
French (fr)
Other versions
WO2010135136A3 (en
Inventor
Deepak Raghunath Attarde
Rajiv Kottomtharayil
Manoj Kumar Vijayan
Original Assignee
Commvault Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Commvault Systems, Inc. filed Critical Commvault Systems, Inc.
Publication of WO2010135136A2 publication Critical patent/WO2010135136A2/en
Publication of WO2010135136A3 publication Critical patent/WO2010135136A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • G06F11/1448Management of the data involved in backup or backup restore
    • G06F11/1453Management of the data involved in backup or backup restore using de-duplication of the data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1415Saving, restoring, recovering or retrying at system level
    • G06F11/1435Saving, restoring, recovering or retrying at system level using file system or storage system metadata
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • G06F11/1458Management of the backup or restore process
    • G06F11/1464Management of the backup or restore process for networked environments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/17Details of further file system functions
    • G06F16/174Redundancy elimination performed by the file system
    • G06F16/1748De-duplication implemented within the file system, e.g. based on file segments
    • G06F16/1752De-duplication implemented within the file system, e.g. based on file segments based on file chunks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0614Improving the reliability of storage systems
    • G06F3/0617Improving the reliability of storage systems in relation to availability
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0638Organizing or formatting or addressing of data
    • G06F3/064Management of blocks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/067Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1446Point-in-time backing up or restoration of persistent data
    • G06F11/1458Management of the backup or restore process
    • G06F11/1469Backup restoration techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/80Database-specific techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/84Using snapshots, i.e. a logical point-in-time copy of the data

Definitions

  • Single instancing in a data storage system typically involves attempting to store only a single instance of a file on a storage device.
  • a separate folder on the file system of the storage device is created for each single instancing storage operation performed.
  • Each file that has been single instanced is stored as a separate individual file in the separate folder.
  • Certain file systems of storage devices may not be capable of adequately providing for storing such large numbers of files.
  • Other file systems may be equipped to handle storing millions of files or more, but may not perform optimally in such situations.
  • Figure 1 is a block diagram illustrating an example of a data storage enterprise that may employ aspects of the invention.
  • Figure 2 is a block diagram depicting in more detail certain components illustrated in Figure 1.
  • Figure 3 is a flow diagram of certain aspects of a process for performing a storage operation.
  • Figure 4 is a flow diagram of other aspects of a process for performing a storage operation.
  • Figures 5A and 5B are diagrams illustrating suitable data structures that may be employed by aspects of the invention.
  • Figures 6A and 6B are diagrams illustrating suitable data structures that may be employed by aspects of the invention.
  • Figure 7 is a diagram illustrating various data structures that aspects of the invention may utilize.
  • Figure 8 is a flow diagram of a process for restoring data.
  • Figure 9 is a flow diagram of a process for pruning data.
  • This application describes in detail, among other things, systems and methods for single instancing (alternatively called deduplicating) blocks of data in a data storage system (alternatively called a data storage network, a data storage environment, or a data storage enterprise).
  • the data storage system stores single instanced blocks of data (alternatively referred to as deduplicated blocks of data) in one or more files and maintains one or more data structures (e.g., index files) that keep track of which blocks of data are referenced.
  • the data storage system may include multiple computing devices or computing systems (e.g., client computing devices) that store primary data (e.g., production data such as system files, user files, etc.).
  • the data storage system may also include a secondary storage computing device, a single instance database, and one or more storage devices that store copies of the primary data (e.g., secondary copies, tertiary copies, etc.).
  • the secondary storage computing device receives blocks of data from the computing devices and accesses the single instance database to determine whether the blocks of data are unique (unique meaning that no instances of the blocks of data are already stored on the storage devices). If a block of data is unique, the single instance database stores it in a file on a storage device. If not, the secondary storage computing device can avoid storing the block of data on the storage devices.
  • the primary data of the computing devices can be divided into data that is eligible for single instancing and data that is not eligible for single instancing.
  • An example of the latter is metadata (e.g., Master File Table (MFT) information) and an example of the former is data (e.g., operating system and/or application files).
  • MFT Master File Table
  • a file typically comprises one or more blocks as tracked by the file systems of the computing devices.
  • the computing devices align data that is eligible for single instancing into blocks of data (which may comprise one or more blocks as tracked by the file systems of the computing devices) and generate identifiers for the blocks of data that the secondary storage computing device uses to determine if the blocks of data are unique.
  • This allows the secondary storage computing device to avoid generating identifiers for the blocks of data, which may be computationally expensive and/or require a long time to perform. Therefore, the distribution of the task of generating identifiers (which can be computationally expensive operations) across numerous computing devices frees up the secondary storage computing device to perform other operations (e.g., storing data, retrieving data, pruning data, etc.).
  • the computing devices send the blocks of data and other data (e.g., metadata and/or the data that is not eligible for single instancing) in a data stream to the secondary storage computing device.
  • the secondary storage computing device receives the data stream and stores blocks of data and their identifiers in buffers in random access memory (RAM).
  • RAM random access memory
  • the secondary storage computing device determines whether a block of data is already stored on a storage device. To do this, the secondary storage computing device determines, by analyzing data structures in the single instance database in view of the block's identifier, whether the block of data is already stored on a storage device. If it is, then the secondary storage computing device 1) stores a link to the already stored block of data in a metadata file and 2) discards the block of data from the memory buffer. If it is not, then the secondary storage computing device stores the block of data in a container file.
  • the secondary storage computing device can keep a single block of data in a single memory buffer while it looks up its identifier in the single instance database. This allows the secondary storage computing device to avoid writing the block of data to disk (an operation which is typically slower than storing the block of data in a RAM buffer) until the secondary storage computing device determines that it needs to store the block of data in a container file on a storage device.
  • the secondary storage computing device stores data that is not eligible for single instancing in metadata files.
  • the secondary storage computing device By storing multiple blocks of data in a single container file, the secondary storage computing device avoids storing each block of data as a separate file on the file systems of the storage devices. This reduces the number of files that would be stored on the file systems of the storage devices, thereby ensuring that the storage devices can adequately store the data of the computing devices in the data storage system.
  • One advantage of these techniques is that they significantly reduce the number of files stored on a file system of a computing device or storage device. This is at least partly due to the storage of data blocks within the container files. Even if the secondary storage computing device performs numerous storage operations, these techniques will result in storing far fewer files on the file system than storage operations where each data block is stored as a separate file. Therefore, the file system of the computing device or storage device may not necessarily have to contend with storing excessively large numbers of files, such as millions of files or more. Accordingly, these techniques enable very large numbers of blocks of data to be stored without regard to limitations of the file system of the computing device or storage device.
  • the secondary storage computing device creates the container files as sparse files (typically only on operating systems that support sparse files, e.g., Windows operating systems, and other operating systems that support sparse files).
  • a sparse file is type of file that may include empty space (e.g., a sparse file may have real data within it, such as at the beginning of the file and/or at the end of the file, but may also have empty space in it that is not storing actual data, such as a contiguous range of bytes all having a value of zero).
  • the secondary storage computing device maintains a separate index that stores an indication of whether blocks of data in container files are referred to by links in metadata files. In some examples, this can be analogized to using another, non-native file system that keeps track of blocks of data in the container files, on top of the existing, native file systems of the storage devices.
  • the secondary storage computing device can prune it.
  • the secondary storage computing device accesses the separate index to determine the blocks of data that are not referred to by links.
  • the secondary storage computing device can free up space in the container files corresponding to those blocks of data by marking the portions of the physical media corresponding to the unreferenced portions of the container file as available for storage (e.g., by zeroing out the corresponding bytes in the container files).
  • the secondary storage computing device can free up space in the container files by truncating the extreme portions of the container files (e.g., the extremities such as the beginnings and/or the ends of the container files), thereby making the corresponding portions of the physical media available to store other data. Freeing up space in container files allows the operating system to utilize the freed-up space in other fashions (e.g., other programs may utilize the freed-up space).
  • FIG. 1 and 2 and the discussion herein provide a brief, general description of a suitable specialized environment in which aspects of the invention can be implemented.
  • Those skilled in the relevant art will appreciate that aspects of the invention can be practiced with other communications, data processing, or computer system configurations, including: Internet appliances, hand-held devices (including personal digital assistants (PDAs)), wearable computers, all manner of cellular phones, mobile phones, and/or mobile devices, multi-processor systems, microprocessor-based or programmable consumer electronics, set-top boxes, network PCs, mini-computers, mainframe computers, and the like.
  • PDAs personal digital assistants
  • the terms "computer,” “server,” “host,” “host system,” “client,” and the like are generally used interchangeably herein, and refer to any of the above devices and systems, as well as any data processor.
  • LAN Local Area Network
  • WAN Wide Area Network
  • program modules may be located in both local and remote memory storage devices.
  • aspects of the invention may be stored or distributed on computer-readable media, including tangible computer-readable storage media such as magnetically or optically readable computer discs, hard-wired or preprogrammed chips (e.g., EEPROM semiconductor chips), nanotechnology memory, biological memory, or other data storage media.
  • computer implemented instructions, data structures, screen displays, and other data under aspects of the invention may be distributed over the Internet or over other networks (including wireless networks), on a propagated signal on a propagation medium (e.g., an electromagnetic wave(s), a sound wave, etc.) over a period of time, or they may be provided on any analog or digital network (packet switched, circuit switched, or other scheme).
  • Figure 1 illustrates an example of a data storage system that may employ aspects of the invention.
  • Figure 2 illustrates in more detail certain components illustrated in Figure 1 that may be used to implement a block-level single instancing system. These components include a secondary storage computing device, a single instancing database, and a storage device that stores only a single instance of blocks of data of one or more computing devices (e.g., client computing devices).
  • Figure 3 illustrates aspects of a process for copying data that a computing device (e.g., a client computing device) may perform. These aspects include determining whether data is eligible for single instancing and transferring data in a data stream to the secondary storage computing device.
  • Figure 4 illustrates aspects of the copy process that the secondary storage computing device may perform upon receipt of the data stream from the computing device. During this part of the copy process, the secondary storage computing device determines whether the data of the computing device is single instanceable.
  • Figures 5A and 5B, 6A and 6B, and 7 are illustrations of various data structures that aspects of the invention may utilize.
  • Figures 5A and 5B depict data streams that the computing device may form during the copy process.
  • Figures 6A and 6B show data structures that may be used by the single instance database to keep track of where blocks of data and references to blocks of data are stored on the storage device.
  • Figure 7 illustrates data structures that may be used to store blocks of data on the storage device.
  • Figures 8 and 9 are process flow diagrams.
  • Figure 8 illustrates an example process that the secondary storage computing device may perform to restore data stored on the storage device, such as to a computing device.
  • Figure 9 depicts an example process that the secondary storage computing device may perform to prune data stored on the storage device when it is no longer required to be stored on the storage device.
  • Figure 1 illustrates an example of one arrangement of resources in a computing network, comprising a data storage system 150.
  • the resources in the data storage system 150 may employ the processes and techniques described herein.
  • the system 150 includes a storage manager 105, one or more data agents 195, one or more secondary storage computing devices 165, one or more storage devices 115, one or more computing devices 130 (called clients 130), one or more data or information stores 160 and 162, and a single instancing database 123.
  • the storage manager 105 includes an index 111 , a jobs agent 120, an interface agent 125, and a management agent 131.
  • the system 150 may represent a modular storage system such as the CommVault QiNetix system, and also the CommVault GALAXY backup system, available from CommVault Systems, Inc. of Oceanport, NJ, aspects of which are further described in the commonly-assigned U.S. Patent Application No. 09/610,738, now U.S. Patent No. 7,035,880, the entirety of which is incorporated by reference herein.
  • the system 150 may also represent a modular storage system such as the CommVault Simpana system, also available from CommVault Systems, Inc.
  • the system 150 may generally include combinations of hardware and software components associated with performing storage operations on electronic data. Storage operations include copying, backing up, creating, storing, retrieving, and/or migrating primary storage data (e.g., data stores 160 and/or 162) and secondary storage data (which may include, for example, snapshot copies, backup copies, hierarchical storage management (HSM) copies, archive copies, and other types of copies of electronic data stored on storage devices 115).
  • the system 150 may provide one or more integrated management consoles for users or system processes to interface with in order to perform certain storage operations on electronic data as further described herein. Such integrated management consoles may be displayed at a central control facility or several similar consoles distributed throughout multiple network locations to provide global or geographically specific network data storage information.
  • storage operations may be performed according to various storage preferences, for example, as expressed by a user preference, a storage policy, a schedule policy, and/or a retention policy.
  • a "storage policy" is generally a data structure or other information source that includes a set of preferences and other storage criteria associated with performing a storage operation.
  • the preferences and storage criteria may include, but are not limited to, a storage location, relationships between system components, network pathways to utilize in a storage operation, data characteristics, compression or encryption requirements, preferred system components to utilize in a storage operation, a single instancing or variable instancing policy to apply to the data, and/or other criteria relating to a storage operation.
  • a storage policy may indicate that certain data is to be stored in the storage device 115, retained for a specified period of time before being aged to another tier of secondary storage, copied to the storage device 115 using a specified number of data streams, etc.
  • a "schedule policy” may specify a frequency with which to perform storage operations and a window of time within which to perform them. For example, a schedule policy may specify that a storage operation is to be performed every Saturday morning from 2:00 a.m. to 4:00 a.m.
  • a "retention policy” may specify how long data is to be retained at specific tiers of storage or what criteria must be met before data may be pruned or moved from one tier of storage to another tier of storage.
  • the storage policy includes information generally specified by the schedule policy and/or the retention policy.
  • Storage policies, schedule policies and/or retention policies may be stored in a database of the storage manager 105, to archive media as metadata for use in restore operations or other storage operations, or to other locations or components of the system 150.
  • the system 150 may comprise a storage operation cell that is one of multiple storage operation cells arranged in a hierarchy or other organization.
  • Storage operation cells may be related to backup cells and provide some or all of the functionality of backup cells as described in the assignee's U.S. Patent Application No. 09/354,058, now U.S. Patent No. 7,395,282, which is incorporated herein by reference in its entirety.
  • storage operation cells may also perform additional types of storage operations and other types of storage management functions that are not generally offered by backup cells.
  • Storage operation cells may contain not only physical devices, but also may represent logical concepts, organizations, and hierarchies.
  • a first storage operation cell may be configured to perform a first type of storage operations such as HSM operations, which may include backup or other types of data migration, and may include a variety of physical components including a storage manager 105 (or management agent 131), a secondary storage computing device 165, a client 130, and other components as described herein.
  • a second storage operation cell may contain the same or similar physical components; however, it may be configured to perform a second type of storage operations, such as storage resource management (SRM) operations, and may include monitoring a primary data copy or performing other known SRM operations.
  • SRM storage resource management
  • each storage operation cell may contain the same or similar physical devices.
  • different storage operation cells may contain some of the same physical devices and not others.
  • a storage operation cell configured to perform SRM tasks may contain a secondary storage computing device 165, client 130, or other network device connected to a primary storage volume
  • a storage operation cell configured to perform HSM tasks may instead include a secondary storage computing device 165, client 130, or other network device connected to a secondary storage volume and not contain the elements or components associated with and including the primary storage volume.
  • connection does not necessarily require a physical connection; rather, it could refer to two devices that are operably coupled to each other, communicably coupled to each other, in communication with each other, or more generally, refer to the capability of two devices to communicate with each other.
  • These two storage operation cells may each include a different storage manager 105 that coordinates storage operations via the same secondary storage computing devices 165 and storage devices 115.
  • This "overlapping" configuration allows storage resources to be accessed by more than one storage manager 105, such that multiple paths exist to each storage device 115 facilitating failover, load balancing, and promoting robust data access via alternative routes.
  • the same storage manager 105 may control two or more storage operation cells (whether or not each storage operation cell has its own dedicated storage manager 105).
  • the extent or type of overlap may be user-defined (through a control console) or may be automatically configured to optimize data storage and/or retrieval.
  • the clients 130 typically include application software for performing various operations.
  • Clients 130 typically also include an operating system on which the application software runs.
  • a file system can be provided to facilitate and control file access by the operating system and application software.
  • File systems can facilitate access to local and remote storage devices for file or data access and storage.
  • Clients 130 can also include local storage such as a media module media drive with fixed or removable media.
  • the clients 130 include storage mechanisms for allowing computer programs or other instructions or data to be loaded into memory for execution.
  • Such storage mechanisms might include, for example, a fixed or removable storage unit and an interface.
  • Examples of such storage units and interfaces can include a program cartridge and cartridge interface, a removable memory (for example, a flash memory or other removable memory module) and memory slot, a PCMCIA slot and card, and other fixed or removable storage units and interfaces that allow software and data to be transferred from the storage unit to memory.
  • Data agent 195 may be a software module or part of a software module that is generally responsible for performing storage operations on the data of the client 130 stored in data store 160/162 or other memory location. Each client 130 may have at least one data agent 195 and the system 150 can support multiple clients 130. Data agent 195 may be distributed between client 130 and storage manager 105 (and any other intermediate components), or it may be deployed from a remote location or its functions approximated by a remote process that performs some or all of the functions of data agent 195.
  • module might describe a given unit of functionality that can be performed in accordance with one or more embodiments of the present invention.
  • a module might be implemented utilizing any form of hardware, software, firmware, or a combination thereof.
  • processors, controllers, ASICs, PLAs, logical components, software routines or other mechanisms might be implemented to make up a module.
  • the various modules described herein might be implemented as discrete modules or the functions and features described can be shared in part or in total among one or more modules.
  • the overall system 150 may employ multiple data agents 195, each of which may perform storage operations on data associated with a different application.
  • different individual data agents 195 may be designed to handle Microsoft Exchange data, Lotus Notes data, Microsoft Windows file system data, Microsoft Active Directory Objects data, Microsoft SQL Server data, Microsoft Sharepoint Server data, and other types of data known in the art.
  • Other embodiments may employ one or more generic data agents 195 that can handle and process multiple data types rather than using the specialized data agents described above.
  • one data agent 195 may be required for each data type to perform storage operations on the data of the client 130. For example, to back up, migrate, and restore all the data on a Microsoft Exchange server, the client 130 may use one Microsoft Exchange Mailbox data agent 195 to back up the Exchange mailboxes, one Microsoft Exchange Database data agent 195 to back up the Exchange databases, one Microsoft Exchange Public Folder data agent 195 to back up the Exchange Public Folders, and one Microsoft Windows File System data agent 195 to back up the file system of the client 130. These data agents 195 would be treated as four separate data agents 195 by the system even though they reside on the same client 130.
  • the overall system 150 may use one or more generic data agents 195, each of which may be capable of handling two or more data types.
  • one generic data agent 195 may be used to back up, migrate and restore Microsoft Exchange Mailbox data and Microsoft Exchange Database data while another generic data agent 195 may handle Microsoft Exchange Public Folder data and Microsoft Windows File System data, etc.
  • Data agents 195 may be responsible for arranging or packing data to be copied or migrated into a certain format such as an archive file. Nonetheless, it will be understood that this represents only one example, and any suitable packing or containerization technique or transfer methodology may be used if desired.
  • Such an archive file may include metadata, a list of files or data objects copied, the file, and data objects themselves.
  • any data moved by the data agents may be tracked within the system by updating indexes associated with appropriate storage managers 105 or secondary storage computing devices 165.
  • a file or a data object refers to any collection or grouping of bytes of data that can be viewed as one or more logical units.
  • storage manager 105 may be a software module or other application that coordinates and controls storage operations performed by the system 150.
  • Storage manager 105 may communicate with some or all elements of the system 150, including clients 130, data agents 195, secondary storage computing devices 165, and storage devices 115, to initiate and manage storage operations (e.g., backups, migrations, data recovery operations, etc.).
  • Storage manager 105 may include a jobs agent 120 that monitors the status of some or all storage operations previously performed, currently being performed, or scheduled to be performed by the system 150. (One or more storage operations are alternatively referred to herein as a "job” or “jobs.") Jobs agent 120 may be communicatively coupled to an interface agent 125 (e.g., a software module or application).
  • Interface agent 125 may include information processing and display software, such as a graphical user interface (“GUI”), an application programming interface (“API”), or other interactive interface through which users and system processes can retrieve information about the status of storage operations.
  • GUI graphical user interface
  • API application programming interface
  • users may optionally issue instructions to various storage operation cells regarding performance of the storage operations as described and contemplated herein. For example, a user may modify a schedule concerning the number of pending snapshot copies or other types of copies scheduled as needed to suit particular needs or requirements. As another example, a user may employ the GUI to view the status of pending storage operations in some or all of the storage operation cells in a given network or to monitor the status of certain components in a particular storage operation cell (e.g., the amount of storage capacity left in a particular storage device 115).
  • Storage manager 105 may also include a management agent 131 that is typically implemented as a software module or application program.
  • management agent 131 provides an interface that allows various management agents 131 in other storage operation cells to communicate with one another. For example, assume a certain network configuration includes multiple storage operation cells hierarchically arranged or otherwise logically related in a WAN or LAN configuration. With this arrangement, each storage operation cell may be connected to the other through each respective interface agent 125. This allows each storage operation cell to send and receive certain pertinent information from other storage operation cells, including status information, routing information, information regarding capacity and utilization, etc. These communications paths may also be used to convey information and instructions regarding storage operations.
  • a management agent 131 in a first storage operation cell may communicate with a management agent 131 in a second storage operation cell regarding the status of storage operations in the second storage operation cell.
  • Another illustrative example includes the case where a management agent 131 in a first storage operation cell communicates with a management agent 131 in a second storage operation cell to control storage manager 105 (and other components) of the second storage operation cell via management agent 131 contained in storage manager 105.
  • management agent 131 in a first storage operation cell communicates directly with and controls the components in a second storage operation cell and bypasses the storage manager 105 in the second storage operation cell.
  • storage operation cells can also be organized hierarchically such that hierarchically superior cells control or pass information to hierarchically subordinate cells or vice versa.
  • Storage manager 105 may also maintain an index, a database, or other data structure 111.
  • the data stored in database 111 may be used to indicate logical associations between components of the system, user preferences, management tasks, media containerization and data storage information or other useful data.
  • the storage manager 105 may use data from database 111 to track logical associations between secondary storage computing device 165 and storage devices 115 (or movement of data as containerized from primary to secondary storage).
  • the secondary storage computing device 165 which may also be referred to as a media agent, may be implemented as a software module that conveys data, as directed by storage manager 105, between a client 130 and one or more storage devices 115 such as a tape library, a magnetic media storage device, an optical media storage device, or any other suitable storage device.
  • secondary storage computing device 165 may be communicatively coupled to and control a storage device 115.
  • a secondary storage computing device 165 may be considered to be associated with a particular storage device 115 if that secondary storage computing device 165 is capable of routing and storing data to that particular storage device 115.
  • a secondary storage computing device 165 associated with a particular storage device 115 may instruct the storage device to use a robotic arm or other retrieval means to load or eject a certain storage media, and to subsequently archive, migrate, or restore data to or from that media.
  • Secondary storage computing device 165 may communicate with a storage device 115 via a suitable communications path such as a SCSI or Fibre Channel communications link.
  • the storage device 115 may be communicatively coupled to the storage manager 105 via a SAN.
  • Each secondary storage computing device 165 may maintain an index, a database, or other data structure 161 that may store index data generated during storage operations for secondary storage (SS) as described herein, including creating a metabase (MB). For example, performing storage operations on Microsoft Exchange data may generate index data. Such index data provides a secondary storage computing device 165 or other external device with a fast and efficient mechanism for locating data stored or backed up.
  • SS secondary storage
  • MB metabase
  • index data provides a secondary storage computing device 165 or other external device with a fast and efficient mechanism for locating data stored or backed up.
  • a secondary storage computing device index 161 may store data associating a client 130 with a particular secondary storage computing device 165 or storage device 115, for example, as specified in a storage policy, while a database or other data structure in secondary storage computing device 165 may indicate where specifically the data of the client 130 is stored in storage device 115, what specific files were stored, and other information associated with storage of the data of the client 130.
  • index data may be stored along with the data backed up in a storage device 115, with an additional copy of the index data written to index cache in a secondary storage device.
  • the data is readily available for use in storage operations and other activities without having to be first retrieved from the storage device 115.
  • information stored in cache is typically recent information that reflects certain particulars about operations that have recently occurred. After a certain period of time, this information is sent to secondary storage and tracked. This information may need to be retrieved and uploaded back into a cache or other memory in a secondary computing device before data can be retrieved from storage device 115.
  • the cached information may include information regarding format or containerization of archives or other files stored on storage device 115.
  • One or more of the secondary storage computing devices 165 may also maintain one or more single instance databases 123. More details as to single instancing may be found in one or more of the following commonly-assigned U.S. patent applications: 1) U.S. Pat. App. No. 11/269,512 (entitled SYSTEM AND METHOD TO SUPPORT SINGLE INSTANCE STORAGE OPERATIONS, Attorney Docket No. 60692-8023.US00); 2) U.S. Pat. App. No. 12/145,347 (entitled APPLICATION-AWARE AND REMOTE SINGLE INSTANCE DATA MANAGEMENT, Attorney Docket No. 60692-8056. USOO); or 3) U.S. Pat. App. No.
  • the secondary storage computing devices 165 maintain one or more variable instance databases.
  • Variable instancing generally refers to storing in secondary storage one or more instances, but fewer than the total number of instances, of each data block (or data object) in a set of data (e.g., primary data). More details as to variable instancing may be found in the commonly-assigned U.S. Pat. App. No. 61/164,803 (entitled STORING A VARIABLE NUMBER OF INSTANCES OF DATA OBJECTS, Attorney Docket No. 60692-8068.US00).
  • a client 130 such as a data agent 195, or a storage manager 105, coordinates and directs local archiving, migration, and retrieval application functions as further described in the previously- referenced U.S. Patent Application No. 09/610,738.
  • This client 130 can function independently or together with other similar clients 130.
  • each secondary storage computing device 165 has its own associated metabase 161.
  • Each client 130 may also have its own associated metabase 170.
  • each "tier" of storage such as primary storage, secondary storage, tertiary storage, etc., may have multiple metabases or a centralized metabase, as described herein.
  • the metabases on this storage tier may be centralized.
  • second and other tiers of storage may have either centralized or distributed metabases.
  • mixed architecture systems may be used if desired, that may include a first tier centralized metabase system coupled to a second tier storage system having distributed metabases and vice versa, etc.
  • a storage manager 105 or other management module may keep track of certain information that allows the storage manager 105 to select, designate, or otherwise identify metabases to be searched in response to certain queries as further described herein. Movement of data between primary and secondary storage may also involve movement of associated metadata and other tracking information as further described herein.
  • primary data may be organized into one or more sub- clients.
  • a sub-client is a portion of the data of one or more clients 130, and can contain either all of the data of the clients 130 or a designated subset thereof.
  • the data store 162 includes two sub-clients.
  • an administrator or other user with the appropriate permissions; the term administrator is used herein for brevity
  • Figure 2 is a block diagram depicting in more detail certain components illustrated in Figure 1.
  • the data agent 195 of the client 130 includes various components, such as a data identification component 202, a block identification component 204, and an identifier generation component 206.
  • the data agent 195 also includes a compression component 210, an encryption component 212, and a data stream generation component 214. Various functions performed by these components are described herein.
  • the client 130 includes data 240.
  • the data 240 includes single instanceable data (Sl data) 242 and non-single instanceable data (non-SI data) 244.
  • SI data 242 includes data that is eligible for single instancing.
  • Non-SI data 244 includes data that is not eligible for single instancing.
  • Non-SI data 244 may include metadata such as access control lists (ACLs), disk partition information, Master File Table (MFT) or File Allocation Table (FAT) information, and/or other metadata.
  • Non-SI data 244 may also include other data that is determined not to be single instanceable.
  • SI data 242 may include data 240 of the client 130 other than non- SI data 244 (e.g., system files, application files, user files, etc.).
  • the secondary storage computing device 165 includes a data stream reception component 220 and an identifier comparison component 222. Various functions performed by these components are also described in detail herein.
  • the secondary storage computing device 165 also includes a memory 230, which includes multiple buffers 232.
  • the secondary storage computing device 165 may also include other components, such as a decompression component and/or a decryption component.
  • the single instance database 123 includes data structures 250 that are used to store data, such as metadata about SI data 242.
  • the storage device 115 also includes data structures 260 that are used to store data, such as SI data 242 and non- SI data 244.
  • the secondary storage computing device 165 includes the components that the client 130 includes, and performs the functions that the client 130 performs.
  • Figures 3 and 4 are flow diagrams illustrating certain aspects of processes 300 and 400, respectively, for performing a storage operation such as a copy operation.
  • a storage operation (alternatively referred to as a job) is typically performed on files stored on file systems of one or more clients 130.
  • One or more of the entities illustrated in the figures e.g., Figures 1 and/or 2) may perform different aspects of the processes 300 and 400.
  • a storage manager 105 instigates the process 300 by sending an indication specifying the storage operation to the data agent 195.
  • the data agent 195 accesses the data of the client 130 (e.g., accesses files stored on the filesystem of the client 130).
  • the data agent 195 sends the data to the secondary storage computing device 165, which then stores the data on one or more storage devices 115. In some examples, less than all of these entities may be involved in performing the storage operation.
  • the process 300 is described as being performed by the data agent 195 and the process 400 is described as being performed by the secondary storage computing device 165. However, those of skill in the art will understand that aspects of the processes 300 and 400 may be performed by any one or more of the entities described herein (e.g., the data agent 195, the storage manager 105, the secondary storage computing device 165, etc.).
  • the process 300 begins at step 305 where the data agent 195 receives an indication to copy data of the client 130.
  • the storage manager 105 may send the indication to the data agent 195 (e.g., according to a storage policy), an administrator may manually start the process 300, and/or the process 300 may be automatically started according to a schedule policy.
  • the data agent 195 accesses the data 240 of the client 130.
  • the data agent 195 e.g., the data identification component 202 determines which portions of the data 240 are SI data 242 and which portions are non-SI data 244. For example, the data agent 195 may determine that metadata (e.g., MFT, FAT, volume information, transaction logs, etc.) on the file system of the client 130 is non-SI data 244, and that data other than metadata is SI data 242 (e.g., system files, user files, etc.).
  • metadata e.g., MFT, FAT, volume information, transaction logs, etc.
  • the data agent 195 (e.g., the data stream generation component 214) forms a data stream of multiple pairs of stream header and stream payload from the SI data 242 and the non-SI data 244.
  • a data stream therefore, comprises multiple pairs of stream header and stream payload.
  • data streams may contain data organized in other fashions.
  • the data agent 195 may set a flag in the stream header to indicate that the corresponding stream payload contains single instanceable data.
  • the data agent 195 aligns the stream header and stream payload into one or more fixed size blocks of data.
  • a block of data is a sequence of bits or bytes having a nominal length (a data block size).
  • the file system of the client 130 may track its data 240 in blocks (alternatively called clusters) in sizes of 512 bytes, 4Kb, 16Kb, or other sizes.
  • a block may be a subset of one or more data objects.
  • a file on the file system of the client 130 typically spans one or more blocks (e.g., a file of size 10Kb may span 3 blocks of size 4Kb).
  • the data agent 195 typically aligns data blocks such that they have the same size, which may be 32Kb, 64Kb, 128Kb, 256Kb, 512Kb, or other sizes. Accordingly, the term data block, as used herein, may comprise one or more blocks as tracked by the file system of the clients 130.
  • these 128Kb data blocks comprise 32 blocks of data 240 as tracked by the file system of the client 130.
  • the data agent 195 determines whether a data block is single instanceable. The data agent 195 does so by analyzing the portion of the one or more corresponding stream headers that indicates whether the data block is single instanceable. For example, the stream headers may contain a flag or bit that indicates whether the successive stream payload contain single instanceable data. (For example, see Figure 5A, illustrating stream headers containing such flags.) If the data block is single instanceable, the process 300 continues at step 330, where the data agent 195 (e.g., the identifier generation component 206) generates an identifier for the data block.
  • the data agent 195 e.g., the identifier generation component 206
  • identifiers include a hash value, message digest, checksum, digital fingerprint, digital signature or other sequence of bytes that substantially uniquely identifies the data block in the data storage system.
  • identifiers could be generated using Message Digest Algorithm 5 (MD5) or Secure Hash Algorithm SHA 512.
  • MD5 Message Digest Algorithm 5
  • SHA 512 Secure Hash Algorithm 512.
  • the phrase "substantially unique" is used to modify the term "identifier" because algorithms used to produce hash values may result in collisions, where two different data objects, when hashed, result in the same hash value.
  • collisions should be suitably rare and thus the identifier generated for a block should be unique throughout the data storage system.
  • probabilistically unique identifier may also be used.
  • the phrase “probabilistically unique” is used to indicate that collisions should be low-probability occurrences, and, therefore, the identifier should be unique throughout the data storage system.
  • the data agent 195 inserts the generated identifier into the data stream.
  • the generated identifier may be comprised in an identifier header and identifier data pair that immediately follows the data block for which it is generated. (See Figure 5B and the accompanying description for additional details of the identifier header and identifier data pair.)
  • the data agent 195 determines whether there are more data blocks. If so, the process 300 returns to step 325. If not, the process 300 continues at step 345, where the data agent 195 transfers the data stream to the secondary storage computing device 165. The process 300 then ends.
  • the data agent 195 may perform additional operations upon the stream header and/or stream payload, such as encrypting the stream payload (e.g., using the encryption component 212) and/or compressing the stream payload (e.g., using the compression component 210).
  • FIG 4 is a flow diagram illustrating certain aspects of the process 400 that the secondary storage computing device 165 performs upon receiving the data stream from the data agent 195.
  • the secondary storage computing device 165 receives the data stream from the data agent 195.
  • the secondary storage computing device 165 stores the stream header and stream payload corresponding to a data block in a buffer 232 of the memory 230.
  • the secondary storage computing device 165 can store the entire stream header and stream payload pairs corresponding to a single block in a single buffer, because the buffer size (e.g., approximately 640Kb) is greater than the size of the stream header and stream payload pairs (e.g., up to approximately 512Kb).
  • the buffer size is typically no greater than 10 times the size of the stream header and stream payload pairs.
  • the memory 230 includes 30 buffers 232, thus allowing the secondary storage computing device 165 to simultaneously store up to 30 different data blocks in fast-access memory.
  • the ability to store multiple data blocks in memory enables the secondary storage computing device 165 to avoid writing the multiple data blocks to disk, which can be a lengthy operation.
  • the secondary storage computing device 165 determines whether the data block is single instanceable. The secondary storage computing device 165 may do so, for example, by analyzing the metadata in the stream header that indicates whether the data block is single instanceable (e.g., a flag or bit that indicates whether the data block is single instanceable).
  • the process 400 continues at step 425, where the secondary storage computing device (e.g., the identifier comparison component 222) obtains the identifier corresponding to the data block (e.g., from the identifier data of the data stream) and looks up the identifier.
  • the secondary storage computing device 165 looks up the identifier in the primary table in the single instance database 123. (Example data structures used by the single instance database 123 are illustrated in Figures 6A and 6B and described with reference to these figures).
  • the secondary storage computing device 165 finds the identifier of the data block in the primary table, this indicates that an instance of the data block is already stored on the storage device 115, and that the block of data should not be stored. Accordingly, the secondary storage computing device 165 can avoid storing another instance of the data block and can instead store a link (alternatively called a pointer) to the location(s) of the already stored instance.
  • the secondary storage computing device 165 adds a link to the location(s) of the already stored instance of the data block to a metadata file. The link refers or points to the already stored instance of the data block.
  • the secondary storage computing device 165 may add as the link to the metadata file the record of the already stored instance of the data block in the primary table.
  • the secondary storage computing device 165 adds an entry to the secondary table in the single instance database. The entry includes the location of the link in the metadata file.
  • the secondary storage computing device 165 also increments a reference count corresponding to the data block in the primary table. The reference count indicates the number of links to the already stored instance of the data block.
  • the secondary storage computing device 165 discards the stream header and stream payload corresponding to the data block from the buffer 232 of the memory 230. Additionally or alternatively, the secondary storage computing device 165 may indicate that the buffer is available for storing another pair of stream header and stream payload.
  • the secondary storage computing device 165 does not find the identifier of the block in the primary table (step 430), this indicates that no instances of the data block are already stored on the storage device 115, and that the block of data should be stored. Accordingly, at step 435 the secondary storage computing device 165 stores the data block in a container file on the storage device 115. (See Figure 7 and the accompanying description for additional details of container files.) At step 440 the secondary storage computing device 165 adds an entry to the primary table in the single instance database. The entry includes the location of the data block in the container file.
  • step 415 the process 400 continues at step 420, where the secondary storage computing device 165 stores the block in a metadata file. (See Figure 7 and the accompanying description for additional details of metadata files.)
  • step 420 the secondary storage computing device 165 determines whether there are more data blocks. If so, the process 400 returns to step 415. If not the process 400 concludes.
  • the secondary storage computing device 165 may perform additional operations during the process 400, such as decrypting the stream payload (e.g., using a decryption component) and/or decompressing the stream payload (e.g., using a decompression component).
  • the secondary storage computing device 165 may also store in the index 161 , for the data blocks, information mapping an archive file and offset to the physical location of the data blocks.
  • An archive file is a logical entity that is created during a storage operation and that corresponds to physical locations of data blocks on the storage device 115.
  • the storage manager 105 may map archive files to physical locations and keep such information in index 111.
  • a variable number of instances of data blocks (e.g., more than one instance and up to N-1 instances, where N is the number of instances of the data block in primary data) is stored on the storage devices 115.
  • the secondary storage computing devices 165 may use techniques described in the previously referenced U.S. Pat. App. No. 61/164,803 to ensure that a sufficient number of instances of the blocks of data are stored on the storage devices 115. Storing multiple instances (up to N-1) of N data blocks provides for less risk of data loss than single instance storage techniques, and generally nearly as less risk of data loss as conventional data protection techniques (which store N instances of N data blocks).
  • Storing multiple instances (up to N-1) of N data blocks also provides for more efficient use of available storage space than conventional data protection techniques, and almost as efficient use as single instance storage techniques. Accordingly, the storing of a variable number of instances of data blocks enables an administrator to tailor data protection to strike an appropriate balance between 1) minimizing the risk of data loss, and 2) making efficient use of available data storage space, in accordance with the administrator's requirements.
  • FIGs 5A and 5B are diagrams of example data streams 500 and 550, respectively, that may be employed by aspects of the invention.
  • the data agent 195 forms the data stream 500 from the data 240 of the client 130.
  • the data stream 500 is composed of multiple pairs of stream header 502 and stream payload 504.
  • a stream payload 504 includes SI data 242 and/or non-SI data 244.
  • a stream header 502 includes metadata about the stream payload 504.
  • This metadata may include, for example, a length of the stream payload 504, an indication of whether the stream payload 504 is encrypted, an indication of whether the stream payload 504 is compressed, an archive file identifier (ID), an indication of whether the stream payload 504 is single instanceable, and an indication of whether the stream payload 504 is a start of a block of data.
  • ID archive file identifier
  • the data stream 550 has the stream header 502 and stream payload 504 aligned into multiple data blocks.
  • the data blocks are of size 64Kb.
  • the first two stream header 502 and stream payload 504 pairs comprise a first data block of size 64Kb.
  • the first stream header 502 indicates that the length of the succeeding stream payload 504 is 63Kb and that it is the start of a data block.
  • the stream header 502 may also include the metadata discussed with reference to the stream headers 502 illustrated in Figure 3A.
  • the next stream header 502 indicates that the succeeding stream payload 504 has a length of 1 Kb and that it is not the start of a new data block.
  • the identifier header 506 includes an indication that the succeeding identifier data 508 includes the identifier for the immediately previous data block.
  • the identifier data 508 includes the identifier that the data agent (e.g., the identifier generation component 206) generated for the data block.
  • the data stream 550 also includes other stream header 502 and stream payload 504 pairs, which may be for SI data 242 and/or for non-SI data 244.
  • Figures 6A and 6B are diagrams illustrating the data structures 250 that may be used by the single instance database 123.
  • the data structures 250 do not form part of a native file system of a storage device storing the single instance database 123.
  • the data structures 250 are not provided by any native file system for storage devices at least as of the time of the filing of the provisional patent application to which this application claims priority.
  • the data structures 250 include a primary table 600 and a secondary table 650.
  • the primary table 600 includes an identifier column 602 in which a data block identifier is stored, a location column 604 in which a location of the data block in a container file is stored, an offset column 606 indicating the offset within the container file corresponding to the location of the data block, and a reference count column 608, which contains a reference count of the number of links that refer to the data block.
  • row 620 includes information about a data block for which the identifier is "0xA1 B3FG.” This data block is located in the container file that is indicated in the location column 606, at an offset of 10 within the container file.
  • this data block is referred to twice, meaning that there are two links that refer to the data block.
  • row 624 includes information about a data block for which the identifier is "0xC13804." The location of this data block is indicated in the location column 604 at an offset of 38 within the container file, and it is referred to one other time, by one link.
  • the secondary table 650 includes information about links that refer to data blocks.
  • the secondary table 650 includes an identifier column 652, a referring location column 654, and an offset column 656.
  • row 660 includes information about a reference to the data block having the identifier of "0xA1 B3FG" (row 620 in the primary table 600).
  • the location of the link is indicated in column 654, at an offset of five within the indicated metadata file.
  • row 662 includes information about another reference to the data block having the identifier of "0xA1 B3FG.” This link is located at the location indicated in column 654, at an offset of 15 within the indicated metadata file.
  • row 664 includes information about a reference to the block for which the identifier is "0xC13804" (row 624 in the primary table 600). The location of the link is indicated in column 654, at an offset of 19 within the indicated metadata file.
  • Figure 7 is a diagram illustrating the data structures 260 that may be used to store blocks of SI data and non-SI data on the storage device 115.
  • the data structures 260 do not form part of a native file system of the storage device 115.
  • the data structures 260 are not provided by any native file systems for storage devices at least as of the time of the filing of the provisional patent application to which this application claims priority.
  • the data structures 260 include one or more volume folders 702, one or more chunk folders 704/705 within a volume folder 702, and multiple files within a chunk folder 704.
  • Each chunk folder 704/705 includes a metadata file 706/707, a metadata index file 708/709, one or more container files 710/711/713, and a container index file 712/714.
  • the metadata file 706/707 stores non-SI data blocks as well as links to SI data blocks stored in container files.
  • the metadata index file 708/709 stores an index to the data in the metadata file 706/707.
  • the container files 710/711/713 store SI data blocks.
  • the container index file 712/714 stores an index to the container files 710/711/713.
  • the container index file 712/714 stores an indication of whether a corresponding block in a container file 710/711/713 is referred to by a link in a metadata file 706/707.
  • data block B2 in the container file 710 is referred to by a link in the metadata file 707 in the chunk folder 705.
  • the corresponding index entry in the container index file 712 indicates that the data block B2 in the container file 710 is referred to.
  • data block B1 in the container file 711 is referred to by a link in the metadata file 707, and so the corresponding index entry in the container index file 712 indicates that this data block is referred to.
  • the data structures 260 illustrated in Figure 7 may have been created as a result of two storage operations involving two clients 130.
  • a first storage operation on a first client 130 could result in the creation of the first chunk folder 704, and a second storage operation on a second client 130 could result in the creation of the second chunk folder 705.
  • the container files 710/711 in the first chunk folder 704 would contain the blocks of SI data 242 of the first client 130.
  • the second storage operation on the data 240 of the second client 130 would result in the secondary storage computing device 165 storing primarily links to the data blocks of the first client 130 that are already stored in the container files 710/711. Accordingly, while a first storage operation may result in storing nearly all of the data subject to the storage operation, subsequent storage operations involving similar data may result in substantial data storage space savings, because links to already stored data blocks can be stored instead of additional instances of data blocks.
  • a sparse file is type of file that may include empty space (e.g., a sparse file may have real data within it, such as at the beginning of the file and/or at the end of the file, but may also have empty space in it that is not storing actual data, such as a contiguous range of bytes all having a value of zero).
  • the secondary storage computing device 165 creates a new container file 710/711/713 when a container file 710/711/713 either includes 100 blocks of data or when the size of the container file 710 exceeds 50 Mb.
  • the secondary storage computing device 165 creates a new container file 710/711/713 when a container file 710/711/713 satisfies other criteria (e.g., it contains from approximately 100 to approximately 1000 blocks or when its size exceeds approximately 50 Mb to 1Gb). Those of skill in the art will understand that the secondary storage computing device 165 can create a new container file 710/711/713 when other criteria are met.
  • a file on which a storage operation is performed may comprise a large number of data blocks.
  • a 100Mb file may be comprised in 400 data blocks of size 256Kb. If such a file is to be stored, its data blocks may span more than one container file, or even more than one chunk folder.
  • a database file of 20Gb may comprise over 40,000 data blocks of size 512Kb. If such a database file is to be stored, its data blocks will likely span multiple container files, multiple chunk folders, and potentially multiple volume folders. As described in detail herein, restoring such files may thus requiring accessing multiple container files, chunk folders, and/or volume folders to obtain the requisite data blocks.
  • One advantage of the data structures 260 illustrated in Figure 7 and/or of the techniques described herein is that they significantly reduce the number of files stored on a file system of the storage device 115. This is at least partly due to the storage of data blocks within the container files 710/711/713. Even if numerous storage operations using these data structures 260 are performed, this will result in far fewer files on the storage device 115 than storage operations where each data block is stored as a separate file. Therefore, the file system of the storage device 115 may not necessarily have to contend with storing excessively large numbers of files, such as millions of files or more. Accordingly, the systems and methods described herein enable very large numbers of blocks of data to be stored without regard to limitations of the file system of the storage device 115.
  • Another advantage is that the data storage system enables a reduction in the amount of blocks of data stored on the storage devices 115, while still maintaining at least one instance of each block of primary data.
  • blocks of primary data can be distributed across two or more storage devices 115, thereby adding a further aspect of redundancy.
  • Another advantage is that the metadata files 706/707, the metadata index files 708/709, the container files 710/711/713, and/or the container index files 712/714 could be used to replicate the data stored in the single instance database 123 or reconstruct the single instance database 123 if the data of the single instance database 123 is ever lost and/or corrupted.
  • the storage of data blocks in the container files may create additional complexities when it comes time to prune data blocks (pruning data blocks may be alternatively referred to as deleting or removing data blocks) that the data storage system no longer need retain. This is because the data blocks are not stored as files on the file system on the storage device 115 and thus cannot be directly referenced by the file system using the file system's data structures (the data structures that are built into or provided with the file system). As described in detail with reference to Figure 9, the secondary storage computing device 165 uses the container index files 712/714 to keep track of which blocks of data are referenced and thus which blocks are not prunable (deletable).
  • the use of the container index files 712/714, the metadata index files 708/709, and/or the primary and secondary tables 600/650 to track data is analogous to a driver, agent or an additional file system that is layered on top of the existing file system of the storage device 115.
  • This driver/agent/additional file system allows the data storage system to efficiently keep track of very large numbers of blocks of data, without regard to any limitations of the file systems of the storage devices 115. Accordingly, the data storage system can store very large numbers of blocks of data.
  • the data structures 260 illustrated in Figure 7 and the techniques described herein enable the performance of multiple storage operations cumulatively involving very large amounts of data, while still allowing for recovery of space on the storage device 115 when storage of certain data blocks is no longer required.
  • the data of numerous clients 130 can be protected without having to store redundant copies or instances of data blocks.
  • Space on the storage device 115 can also be recovered when it is no longer necessary to store certain data blocks. Accordingly, storage operations involving very large amounts of data are enabled and optimized by the techniques described herein. Process for Restoring Data
  • FIG. 8 is a flow diagram of a process 800 for restoring one or more blocks of data.
  • the process 800 is described as being performed by the secondary storage computing device 165, although those of skill in the art will understand that aspects of the process 800 may be performed by any of the entities described herein.
  • the process 800 begins at step 805 where the secondary storage computing device 165 receives a selection of data to restore (e.g., one or more files).
  • a selection of data to restore e.g., one or more files.
  • an administrator may utilize an integrated management console that provides an interface for allowing the administrator to specify one or more data blocks to be restored (e.g., by allowing the administrator to specify one or more files to be restored).
  • a client 130 may request that a data block that had been previously copied from the client 130 be restored to the client 130.
  • the secondary storage computing device 165 determines an archive file and offset within the archive file corresponding to the data to be restored.
  • the secondary storage computing device 165 may analyze the index 111 of the storage manager
  • the secondary storage computing device 165 determines volume folders and chunk folders corresponding to the archive file and offset. The secondary storage computing device 165 may do so by analyzing the index 161 to determine the volume folders and chunk folders. The determined volume folders and chunk folders contain the requested data.
  • the secondary storage computing device 165 accesses an index file within the determined volume folders and chunk folders that corresponds to the data to be restored. This may be the metadata index file 708/709 when the requested data is non-SI data 244 or the container index file 712/714 when the requested data is SI data 242.
  • the secondary storage computing device 165 determines, from the index file, the offset within the metadata file 706/707 or the offset within the container file 710/711/13 corresponding to the requested data.
  • the secondary storage computing device 165 accesses the metadata file 706/707 or the container file 710/711/13 and seeks to the determined offset.
  • the secondary storage computing device 165 retrieves the data from the metadata file 706/707 or the container file 710/711/13.
  • the secondary storage computing device restores the data to a selected location (e.g., to a client 130 and/or to another location). The process 800 then concludes.
  • restoring a file may necessitate accessing multiple container files, chunk folders, and/or volume folders to obtain the data blocks that comprise the file.
  • the secondary storage computing device 165 may thus have to obtain a first data block from a first container file and a second data block from a second container file.
  • the secondary storage computing device 165 may thus have to obtain a first data block from a first container file within a first folder and a second data block from a second container file within a second folder. To do so, the secondary storage computing device 165 may have to access multiple index files or other data structures to locate the requisite blocks of data.
  • Those of skill in the art will understand that various techniques may be used to restore data such as files and other data.
  • FIG. 9 is a flow diagram of a process 900 for pruning data.
  • the process 900 is described as being performed by the secondary storage computing device 165, although those of skill in the art will understand that aspects of the process 900 may be performed by any of the entities described herein.
  • the process 900 begins when the secondary storage computing device 165 receives instructions to prune data corresponding to a storage operation (job). Additionally or alternatively, one or more files can be selected to be pruned, and/or one or more data blocks can be selected to be pruned. This selection of a job or other data to be deleted can be made manually, such as by an administrator, or automatically, such as by the job, files, and/or data blocks aging out by a retention policy.
  • the data structures 260 illustrated in Figure 7 may have been created as a result of two jobs involving two clients 130.
  • a first job on a first client 130 could result in the creation of the first chunk folder 704, and a second job on a second client 130 could result in the creation of the second chunk folder 705.
  • the process 900 is described using this example. More specifically, the process 900 is described below as pruning the data created as a result of the first job. Of course, a similar process may be used to delete other jobs, or even smaller increments of data or data objects, such as individual files or blocks.
  • the secondary storage computing device 165 determines the file, e.g., archive file, and the volume folders 702 and chunk folder 704 corresponding to the job to be pruned. The secondary storage computing device 165 may do so, for example, by analyzing the index 111 and/or the index 161 to determine this information.
  • the secondary storage computing device 165 deletes the metadata file 706 and the metadata index file 708 in the chunk folder 704. The secondary storage computing device 165 can delete the metadata file 706 and the metadata index file 708 in this example because these files include non-SI data 244, which is not referenced by any other data.
  • the secondary storage computing device 165 accesses the container file 710 and the container index file 712 in the chunk folder 704.
  • the secondary storage computing device 165 begins iterating through the data blocks in the container files 710.
  • the secondary storage computing device 165 accesses the primary table 600 in the single instance database 123.
  • the secondary storage computing device 165 determines from the primary table 600 whether the reference count of a data block in the container file 710 is equal to zero. If so, this indicates that there are no references to the data block.
  • the process 900 then continues at step 925, where the secondary storage computing device 165 sets the entry in the container index file 712 corresponding to the data block equal to zero, thus indicating that there are no references to the data block, and therefore prunable.
  • the process 900 continues at step 930.
  • the secondary storage computing device 165 determines whether there are more data blocks in the container file 710. If so, the process 900 returns to step 920, where it accesses the next data block. If there are no more data blocks in the container file 710, the process 900 continues at step 932, where the secondary storage computing device 165 determines whether all the entries in the container index file 712 corresponding to the container file 710 are equal to zero. As illustrated in Figure 7, the second index entry in the container index file 712 is not equal to zero, thus indicating that the corresponding block in container file 710 is referenced (by data in the chunk folder 705, as earlier described). Accordingly, the container file 710 cannot be deleted.
  • the secondary storage computing device 165 would delete the container file 710.
  • the process would then continue at step 935, where the secondary storage computing device 165 determines whether there are more container files. According to the example as illustrated in Figure 7, there is an additional container file 711.
  • the process 900 then returns to step 915, where it performs the same steps 920- 933 for container file 711.
  • the secondary storage computing device 165 would also determine that the container file 711 cannot be deleted, because it contains a data block that is referenced (by data in the chunk folder 705, as earlier described).
  • the process 900 continues at step 940, where the secondary storage computing device 165 determines whether to free up storage space in the container files 710/711.
  • the secondary storage computing device 165 may do so using various techniques. For example, if the operating system of the secondary storage computing device 165 supports sparse files, then the secondary storage computing device 165 may free up space by zeroing out the bytes in the container files corresponding to the space to be freed up.
  • the secondary storage computing device 165 may mark these portions of the container files 710/711 as available for storage by the operating system or the file system.
  • the secondary storage computing device 165 may do so by calling an API of the operating system to mark the unreferenced portions of the container files 710/711 as available for storage.
  • the secondary storage computing device 165 may use certain optimizations to manage the number of times portions of the container file are specified or marked as available for storage, such as only zeroing out bytes in container files when a threshold number of unreferenced contiguous blocks is reached (e.g., three or more unreferenced contiguous blocks). These optimizations may result in less overhead for the operating system because it reduces the number of contiguous ranges of zero-value bytes in the container files 710/711 that the operating system must keep track of (e.g., it reduces the amount of metadata about portions of the container files 710/711 that are available for storage).
  • the secondary storage computing device 165 may free up space by truncating either the beginning or the end of the container files 710/711 (removing or deleting data at the beginning or end of the container files 710/711).
  • the secondary storage computing device 165 may do so by calling an API of the operating system, or by operating directly on the container files 710/711. For example, if a certain number of the last blocks of the container file are not being referred to, the secondary storage computing device 165 may truncate these portions of the container files 710/711. Other techniques may be used to free up space in the container files 710/711 for storage of other data.
  • the secondary storage computing device 165 frees up space in the container files 710/711. The process 900 then concludes.
  • the chunk folder 704 would contain only the container files 710/711 and the container index file 712.
  • the chunk folder 705 is pruned (that is, when the job that created this chunk folder is selected to be pruned)
  • the container files 710/711 in the chunk folder 704 can be deleted, because they no longer contain data blocks that is referenced by other data. Therefore, pruning data corresponding to a job may also result in pruning data corresponding to an earlier job, because the data corresponding to the earlier job is no longer referenced by the later job.
  • the process 900 is described with reference to the pruning of data corresponding to jobs (one or more storage operations), other data can also be pruned.
  • an administrator may wish to delete SI data 242 but retain non-SI data 244.
  • the administrator may instruct the secondary storage computing device 165 to delete the container files 710/711/713 but retain the metadata files 706/707 and metadata index files 708/709.
  • an administrator or storage policy may delete one or more specific files.
  • the secondary storage computing device 165 deletes the data blocks in the container files 710/711/713 corresponding to the specific files but retains other data blocks.
  • the process 900 may include fewer or more steps than those described herein to accommodate these other pruning examples. Those of skill in the art will understand that data can be pruned in various fashions and therefore, that the process 900 is not limited to the steps described herein.
  • One advantage of the process 900 and the techniques described herein is that they enable the deletion of data on the storage devices 115 that no longer needs to be stored while still retaining data that needs to be stored, and doing so in a space- efficient manner. Space previously allocated for data blocks that no longer need to be stored can be reclaimed by the data storage system, and used to store other data. Accordingly, the techniques described herein provide for efficient use of available storage space (available on physical media).
  • the secondary storage computing device 165 may keep track of which blocks of data in container files 710 are not referenced, instead of keeping track of which blocks of data are referred to by links.
  • non-SI data 244 may not be aligned into blocks of data. Accordingly, the invention is not limited except as by the appended claims.
  • module does not imply that the components or functionality described or claimed as part of the module are all configured in a common package. Indeed, any or all of the various components of a module, whether control logic or other components, can be combined in a single package or separately maintained and can further be distributed in multiple groupings or packages or across multiple locations.
  • a synchronization process or synchronization processes are described herein, it is not intended to require that multiple synchronizations occur simultaneously or that multiple computing systems being synchronized each receive the same data.
  • the data can be broadcast to all participating computing systems simultaneously (or close to simultaneously), in other examples the data can be sent to different computing systems or groups of computing systems at different times.
  • the same data, or the same subset of the data can be sent to all computing systems.
  • subsets of the data can be tailored for a given computing system or group of computing systems.

Abstract

Described in detail herein are systems and methods for single instancing blocks of data in a data storage system. For example, the data storage system may include multiple computing devices (e.g., client computing devices) that store primary data. The data storage system may also include a secondary storage computing device, a single instance database, and one or more storage devices that store copies of the primary data (e.g., secondary copies, tertiary copies, etc.). The secondary storage computing device receives blocks of data from the computing devices and accesses the single instance database to determine whether the blocks of data are unique (meaning that no instances of the blocks of data are stored on the storage devices). If a block of data is unique, the single instance database stores it on a storage device. If not, the secondary storage computing device can avoid storing the block of data on the storage devices.

Description

BLOCK-LEVEL SINGLE INSTANCING
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] This application claims the benefit of U.S. Patent Application No. 61/180,791 filed on May 22, 2009 (entitled BLOCK-LEVEL SINGLE INSTANCING, Attorney Docket No. 60692-8073. USOO), and is related to U.S. Patent Application No. 12/565,576 filed on September 23, 2009 (entitled SYSTEMS AND METHODS FOR MANAGING SINGLE INSTANCING DATA, Attorney Docket No. 60692-8067.US01), each of which is incorporated by reference in its entirety.
BACKGROUND
[0002] Single instancing in a data storage system typically involves attempting to store only a single instance of a file on a storage device. In certain single instancing systems, a separate folder on the file system of the storage device is created for each single instancing storage operation performed. Each file that has been single instanced is stored as a separate individual file in the separate folder.
[0003] Because there may be numerous computing systems in the data storage system, each requiring one or more storage operations, these techniques may result in the creation of numerous folders, each containing numerous files. For example, if there are hundreds of computing systems, each having thousands of files, backing up or copying all of these files may potentially result in the creation of millions of files on the storage device.
[0004] Certain file systems of storage devices may not be capable of adequately providing for storing such large numbers of files. Other file systems may be equipped to handle storing millions of files or more, but may not perform optimally in such situations.
[0005] The need exists for systems and methods that overcome the above problems, as well as that provide additional benefits. Overall, the examples herein of some prior or related systems and their associated limitations are intended to be illustrative and not exclusive. Other limitations of existing or prior systems will become apparent to those of skill in the art upon reading the following Detailed Description.
BRIEF DESCRIPTION OF THE DRAWINGS
[0006] Figure 1 is a block diagram illustrating an example of a data storage enterprise that may employ aspects of the invention.
[0007] Figure 2 is a block diagram depicting in more detail certain components illustrated in Figure 1.
[0008] Figure 3 is a flow diagram of certain aspects of a process for performing a storage operation.
[0009] Figure 4 is a flow diagram of other aspects of a process for performing a storage operation.
[0010] Figures 5A and 5B are diagrams illustrating suitable data structures that may be employed by aspects of the invention.
[0011] Figures 6A and 6B are diagrams illustrating suitable data structures that may be employed by aspects of the invention.
[0012] Figure 7 is a diagram illustrating various data structures that aspects of the invention may utilize.
[0013] Figure 8 is a flow diagram of a process for restoring data. [0014] Figure 9 is a flow diagram of a process for pruning data.
DETAILED DESCRIPTION
[0015] The headings provided herein are for convenience only and do not necessarily affect the scope or meaning of the claimed invention.
Overview
[0016] This application describes in detail, among other things, systems and methods for single instancing (alternatively called deduplicating) blocks of data in a data storage system (alternatively called a data storage network, a data storage environment, or a data storage enterprise). The data storage system stores single instanced blocks of data (alternatively referred to as deduplicated blocks of data) in one or more files and maintains one or more data structures (e.g., index files) that keep track of which blocks of data are referenced. This allows the data storage system to, among other things: 1) single-instance data at a more granular level (at a block-level instead of at a file-level); 2) reduce or eliminate redundantly stored data, thereby saving storage space; 3) store very large numbers of blocks of data without regard to file system limitations; and 4) delete data that no longer needs to be stored, while still maintaining data that needs to be stored.
[0017] The data storage system, for example, may include multiple computing devices or computing systems (e.g., client computing devices) that store primary data (e.g., production data such as system files, user files, etc.). The data storage system may also include a secondary storage computing device, a single instance database, and one or more storage devices that store copies of the primary data (e.g., secondary copies, tertiary copies, etc.). The secondary storage computing device receives blocks of data from the computing devices and accesses the single instance database to determine whether the blocks of data are unique (unique meaning that no instances of the blocks of data are already stored on the storage devices). If a block of data is unique, the single instance database stores it in a file on a storage device. If not, the secondary storage computing device can avoid storing the block of data on the storage devices.
[0018] The primary data of the computing devices can be divided into data that is eligible for single instancing and data that is not eligible for single instancing. An example of the latter is metadata (e.g., Master File Table (MFT) information) and an example of the former is data (e.g., operating system and/or application files). A file typically comprises one or more blocks as tracked by the file systems of the computing devices.
[0019] The computing devices align data that is eligible for single instancing into blocks of data (which may comprise one or more blocks as tracked by the file systems of the computing devices) and generate identifiers for the blocks of data that the secondary storage computing device uses to determine if the blocks of data are unique. This allows the secondary storage computing device to avoid generating identifiers for the blocks of data, which may be computationally expensive and/or require a long time to perform. Therefore, the distribution of the task of generating identifiers (which can be computationally expensive operations) across numerous computing devices frees up the secondary storage computing device to perform other operations (e.g., storing data, retrieving data, pruning data, etc.).
[0020] The computing devices send the blocks of data and other data (e.g., metadata and/or the data that is not eligible for single instancing) in a data stream to the secondary storage computing device. The secondary storage computing device receives the data stream and stores blocks of data and their identifiers in buffers in random access memory (RAM). The secondary storage computing device determines whether a block of data is already stored on a storage device. To do this, the secondary storage computing device determines, by analyzing data structures in the single instance database in view of the block's identifier, whether the block of data is already stored on a storage device. If it is, then the secondary storage computing device 1) stores a link to the already stored block of data in a metadata file and 2) discards the block of data from the memory buffer. If it is not, then the secondary storage computing device stores the block of data in a container file.
[0021] Because the size of a block of data and associated metadata is typically less then the size of a memory buffer, the secondary storage computing device can keep a single block of data in a single memory buffer while it looks up its identifier in the single instance database. This allows the secondary storage computing device to avoid writing the block of data to disk (an operation which is typically slower than storing the block of data in a RAM buffer) until the secondary storage computing device determines that it needs to store the block of data in a container file on a storage device. The secondary storage computing device stores data that is not eligible for single instancing in metadata files.
[0022] By storing multiple blocks of data in a single container file, the secondary storage computing device avoids storing each block of data as a separate file on the file systems of the storage devices. This reduces the number of files that would be stored on the file systems of the storage devices, thereby ensuring that the storage devices can adequately store the data of the computing devices in the data storage system.
-A- [0023] One advantage of these techniques is that they significantly reduce the number of files stored on a file system of a computing device or storage device. This is at least partly due to the storage of data blocks within the container files. Even if the secondary storage computing device performs numerous storage operations, these techniques will result in storing far fewer files on the file system than storage operations where each data block is stored as a separate file. Therefore, the file system of the computing device or storage device may not necessarily have to contend with storing excessively large numbers of files, such as millions of files or more. Accordingly, these techniques enable very large numbers of blocks of data to be stored without regard to limitations of the file system of the computing device or storage device.
[0024] However, the storage of blocks of data in container files may create additional complexities when it comes time to prune data. This is because a container file may contain blocks of data that are referenced by links in metadata files and thus cannot be deleted, because referenced blocks of data typically still need to be stored on the storage devices. Furthermore, because the blocks of data are not stored as files on the file systems of the storage devices, they cannot be directly referenced by the file system.
[0025] The systems and methods described herein provide solutions to these problems. The secondary storage computing device creates the container files as sparse files (typically only on operating systems that support sparse files, e.g., Windows operating systems, and other operating systems that support sparse files). A sparse file is type of file that may include empty space (e.g., a sparse file may have real data within it, such as at the beginning of the file and/or at the end of the file, but may also have empty space in it that is not storing actual data, such as a contiguous range of bytes all having a value of zero). Second, the secondary storage computing device maintains a separate index that stores an indication of whether blocks of data in container files are referred to by links in metadata files. In some examples, this can be analogized to using another, non-native file system that keeps track of blocks of data in the container files, on top of the existing, native file systems of the storage devices.
[0026] When a block of data is not referred to and does not need to be stored, the secondary storage computing device can prune it. To prune data, the secondary storage computing device accesses the separate index to determine the blocks of data that are not referred to by links. On operating systems that support sparse files, the secondary storage computing device can free up space in the container files corresponding to those blocks of data by marking the portions of the physical media corresponding to the unreferenced portions of the container file as available for storage (e.g., by zeroing out the corresponding bytes in the container files). On operating systems that do not support sparse files, the secondary storage computing device can free up space in the container files by truncating the extreme portions of the container files (e.g., the extremities such as the beginnings and/or the ends of the container files), thereby making the corresponding portions of the physical media available to store other data. Freeing up space in container files allows the operating system to utilize the freed-up space in other fashions (e.g., other programs may utilize the freed-up space).
[0027] Various examples of the invention will now be described. The following description provides specific details for a thorough understanding and enabling description of these examples. One skilled in the relevant art will understand, however, that the invention may be practiced without many of these details. Likewise, one skilled in the relevant art will also understand that the invention may include many other obvious features not described in detail herein. Additionally, some well-known structures or functions may not be shown or described in detail below, so as to avoid unnecessarily obscuring the relevant description.
[0028] The terminology used below is to be interpreted in its broadest reasonable manner, even though it is being used in conjunction with a detailed description of certain specific examples of the invention. Indeed, certain terms may even be emphasized below; however, any terminology intended to be interpreted in any restricted manner will be overtly and specifically defined as such in this Detailed Description section.
[0029] Figures 1 and 2 and the discussion herein provide a brief, general description of a suitable specialized environment in which aspects of the invention can be implemented. Those skilled in the relevant art will appreciate that aspects of the invention can be practiced with other communications, data processing, or computer system configurations, including: Internet appliances, hand-held devices (including personal digital assistants (PDAs)), wearable computers, all manner of cellular phones, mobile phones, and/or mobile devices, multi-processor systems, microprocessor-based or programmable consumer electronics, set-top boxes, network PCs, mini-computers, mainframe computers, and the like. The terms "computer," "server," "host," "host system," "client," and the like are generally used interchangeably herein, and refer to any of the above devices and systems, as well as any data processor.
[0030] While aspects of the invention, such as certain functions, are described as being performed exclusively on a single device, the invention can also be practiced in distributed environments where functions or modules are shared among disparate processing devices, which are linked through a communications network, such as a Local Area Network (LAN), Wide Area Network (WAN), and/or the Internet. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
[0031] Aspects of the invention may be stored or distributed on computer-readable media, including tangible computer-readable storage media such as magnetically or optically readable computer discs, hard-wired or preprogrammed chips (e.g., EEPROM semiconductor chips), nanotechnology memory, biological memory, or other data storage media. Alternatively, computer implemented instructions, data structures, screen displays, and other data under aspects of the invention may be distributed over the Internet or over other networks (including wireless networks), on a propagated signal on a propagation medium (e.g., an electromagnetic wave(s), a sound wave, etc.) over a period of time, or they may be provided on any analog or digital network (packet switched, circuit switched, or other scheme).
[0032] Aspects of the invention will now be described in detail with respect to Figures 1 through 9. Figure 1 illustrates an example of a data storage system that may employ aspects of the invention. Figure 2 illustrates in more detail certain components illustrated in Figure 1 that may be used to implement a block-level single instancing system. These components include a secondary storage computing device, a single instancing database, and a storage device that stores only a single instance of blocks of data of one or more computing devices (e.g., client computing devices).
[0033] Figure 3 illustrates aspects of a process for copying data that a computing device (e.g., a client computing device) may perform. These aspects include determining whether data is eligible for single instancing and transferring data in a data stream to the secondary storage computing device. Figure 4 illustrates aspects of the copy process that the secondary storage computing device may perform upon receipt of the data stream from the computing device. During this part of the copy process, the secondary storage computing device determines whether the data of the computing device is single instanceable.
[0034] Figures 5A and 5B, 6A and 6B, and 7 are illustrations of various data structures that aspects of the invention may utilize. Figures 5A and 5B depict data streams that the computing device may form during the copy process. Figures 6A and 6B show data structures that may be used by the single instance database to keep track of where blocks of data and references to blocks of data are stored on the storage device. Figure 7 illustrates data structures that may be used to store blocks of data on the storage device.
[0035] Figures 8 and 9 are process flow diagrams. Figure 8 illustrates an example process that the secondary storage computing device may perform to restore data stored on the storage device, such as to a computing device. Figure 9 depicts an example process that the secondary storage computing device may perform to prune data stored on the storage device when it is no longer required to be stored on the storage device.
Suitable Data Storage System
[0036] Figure 1 illustrates an example of one arrangement of resources in a computing network, comprising a data storage system 150. The resources in the data storage system 150 may employ the processes and techniques described herein. The system 150 includes a storage manager 105, one or more data agents 195, one or more secondary storage computing devices 165, one or more storage devices 115, one or more computing devices 130 (called clients 130), one or more data or information stores 160 and 162, and a single instancing database 123. The storage manager 105 includes an index 111 , a jobs agent 120, an interface agent 125, and a management agent 131. The system 150 may represent a modular storage system such as the CommVault QiNetix system, and also the CommVault GALAXY backup system, available from CommVault Systems, Inc. of Oceanport, NJ, aspects of which are further described in the commonly-assigned U.S. Patent Application No. 09/610,738, now U.S. Patent No. 7,035,880, the entirety of which is incorporated by reference herein. The system 150 may also represent a modular storage system such as the CommVault Simpana system, also available from CommVault Systems, Inc.
[0037] The system 150 may generally include combinations of hardware and software components associated with performing storage operations on electronic data. Storage operations include copying, backing up, creating, storing, retrieving, and/or migrating primary storage data (e.g., data stores 160 and/or 162) and secondary storage data (which may include, for example, snapshot copies, backup copies, hierarchical storage management (HSM) copies, archive copies, and other types of copies of electronic data stored on storage devices 115). The system 150 may provide one or more integrated management consoles for users or system processes to interface with in order to perform certain storage operations on electronic data as further described herein. Such integrated management consoles may be displayed at a central control facility or several similar consoles distributed throughout multiple network locations to provide global or geographically specific network data storage information.
[0038] In one example, storage operations may be performed according to various storage preferences, for example, as expressed by a user preference, a storage policy, a schedule policy, and/or a retention policy. A "storage policy" is generally a data structure or other information source that includes a set of preferences and other storage criteria associated with performing a storage operation. The preferences and storage criteria may include, but are not limited to, a storage location, relationships between system components, network pathways to utilize in a storage operation, data characteristics, compression or encryption requirements, preferred system components to utilize in a storage operation, a single instancing or variable instancing policy to apply to the data, and/or other criteria relating to a storage operation. For example, a storage policy may indicate that certain data is to be stored in the storage device 115, retained for a specified period of time before being aged to another tier of secondary storage, copied to the storage device 115 using a specified number of data streams, etc.
[0039] A "schedule policy" may specify a frequency with which to perform storage operations and a window of time within which to perform them. For example, a schedule policy may specify that a storage operation is to be performed every Saturday morning from 2:00 a.m. to 4:00 a.m. A "retention policy" may specify how long data is to be retained at specific tiers of storage or what criteria must be met before data may be pruned or moved from one tier of storage to another tier of storage. In some cases, the storage policy includes information generally specified by the schedule policy and/or the retention policy. (Put another way, the storage policy includes the schedule policy and/or the retention policy.) Storage policies, schedule policies and/or retention policies may be stored in a database of the storage manager 105, to archive media as metadata for use in restore operations or other storage operations, or to other locations or components of the system 150.
[0040] The system 150 may comprise a storage operation cell that is one of multiple storage operation cells arranged in a hierarchy or other organization. Storage operation cells may be related to backup cells and provide some or all of the functionality of backup cells as described in the assignee's U.S. Patent Application No. 09/354,058, now U.S. Patent No. 7,395,282, which is incorporated herein by reference in its entirety. However, storage operation cells may also perform additional types of storage operations and other types of storage management functions that are not generally offered by backup cells.
[0041] Storage operation cells may contain not only physical devices, but also may represent logical concepts, organizations, and hierarchies. For example, a first storage operation cell may be configured to perform a first type of storage operations such as HSM operations, which may include backup or other types of data migration, and may include a variety of physical components including a storage manager 105 (or management agent 131), a secondary storage computing device 165, a client 130, and other components as described herein. A second storage operation cell may contain the same or similar physical components; however, it may be configured to perform a second type of storage operations, such as storage resource management (SRM) operations, and may include monitoring a primary data copy or performing other known SRM operations.
[0042] Thus, as can be seen from the above, although the first and second storage operation cells are logically distinct entities configured to perform different management functions (i.e., HSM and SRM, respectively), each storage operation cell may contain the same or similar physical devices. Alternatively, different storage operation cells may contain some of the same physical devices and not others. For example, a storage operation cell configured to perform SRM tasks may contain a secondary storage computing device 165, client 130, or other network device connected to a primary storage volume, while a storage operation cell configured to perform HSM tasks may instead include a secondary storage computing device 165, client 130, or other network device connected to a secondary storage volume and not contain the elements or components associated with and including the primary storage volume. (The term "connected" as used herein does not necessarily require a physical connection; rather, it could refer to two devices that are operably coupled to each other, communicably coupled to each other, in communication with each other, or more generally, refer to the capability of two devices to communicate with each other.) These two storage operation cells, however, may each include a different storage manager 105 that coordinates storage operations via the same secondary storage computing devices 165 and storage devices 115. This "overlapping" configuration allows storage resources to be accessed by more than one storage manager 105, such that multiple paths exist to each storage device 115 facilitating failover, load balancing, and promoting robust data access via alternative routes.
[0043] Alternatively or additionally, the same storage manager 105 may control two or more storage operation cells (whether or not each storage operation cell has its own dedicated storage manager 105). Moreover, in certain embodiments, the extent or type of overlap may be user-defined (through a control console) or may be automatically configured to optimize data storage and/or retrieval.
[0044] The clients 130 typically include application software for performing various operations. Clients 130 typically also include an operating system on which the application software runs. A file system can be provided to facilitate and control file access by the operating system and application software. File systems can facilitate access to local and remote storage devices for file or data access and storage. Clients 130 can also include local storage such as a media module media drive with fixed or removable media.
[0045] In some examples, the clients 130 include storage mechanisms for allowing computer programs or other instructions or data to be loaded into memory for execution. Such storage mechanisms might include, for example, a fixed or removable storage unit and an interface. Examples of such storage units and interfaces can include a program cartridge and cartridge interface, a removable memory (for example, a flash memory or other removable memory module) and memory slot, a PCMCIA slot and card, and other fixed or removable storage units and interfaces that allow software and data to be transferred from the storage unit to memory.
[0046] Data agent 195 may be a software module or part of a software module that is generally responsible for performing storage operations on the data of the client 130 stored in data store 160/162 or other memory location. Each client 130 may have at least one data agent 195 and the system 150 can support multiple clients 130. Data agent 195 may be distributed between client 130 and storage manager 105 (and any other intermediate components), or it may be deployed from a remote location or its functions approximated by a remote process that performs some or all of the functions of data agent 195.
[0047] As used herein, the term module might describe a given unit of functionality that can be performed in accordance with one or more embodiments of the present invention. As used herein, a module might be implemented utilizing any form of hardware, software, firmware, or a combination thereof. For example, one or more processors, controllers, ASICs, PLAs, logical components, software routines or other mechanisms might be implemented to make up a module. In implementation, the various modules described herein might be implemented as discrete modules or the functions and features described can be shared in part or in total among one or more modules. In other words, as would be apparent to one of ordinary skill in the art after reading this description, the various features and functionality described herein may be implemented in any given application and can be implemented in one or more separate or shared modules in various combinations and permutations. Even though various features or elements of functionality may be individually described or claimed as separate modules, one of ordinary skill in the art will understand that these features and functionality can be shared among one or more common software and hardware elements, and such description shall not require or imply that separate hardware or software components are used to implement such features or functionality.
[0048] The overall system 150 may employ multiple data agents 195, each of which may perform storage operations on data associated with a different application. For example, different individual data agents 195 may be designed to handle Microsoft Exchange data, Lotus Notes data, Microsoft Windows file system data, Microsoft Active Directory Objects data, Microsoft SQL Server data, Microsoft Sharepoint Server data, and other types of data known in the art. Other embodiments may employ one or more generic data agents 195 that can handle and process multiple data types rather than using the specialized data agents described above.
[0049] If a client 130 has two or more types of data, one data agent 195 may be required for each data type to perform storage operations on the data of the client 130. For example, to back up, migrate, and restore all the data on a Microsoft Exchange server, the client 130 may use one Microsoft Exchange Mailbox data agent 195 to back up the Exchange mailboxes, one Microsoft Exchange Database data agent 195 to back up the Exchange databases, one Microsoft Exchange Public Folder data agent 195 to back up the Exchange Public Folders, and one Microsoft Windows File System data agent 195 to back up the file system of the client 130. These data agents 195 would be treated as four separate data agents 195 by the system even though they reside on the same client 130.
[0050] Alternatively, the overall system 150 may use one or more generic data agents 195, each of which may be capable of handling two or more data types. For example, one generic data agent 195 may be used to back up, migrate and restore Microsoft Exchange Mailbox data and Microsoft Exchange Database data while another generic data agent 195 may handle Microsoft Exchange Public Folder data and Microsoft Windows File System data, etc.
[0051] Data agents 195 may be responsible for arranging or packing data to be copied or migrated into a certain format such as an archive file. Nonetheless, it will be understood that this represents only one example, and any suitable packing or containerization technique or transfer methodology may be used if desired. Such an archive file may include metadata, a list of files or data objects copied, the file, and data objects themselves. Moreover, any data moved by the data agents may be tracked within the system by updating indexes associated with appropriate storage managers 105 or secondary storage computing devices 165. As used herein, a file or a data object refers to any collection or grouping of bytes of data that can be viewed as one or more logical units. [0052] Generally speaking, storage manager 105 may be a software module or other application that coordinates and controls storage operations performed by the system 150. Storage manager 105 may communicate with some or all elements of the system 150, including clients 130, data agents 195, secondary storage computing devices 165, and storage devices 115, to initiate and manage storage operations (e.g., backups, migrations, data recovery operations, etc.).
[0053] Storage manager 105 may include a jobs agent 120 that monitors the status of some or all storage operations previously performed, currently being performed, or scheduled to be performed by the system 150. (One or more storage operations are alternatively referred to herein as a "job" or "jobs.") Jobs agent 120 may be communicatively coupled to an interface agent 125 (e.g., a software module or application). Interface agent 125 may include information processing and display software, such as a graphical user interface ("GUI"), an application programming interface ("API"), or other interactive interface through which users and system processes can retrieve information about the status of storage operations. For example, in an arrangement of multiple storage operations cell, through interface agent 125, users may optionally issue instructions to various storage operation cells regarding performance of the storage operations as described and contemplated herein. For example, a user may modify a schedule concerning the number of pending snapshot copies or other types of copies scheduled as needed to suit particular needs or requirements. As another example, a user may employ the GUI to view the status of pending storage operations in some or all of the storage operation cells in a given network or to monitor the status of certain components in a particular storage operation cell (e.g., the amount of storage capacity left in a particular storage device 115).
[0054] Storage manager 105 may also include a management agent 131 that is typically implemented as a software module or application program. In general, management agent 131 provides an interface that allows various management agents 131 in other storage operation cells to communicate with one another. For example, assume a certain network configuration includes multiple storage operation cells hierarchically arranged or otherwise logically related in a WAN or LAN configuration. With this arrangement, each storage operation cell may be connected to the other through each respective interface agent 125. This allows each storage operation cell to send and receive certain pertinent information from other storage operation cells, including status information, routing information, information regarding capacity and utilization, etc. These communications paths may also be used to convey information and instructions regarding storage operations.
[0055] For example, a management agent 131 in a first storage operation cell may communicate with a management agent 131 in a second storage operation cell regarding the status of storage operations in the second storage operation cell. Another illustrative example includes the case where a management agent 131 in a first storage operation cell communicates with a management agent 131 in a second storage operation cell to control storage manager 105 (and other components) of the second storage operation cell via management agent 131 contained in storage manager 105.
[0056] Another illustrative example is the case where management agent 131 in a first storage operation cell communicates directly with and controls the components in a second storage operation cell and bypasses the storage manager 105 in the second storage operation cell. If desired, storage operation cells can also be organized hierarchically such that hierarchically superior cells control or pass information to hierarchically subordinate cells or vice versa.
[0057] Storage manager 105 may also maintain an index, a database, or other data structure 111. The data stored in database 111 may be used to indicate logical associations between components of the system, user preferences, management tasks, media containerization and data storage information or other useful data. For example, the storage manager 105 may use data from database 111 to track logical associations between secondary storage computing device 165 and storage devices 115 (or movement of data as containerized from primary to secondary storage).
[0058] Generally speaking, the secondary storage computing device 165, which may also be referred to as a media agent, may be implemented as a software module that conveys data, as directed by storage manager 105, between a client 130 and one or more storage devices 115 such as a tape library, a magnetic media storage device, an optical media storage device, or any other suitable storage device. In one embodiment, secondary storage computing device 165 may be communicatively coupled to and control a storage device 115. A secondary storage computing device 165 may be considered to be associated with a particular storage device 115 if that secondary storage computing device 165 is capable of routing and storing data to that particular storage device 115.
[0059] In operation, a secondary storage computing device 165 associated with a particular storage device 115 may instruct the storage device to use a robotic arm or other retrieval means to load or eject a certain storage media, and to subsequently archive, migrate, or restore data to or from that media. Secondary storage computing device 165 may communicate with a storage device 115 via a suitable communications path such as a SCSI or Fibre Channel communications link. In some embodiments, the storage device 115 may be communicatively coupled to the storage manager 105 via a SAN.
[0060] Each secondary storage computing device 165 may maintain an index, a database, or other data structure 161 that may store index data generated during storage operations for secondary storage (SS) as described herein, including creating a metabase (MB). For example, performing storage operations on Microsoft Exchange data may generate index data. Such index data provides a secondary storage computing device 165 or other external device with a fast and efficient mechanism for locating data stored or backed up. Thus, a secondary storage computing device index 161 , or a database 111 of a storage manager 105, may store data associating a client 130 with a particular secondary storage computing device 165 or storage device 115, for example, as specified in a storage policy, while a database or other data structure in secondary storage computing device 165 may indicate where specifically the data of the client 130 is stored in storage device 115, what specific files were stored, and other information associated with storage of the data of the client 130. In some embodiments, such index data may be stored along with the data backed up in a storage device 115, with an additional copy of the index data written to index cache in a secondary storage device. Thus the data is readily available for use in storage operations and other activities without having to be first retrieved from the storage device 115.
[0061] Generally speaking, information stored in cache is typically recent information that reflects certain particulars about operations that have recently occurred. After a certain period of time, this information is sent to secondary storage and tracked. This information may need to be retrieved and uploaded back into a cache or other memory in a secondary computing device before data can be retrieved from storage device 115. In some embodiments, the cached information may include information regarding format or containerization of archives or other files stored on storage device 115.
[0062] One or more of the secondary storage computing devices 165 may also maintain one or more single instance databases 123. More details as to single instancing may be found in one or more of the following commonly-assigned U.S. patent applications: 1) U.S. Pat. App. No. 11/269,512 (entitled SYSTEM AND METHOD TO SUPPORT SINGLE INSTANCE STORAGE OPERATIONS, Attorney Docket No. 60692-8023.US00); 2) U.S. Pat. App. No. 12/145,347 (entitled APPLICATION-AWARE AND REMOTE SINGLE INSTANCE DATA MANAGEMENT, Attorney Docket No. 60692-8056. USOO); or 3) U.S. Pat. App. No. 12/145,342 (entitled APPLICATION- AWARE AND REMOTE SINGLE INSTANCE DATA MANAGEMENT, Attorney Docket No. 60692-8057.US00), 4) U.S. Pat. App. No. 11/963,623 (entitled SYSTEM AND METHOD FOR STORING REDUNDANT INFORMATION, Attorney Docket No. 60692- 8036.US02); 5) U.S. Pat. App. No. 11/950,376 (entitled SYSTEMS AND METHODS FOR CREATING COPIES OF DATA SUCH AS ARCHIVE COPIES, Attorney Docket No. 60692-8037.US01); or 6) the previously referenced U.S. Patent Application No. 12/565,576, each of which is incorporated by reference herein in its entirety.
[0063] In some examples, the secondary storage computing devices 165 maintain one or more variable instance databases. Variable instancing generally refers to storing in secondary storage one or more instances, but fewer than the total number of instances, of each data block (or data object) in a set of data (e.g., primary data). More details as to variable instancing may be found in the commonly-assigned U.S. Pat. App. No. 61/164,803 (entitled STORING A VARIABLE NUMBER OF INSTANCES OF DATA OBJECTS, Attorney Docket No. 60692-8068.US00).
[0064] In some embodiments, certain components may reside and execute on the same computer. For example, in some embodiments, a client 130 such as a data agent 195, or a storage manager 105, coordinates and directs local archiving, migration, and retrieval application functions as further described in the previously- referenced U.S. Patent Application No. 09/610,738. This client 130 can function independently or together with other similar clients 130.
[0065] As shown in Figure 1 , each secondary storage computing device 165 has its own associated metabase 161. Each client 130 may also have its own associated metabase 170. However in some embodiments, each "tier" of storage, such as primary storage, secondary storage, tertiary storage, etc., may have multiple metabases or a centralized metabase, as described herein. For example, rather than a separate metabase or index associated with each client 130 in Figure 1 , the metabases on this storage tier may be centralized. Similarly, second and other tiers of storage may have either centralized or distributed metabases. Moreover, mixed architecture systems may be used if desired, that may include a first tier centralized metabase system coupled to a second tier storage system having distributed metabases and vice versa, etc.
[0066] Moreover, in operation, a storage manager 105 or other management module may keep track of certain information that allows the storage manager 105 to select, designate, or otherwise identify metabases to be searched in response to certain queries as further described herein. Movement of data between primary and secondary storage may also involve movement of associated metadata and other tracking information as further described herein.
[0067] In some examples, primary data may be organized into one or more sub- clients. A sub-client is a portion of the data of one or more clients 130, and can contain either all of the data of the clients 130 or a designated subset thereof. As depicted in Figure 1 , the data store 162 includes two sub-clients. For example, an administrator (or other user with the appropriate permissions; the term administrator is used herein for brevity) may find it preferable to separate email data from financial data using two different sub-clients having different storage preferences, retention criteria, etc.
Components of a Block-Level Single Instancing System
[0068] Figure 2 is a block diagram depicting in more detail certain components illustrated in Figure 1. The data agent 195 of the client 130 includes various components, such as a data identification component 202, a block identification component 204, and an identifier generation component 206. The data agent 195 also includes a compression component 210, an encryption component 212, and a data stream generation component 214. Various functions performed by these components are described herein.
[0069] In addition to the data agent 195, the client 130 includes data 240. The data 240 includes single instanceable data (Sl data) 242 and non-single instanceable data (non-SI data) 244. SI data 242 includes data that is eligible for single instancing. Non-SI data 244 includes data that is not eligible for single instancing. Non-SI data 244 may include metadata such as access control lists (ACLs), disk partition information, Master File Table (MFT) or File Allocation Table (FAT) information, and/or other metadata. Non-SI data 244 may also include other data that is determined not to be single instanceable. SI data 242 may include data 240 of the client 130 other than non- SI data 244 (e.g., system files, application files, user files, etc.).
[0070] The secondary storage computing device 165 includes a data stream reception component 220 and an identifier comparison component 222. Various functions performed by these components are also described in detail herein. The secondary storage computing device 165 also includes a memory 230, which includes multiple buffers 232. The secondary storage computing device 165 may also include other components, such as a decompression component and/or a decryption component. The single instance database 123 includes data structures 250 that are used to store data, such as metadata about SI data 242. The storage device 115 also includes data structures 260 that are used to store data, such as SI data 242 and non- SI data 244. In some examples, the secondary storage computing device 165 includes the components that the client 130 includes, and performs the functions that the client 130 performs.
Processes for Performing Storage Operations
[0071] Figures 3 and 4 are flow diagrams illustrating certain aspects of processes 300 and 400, respectively, for performing a storage operation such as a copy operation. A storage operation (alternatively referred to as a job) is typically performed on files stored on file systems of one or more clients 130. One or more of the entities illustrated in the figures (e.g., Figures 1 and/or 2) may perform different aspects of the processes 300 and 400. In some examples, a storage manager 105 instigates the process 300 by sending an indication specifying the storage operation to the data agent 195. The data agent 195 accesses the data of the client 130 (e.g., accesses files stored on the filesystem of the client 130). The data agent 195 sends the data to the secondary storage computing device 165, which then stores the data on one or more storage devices 115. In some examples, less than all of these entities may be involved in performing the storage operation. The process 300 is described as being performed by the data agent 195 and the process 400 is described as being performed by the secondary storage computing device 165. However, those of skill in the art will understand that aspects of the processes 300 and 400 may be performed by any one or more of the entities described herein (e.g., the data agent 195, the storage manager 105, the secondary storage computing device 165, etc.).
[0072] The process 300 begins at step 305 where the data agent 195 receives an indication to copy data of the client 130. The storage manager 105 may send the indication to the data agent 195 (e.g., according to a storage policy), an administrator may manually start the process 300, and/or the process 300 may be automatically started according to a schedule policy.
[0073] At step 310 the data agent 195 accesses the data 240 of the client 130. The data agent 195 (e.g., the data identification component 202) determines which portions of the data 240 are SI data 242 and which portions are non-SI data 244. For example, the data agent 195 may determine that metadata (e.g., MFT, FAT, volume information, transaction logs, etc.) on the file system of the client 130 is non-SI data 244, and that data other than metadata is SI data 242 (e.g., system files, user files, etc.). At step 315 the data agent 195 (e.g., the data stream generation component 214) forms a data stream of multiple pairs of stream header and stream payload from the SI data 242 and the non-SI data 244. (An example data stream is illustrated in Figure 5A and is described in detail below.) A data stream, therefore, comprises multiple pairs of stream header and stream payload. However, those of skill in the art will understand that data streams may contain data organized in other fashions. For the SI data 242, the data agent 195 may set a flag in the stream header to indicate that the corresponding stream payload contains single instanceable data.
[0074] At step 320, the data agent 195 (e.g., the identifier generation component 206) aligns the stream header and stream payload into one or more fixed size blocks of data. (An example data stream with stream header and stream payload aligned into multiple blocks is illustrated in Figure 5B and is described in detail below.) A block of data (alternatively called a data block) is a sequence of bits or bytes having a nominal length (a data block size). The file system of the client 130 may track its data 240 in blocks (alternatively called clusters) in sizes of 512 bytes, 4Kb, 16Kb, or other sizes. (Put another way, a block may be a subset of one or more data objects.) A file on the file system of the client 130 typically spans one or more blocks (e.g., a file of size 10Kb may span 3 blocks of size 4Kb). The data agent 195 typically aligns data blocks such that they have the same size, which may be 32Kb, 64Kb, 128Kb, 256Kb, 512Kb, or other sizes. Accordingly, the term data block, as used herein, may comprise one or more blocks as tracked by the file system of the clients 130. For example, if the file system of a client 130 tracks its data 240 in blocks of size 4Kb and if the data agent 195 aligns the client's 130 data 240 into data blocks of size 128Kb, then these 128Kb data blocks comprise 32 blocks of data 240 as tracked by the file system of the client 130.
[0075] At step 325 the data agent 195 determines whether a data block is single instanceable. The data agent 195 does so by analyzing the portion of the one or more corresponding stream headers that indicates whether the data block is single instanceable. For example, the stream headers may contain a flag or bit that indicates whether the successive stream payload contain single instanceable data. (For example, see Figure 5A, illustrating stream headers containing such flags.) If the data block is single instanceable, the process 300 continues at step 330, where the data agent 195 (e.g., the identifier generation component 206) generates an identifier for the data block.
[0076] Examples of identifiers include a hash value, message digest, checksum, digital fingerprint, digital signature or other sequence of bytes that substantially uniquely identifies the data block in the data storage system. For example, identifiers could be generated using Message Digest Algorithm 5 (MD5) or Secure Hash Algorithm SHA 512. In some instances, the phrase "substantially unique" is used to modify the term "identifier" because algorithms used to produce hash values may result in collisions, where two different data objects, when hashed, result in the same hash value. However, depending upon the algorithm or cryptographic hash function used, collisions should be suitably rare and thus the identifier generated for a block should be unique throughout the data storage system. The term "probabilistically unique identifier" may also be used. In this case, the phrase "probabilistically unique" is used to indicate that collisions should be low-probability occurrences, and, therefore, the identifier should be unique throughout the data storage system.
[0077] At step 335 the data agent 195 (e.g., the identifier generation component 206) inserts the generated identifier into the data stream. The generated identifier may be comprised in an identifier header and identifier data pair that immediately follows the data block for which it is generated. (See Figure 5B and the accompanying description for additional details of the identifier header and identifier data pair.) At step 340 the data agent 195 determines whether there are more data blocks. If so, the process 300 returns to step 325. If not, the process 300 continues at step 345, where the data agent 195 transfers the data stream to the secondary storage computing device 165. The process 300 then ends. In some examples, the data agent 195 may perform additional operations upon the stream header and/or stream payload, such as encrypting the stream payload (e.g., using the encryption component 212) and/or compressing the stream payload (e.g., using the compression component 210).
[0078] Figure 4 is a flow diagram illustrating certain aspects of the process 400 that the secondary storage computing device 165 performs upon receiving the data stream from the data agent 195. At step 405 the secondary storage computing device 165 receives the data stream from the data agent 195. At step 410, the secondary storage computing device 165 stores the stream header and stream payload corresponding to a data block in a buffer 232 of the memory 230. The secondary storage computing device 165 can store the entire stream header and stream payload pairs corresponding to a single block in a single buffer, because the buffer size (e.g., approximately 640Kb) is greater than the size of the stream header and stream payload pairs (e.g., up to approximately 512Kb). The buffer size is typically no greater than 10 times the size of the stream header and stream payload pairs. In some examples, the memory 230 includes 30 buffers 232, thus allowing the secondary storage computing device 165 to simultaneously store up to 30 different data blocks in fast-access memory. The ability to store multiple data blocks in memory enables the secondary storage computing device 165 to avoid writing the multiple data blocks to disk, which can be a lengthy operation. [0079] At step 415 the secondary storage computing device 165 determines whether the data block is single instanceable. The secondary storage computing device 165 may do so, for example, by analyzing the metadata in the stream header that indicates whether the data block is single instanceable (e.g., a flag or bit that indicates whether the data block is single instanceable).
[0080] If the data block is single instanceable, the process 400 continues at step 425, where the secondary storage computing device (e.g., the identifier comparison component 222) obtains the identifier corresponding to the data block (e.g., from the identifier data of the data stream) and looks up the identifier. The secondary storage computing device 165 looks up the identifier in the primary table in the single instance database 123. (Example data structures used by the single instance database 123 are illustrated in Figures 6A and 6B and described with reference to these figures).
[0081] At step 430, if the secondary storage computing device 165 finds the identifier of the data block in the primary table, this indicates that an instance of the data block is already stored on the storage device 115, and that the block of data should not be stored. Accordingly, the secondary storage computing device 165 can avoid storing another instance of the data block and can instead store a link (alternatively called a pointer) to the location(s) of the already stored instance. At step 445 the secondary storage computing device 165 adds a link to the location(s) of the already stored instance of the data block to a metadata file. The link refers or points to the already stored instance of the data block. For example, the secondary storage computing device 165 may add as the link to the metadata file the record of the already stored instance of the data block in the primary table. At step 450 the secondary storage computing device 165 adds an entry to the secondary table in the single instance database. The entry includes the location of the link in the metadata file. The secondary storage computing device 165 also increments a reference count corresponding to the data block in the primary table. The reference count indicates the number of links to the already stored instance of the data block. At step 455 the secondary storage computing device 165 discards the stream header and stream payload corresponding to the data block from the buffer 232 of the memory 230. Additionally or alternatively, the secondary storage computing device 165 may indicate that the buffer is available for storing another pair of stream header and stream payload.
[0082] If the secondary storage computing device 165 does not find the identifier of the block in the primary table (step 430), this indicates that no instances of the data block are already stored on the storage device 115, and that the block of data should be stored. Accordingly, at step 435 the secondary storage computing device 165 stores the data block in a container file on the storage device 115. (See Figure 7 and the accompanying description for additional details of container files.) At step 440 the secondary storage computing device 165 adds an entry to the primary table in the single instance database. The entry includes the location of the data block in the container file.
[0083] If the data block is not single instanceable (step 415), the process 400 continues at step 420, where the secondary storage computing device 165 stores the block in a metadata file. (See Figure 7 and the accompanying description for additional details of metadata files.) The three branches of the process 400 converge at step 460, where the secondary storage computing device 165 determines whether there are more data blocks. If so, the process 400 returns to step 415. If not the process 400 concludes.
[0084] In some examples, the secondary storage computing device 165 may perform additional operations during the process 400, such as decrypting the stream payload (e.g., using a decryption component) and/or decompressing the stream payload (e.g., using a decompression component). The secondary storage computing device 165 may also store in the index 161 , for the data blocks, information mapping an archive file and offset to the physical location of the data blocks. An archive file is a logical entity that is created during a storage operation and that corresponds to physical locations of data blocks on the storage device 115. The storage manager 105 may map archive files to physical locations and keep such information in index 111.
[0085] In some examples, a variable number of instances of data blocks (e.g., more than one instance and up to N-1 instances, where N is the number of instances of the data block in primary data) is stored on the storage devices 115. In such examples, the secondary storage computing devices 165 may use techniques described in the previously referenced U.S. Pat. App. No. 61/164,803 to ensure that a sufficient number of instances of the blocks of data are stored on the storage devices 115. Storing multiple instances (up to N-1) of N data blocks provides for less risk of data loss than single instance storage techniques, and generally nearly as less risk of data loss as conventional data protection techniques (which store N instances of N data blocks). Storing multiple instances (up to N-1) of N data blocks also provides for more efficient use of available storage space than conventional data protection techniques, and almost as efficient use as single instance storage techniques. Accordingly, the storing of a variable number of instances of data blocks enables an administrator to tailor data protection to strike an appropriate balance between 1) minimizing the risk of data loss, and 2) making efficient use of available data storage space, in accordance with the administrator's requirements.
Suitable Data Structures
[0086] Figures 5A and 5B are diagrams of example data streams 500 and 550, respectively, that may be employed by aspects of the invention. Referring to Figure 5A, the data agent 195 forms the data stream 500 from the data 240 of the client 130. The data stream 500 is composed of multiple pairs of stream header 502 and stream payload 504. A stream payload 504 includes SI data 242 and/or non-SI data 244. A stream header 502 includes metadata about the stream payload 504. This metadata may include, for example, a length of the stream payload 504, an indication of whether the stream payload 504 is encrypted, an indication of whether the stream payload 504 is compressed, an archive file identifier (ID), an indication of whether the stream payload 504 is single instanceable, and an indication of whether the stream payload 504 is a start of a block of data.
[0087] Referring to Figure 5B, the data stream 550 has the stream header 502 and stream payload 504 aligned into multiple data blocks. In this example, the data blocks are of size 64Kb. The first two stream header 502 and stream payload 504 pairs comprise a first data block of size 64Kb. The first stream header 502 indicates that the length of the succeeding stream payload 504 is 63Kb and that it is the start of a data block. (The stream header 502 may also include the metadata discussed with reference to the stream headers 502 illustrated in Figure 3A.) The next stream header 502 indicates that the succeeding stream payload 504 has a length of 1 Kb and that it is not the start of a new data block. Immediately following stream payload 504 are an identifier header 506 and identifier data 508 pair. The identifier header 506 includes an indication that the succeeding identifier data 508 includes the identifier for the immediately previous data block. The identifier data 508 includes the identifier that the data agent (e.g., the identifier generation component 206) generated for the data block. The data stream 550 also includes other stream header 502 and stream payload 504 pairs, which may be for SI data 242 and/or for non-SI data 244.
[0088] Figures 6A and 6B are diagrams illustrating the data structures 250 that may be used by the single instance database 123. The data structures 250 do not form part of a native file system of a storage device storing the single instance database 123. Alternatively, the data structures 250 are not provided by any native file system for storage devices at least as of the time of the filing of the provisional patent application to which this application claims priority. The data structures 250 include a primary table 600 and a secondary table 650.
[0089] Referring to Figure 6A, the primary table 600 includes an identifier column 602 in which a data block identifier is stored, a location column 604 in which a location of the data block in a container file is stored, an offset column 606 indicating the offset within the container file corresponding to the location of the data block, and a reference count column 608, which contains a reference count of the number of links that refer to the data block. For example, row 620 includes information about a data block for which the identifier is "0xA1 B3FG." This data block is located in the container file that is indicated in the location column 606, at an offset of 10 within the container file. As indicated in the reference count column 608, this data block is referred to twice, meaning that there are two links that refer to the data block. As another example, row 624 includes information about a data block for which the identifier is "0xC13804." The location of this data block is indicated in the location column 604 at an offset of 38 within the container file, and it is referred to one other time, by one link.
[0090] Referring to Figure 6B, the secondary table 650 includes information about links that refer to data blocks. The secondary table 650 includes an identifier column 652, a referring location column 654, and an offset column 656. For example, row 660 includes information about a reference to the data block having the identifier of "0xA1 B3FG" (row 620 in the primary table 600). The location of the link is indicated in column 654, at an offset of five within the indicated metadata file. As another example, row 662 includes information about another reference to the data block having the identifier of "0xA1 B3FG." This link is located at the location indicated in column 654, at an offset of 15 within the indicated metadata file. As another example, row 664 includes information about a reference to the block for which the identifier is "0xC13804" (row 624 in the primary table 600). The location of the link is indicated in column 654, at an offset of 19 within the indicated metadata file.
[0091] Figure 7 is a diagram illustrating the data structures 260 that may be used to store blocks of SI data and non-SI data on the storage device 115. The data structures 260 do not form part of a native file system of the storage device 115. Alternatively, the data structures 260 are not provided by any native file systems for storage devices at least as of the time of the filing of the provisional patent application to which this application claims priority.
[0092] The data structures 260 include one or more volume folders 702, one or more chunk folders 704/705 within a volume folder 702, and multiple files within a chunk folder 704. Each chunk folder 704/705 includes a metadata file 706/707, a metadata index file 708/709, one or more container files 710/711/713, and a container index file 712/714. The metadata file 706/707 stores non-SI data blocks as well as links to SI data blocks stored in container files. The metadata index file 708/709 stores an index to the data in the metadata file 706/707. The container files 710/711/713 store SI data blocks. The container index file 712/714 stores an index to the container files 710/711/713. Among other things, the container index file 712/714 stores an indication of whether a corresponding block in a container file 710/711/713 is referred to by a link in a metadata file 706/707. For example, data block B2 in the container file 710 is referred to by a link in the metadata file 707 in the chunk folder 705. Accordingly, the corresponding index entry in the container index file 712 indicates that the data block B2 in the container file 710 is referred to. As another example, data block B1 in the container file 711 is referred to by a link in the metadata file 707, and so the corresponding index entry in the container index file 712 indicates that this data block is referred to. [0093] As an example, the data structures 260 illustrated in Figure 7 may have been created as a result of two storage operations involving two clients 130. For example, a first storage operation on a first client 130 could result in the creation of the first chunk folder 704, and a second storage operation on a second client 130 could result in the creation of the second chunk folder 705. The container files 710/711 in the first chunk folder 704 would contain the blocks of SI data 242 of the first client 130. If the two clients 130 have substantially similar data 240, the second storage operation on the data 240 of the second client 130 would result in the secondary storage computing device 165 storing primarily links to the data blocks of the first client 130 that are already stored in the container files 710/711. Accordingly, while a first storage operation may result in storing nearly all of the data subject to the storage operation, subsequent storage operations involving similar data may result in substantial data storage space savings, because links to already stored data blocks can be stored instead of additional instances of data blocks.
[0094] If the operating system of the secondary storage computing device 165 supports sparse files, then when the secondary storage computing device 165 creates container files 710/711/713, it can create them as sparse files. As previously described, a sparse file is type of file that may include empty space (e.g., a sparse file may have real data within it, such as at the beginning of the file and/or at the end of the file, but may also have empty space in it that is not storing actual data, such as a contiguous range of bytes all having a value of zero). Having the container files 710/711/713 be sparse files allows the secondary storage computing device 165 to free up space in the container files 710/711/713 when blocks of data in the container files 710/711/713 no longer need to be stored on the storage devices 115. In some examples, the secondary storage computing device 165 creates a new container file 710/711/713 when a container file 710/711/713 either includes 100 blocks of data or when the size of the container file 710 exceeds 50 Mb. In other examples, the secondary storage computing device 165 creates a new container file 710/711/713 when a container file 710/711/713 satisfies other criteria (e.g., it contains from approximately 100 to approximately 1000 blocks or when its size exceeds approximately 50 Mb to 1Gb). Those of skill in the art will understand that the secondary storage computing device 165 can create a new container file 710/711/713 when other criteria are met.
[0095] In some cases, a file on which a storage operation is performed may comprise a large number of data blocks. For example, a 100Mb file may be comprised in 400 data blocks of size 256Kb. If such a file is to be stored, its data blocks may span more than one container file, or even more than one chunk folder. As another example, a database file of 20Gb may comprise over 40,000 data blocks of size 512Kb. If such a database file is to be stored, its data blocks will likely span multiple container files, multiple chunk folders, and potentially multiple volume folders. As described in detail herein, restoring such files may thus requiring accessing multiple container files, chunk folders, and/or volume folders to obtain the requisite data blocks.
[0096] One advantage of the data structures 260 illustrated in Figure 7 and/or of the techniques described herein is that they significantly reduce the number of files stored on a file system of the storage device 115. This is at least partly due to the storage of data blocks within the container files 710/711/713. Even if numerous storage operations using these data structures 260 are performed, this will result in far fewer files on the storage device 115 than storage operations where each data block is stored as a separate file. Therefore, the file system of the storage device 115 may not necessarily have to contend with storing excessively large numbers of files, such as millions of files or more. Accordingly, the systems and methods described herein enable very large numbers of blocks of data to be stored without regard to limitations of the file system of the storage device 115.
[0097] Another advantage is that the data storage system enables a reduction in the amount of blocks of data stored on the storage devices 115, while still maintaining at least one instance of each block of primary data. In examples where the data storage system stores a variable number of instances of blocks of primary data, blocks of primary data can be distributed across two or more storage devices 115, thereby adding a further aspect of redundancy.
[0098] Another advantage is that the metadata files 706/707, the metadata index files 708/709, the container files 710/711/713, and/or the container index files 712/714 could be used to replicate the data stored in the single instance database 123 or reconstruct the single instance database 123 if the data of the single instance database 123 is ever lost and/or corrupted.
[0099] The storage of data blocks in the container files may create additional complexities when it comes time to prune data blocks (pruning data blocks may be alternatively referred to as deleting or removing data blocks) that the data storage system no longer need retain. This is because the data blocks are not stored as files on the file system on the storage device 115 and thus cannot be directly referenced by the file system using the file system's data structures (the data structures that are built into or provided with the file system). As described in detail with reference to Figure 9, the secondary storage computing device 165 uses the container index files 712/714 to keep track of which blocks of data are referenced and thus which blocks are not prunable (deletable).
[00100] In some examples, the use of the container index files 712/714, the metadata index files 708/709, and/or the primary and secondary tables 600/650 to track data is analogous to a driver, agent or an additional file system that is layered on top of the existing file system of the storage device 115. This driver/agent/additional file system allows the data storage system to efficiently keep track of very large numbers of blocks of data, without regard to any limitations of the file systems of the storage devices 115. Accordingly, the data storage system can store very large numbers of blocks of data.
[00101] Accordingly, the data structures 260 illustrated in Figure 7 and the techniques described herein enable the performance of multiple storage operations cumulatively involving very large amounts of data, while still allowing for recovery of space on the storage device 115 when storage of certain data blocks is no longer required. For example, the data of numerous clients 130 can be protected without having to store redundant copies or instances of data blocks. Space on the storage device 115 can also be recovered when it is no longer necessary to store certain data blocks. Accordingly, storage operations involving very large amounts of data are enabled and optimized by the techniques described herein. Process for Restoring Data
[00102] Figure 8 is a flow diagram of a process 800 for restoring one or more blocks of data. The process 800 is described as being performed by the secondary storage computing device 165, although those of skill in the art will understand that aspects of the process 800 may be performed by any of the entities described herein. The process 800 begins at step 805 where the secondary storage computing device 165 receives a selection of data to restore (e.g., one or more files). For example, an administrator may utilize an integrated management console that provides an interface for allowing the administrator to specify one or more data blocks to be restored (e.g., by allowing the administrator to specify one or more files to be restored). As another example, a client 130 may request that a data block that had been previously copied from the client 130 be restored to the client 130. At step 810 the secondary storage computing device 165 determines an archive file and offset within the archive file corresponding to the data to be restored. The secondary storage computing device 165 may analyze the index 111 of the storage manager 105 to determine the archive file and offset.
[00103] At step 815 the secondary storage computing device 165 determines volume folders and chunk folders corresponding to the archive file and offset. The secondary storage computing device 165 may do so by analyzing the index 161 to determine the volume folders and chunk folders. The determined volume folders and chunk folders contain the requested data. At step 820 the secondary storage computing device 165 accesses an index file within the determined volume folders and chunk folders that corresponds to the data to be restored. This may be the metadata index file 708/709 when the requested data is non-SI data 244 or the container index file 712/714 when the requested data is SI data 242. At step 825 the secondary storage computing device 165 determines, from the index file, the offset within the metadata file 706/707 or the offset within the container file 710/711/13 corresponding to the requested data. At step 830 the secondary storage computing device 165 accesses the metadata file 706/707 or the container file 710/711/13 and seeks to the determined offset. At step 835 the secondary storage computing device 165 retrieves the data from the metadata file 706/707 or the container file 710/711/13. At step 840 the secondary storage computing device restores the data to a selected location (e.g., to a client 130 and/or to another location). The process 800 then concludes.
[00104] As previously noted, restoring a file may necessitate accessing multiple container files, chunk folders, and/or volume folders to obtain the data blocks that comprise the file. The secondary storage computing device 165 may thus have to obtain a first data block from a first container file and a second data block from a second container file. As another example, the secondary storage computing device 165 may thus have to obtain a first data block from a first container file within a first folder and a second data block from a second container file within a second folder. To do so, the secondary storage computing device 165 may have to access multiple index files or other data structures to locate the requisite blocks of data. Those of skill in the art will understand that various techniques may be used to restore data such as files and other data.
Process for Pruning Data
[00105] Figure 9 is a flow diagram of a process 900 for pruning data. The process 900 is described as being performed by the secondary storage computing device 165, although those of skill in the art will understand that aspects of the process 900 may be performed by any of the entities described herein. The process 900 begins when the secondary storage computing device 165 receives instructions to prune data corresponding to a storage operation (job). Additionally or alternatively, one or more files can be selected to be pruned, and/or one or more data blocks can be selected to be pruned. This selection of a job or other data to be deleted can be made manually, such as by an administrator, or automatically, such as by the job, files, and/or data blocks aging out by a retention policy.
[00106] As previously noted, the data structures 260 illustrated in Figure 7 may have been created as a result of two jobs involving two clients 130. For example, a first job on a first client 130 could result in the creation of the first chunk folder 704, and a second job on a second client 130 could result in the creation of the second chunk folder 705. The process 900 is described using this example. More specifically, the process 900 is described below as pruning the data created as a result of the first job. Of course, a similar process may be used to delete other jobs, or even smaller increments of data or data objects, such as individual files or blocks.
[00107] At step 907 the secondary storage computing device 165 determines the file, e.g., archive file, and the volume folders 702 and chunk folder 704 corresponding to the job to be pruned. The secondary storage computing device 165 may do so, for example, by analyzing the index 111 and/or the index 161 to determine this information. At step 910 the secondary storage computing device 165 deletes the metadata file 706 and the metadata index file 708 in the chunk folder 704. The secondary storage computing device 165 can delete the metadata file 706 and the metadata index file 708 in this example because these files include non-SI data 244, which is not referenced by any other data.
[00108] At step 915 the secondary storage computing device 165 accesses the container file 710 and the container index file 712 in the chunk folder 704. The secondary storage computing device 165 begins iterating through the data blocks in the container files 710. At step 920, beginning with a first block in the container file 710, the secondary storage computing device 165 accesses the primary table 600 in the single instance database 123. The secondary storage computing device 165 determines from the primary table 600 whether the reference count of a data block in the container file 710 is equal to zero. If so, this indicates that there are no references to the data block. The process 900 then continues at step 925, where the secondary storage computing device 165 sets the entry in the container index file 712 corresponding to the data block equal to zero, thus indicating that there are no references to the data block, and therefore prunable.
[00109] If the reference count of a data block is not equal to zero, then the data block is not prunable, and the process 900 continues at step 930. At this step, the secondary storage computing device 165 determines whether there are more data blocks in the container file 710. If so, the process 900 returns to step 920, where it accesses the next data block. If there are no more data blocks in the container file 710, the process 900 continues at step 932, where the secondary storage computing device 165 determines whether all the entries in the container index file 712 corresponding to the container file 710 are equal to zero. As illustrated in Figure 7, the second index entry in the container index file 712 is not equal to zero, thus indicating that the corresponding block in container file 710 is referenced (by data in the chunk folder 705, as earlier described). Accordingly, the container file 710 cannot be deleted.
[00110] However, if the container file 710 did not contain any referenced data blocks, then at step 933, the secondary storage computing device 165 would delete the container file 710. The process would then continue at step 935, where the secondary storage computing device 165 determines whether there are more container files. According to the example as illustrated in Figure 7, there is an additional container file 711. The process 900 then returns to step 915, where it performs the same steps 920- 933 for container file 711. As a result of performing these steps, the secondary storage computing device 165 would also determine that the container file 711 cannot be deleted, because it contains a data block that is referenced (by data in the chunk folder 705, as earlier described).
[00111] After processing container files 710/711 , the process 900 continues at step 940, where the secondary storage computing device 165 determines whether to free up storage space in the container files 710/711. The secondary storage computing device 165 may do so using various techniques. For example, if the operating system of the secondary storage computing device 165 supports sparse files, then the secondary storage computing device 165 may free up space by zeroing out the bytes in the container files corresponding to the space to be freed up. For a certain number of contiguous blocks (e.g., a threshold number of contiguous blocks, such as three contiguous blocks) for which the corresponding entries in the container index file 712 indicate that the blocks are not being referred to, then the secondary storage computing device 165 may mark these portions of the container files 710/711 as available for storage by the operating system or the file system. The secondary storage computing device 165 may do so by calling an API of the operating system to mark the unreferenced portions of the container files 710/711 as available for storage.
[00112] The secondary storage computing device 165 may use certain optimizations to manage the number of times portions of the container file are specified or marked as available for storage, such as only zeroing out bytes in container files when a threshold number of unreferenced contiguous blocks is reached (e.g., three or more unreferenced contiguous blocks). These optimizations may result in less overhead for the operating system because it reduces the number of contiguous ranges of zero-value bytes in the container files 710/711 that the operating system must keep track of (e.g., it reduces the amount of metadata about portions of the container files 710/711 that are available for storage).
[00113] If the operating system of the secondary storage computing device 165 does not support sparse files, then the secondary storage computing device 165 may free up space by truncating either the beginning or the end of the container files 710/711 (removing or deleting data at the beginning or end of the container files 710/711). The secondary storage computing device 165 may do so by calling an API of the operating system, or by operating directly on the container files 710/711. For example, if a certain number of the last blocks of the container file are not being referred to, the secondary storage computing device 165 may truncate these portions of the container files 710/711. Other techniques may be used to free up space in the container files 710/711 for storage of other data. At step 945 the secondary storage computing device 165 frees up space in the container files 710/711. The process 900 then concludes.
[00114] As a result of the process 900, the chunk folder 704 would contain only the container files 710/711 and the container index file 712. At a later time, when the chunk folder 705 is pruned (that is, when the job that created this chunk folder is selected to be pruned), then the container files 710/711 in the chunk folder 704 can be deleted, because they no longer contain data blocks that is referenced by other data. Therefore, pruning data corresponding to a job may also result in pruning data corresponding to an earlier job, because the data corresponding to the earlier job is no longer referenced by the later job.
[00115] Although the process 900 is described with reference to the pruning of data corresponding to jobs (one or more storage operations), other data can also be pruned. For example, an administrator may wish to delete SI data 242 but retain non-SI data 244. In such case, the administrator may instruct the secondary storage computing device 165 to delete the container files 710/711/713 but retain the metadata files 706/707 and metadata index files 708/709. As another example, an administrator or storage policy may delete one or more specific files. In such case, the secondary storage computing device 165 deletes the data blocks in the container files 710/711/713 corresponding to the specific files but retains other data blocks. The process 900 may include fewer or more steps than those described herein to accommodate these other pruning examples. Those of skill in the art will understand that data can be pruned in various fashions and therefore, that the process 900 is not limited to the steps described herein.
[00116] One advantage of the process 900 and the techniques described herein is that they enable the deletion of data on the storage devices 115 that no longer needs to be stored while still retaining data that needs to be stored, and doing so in a space- efficient manner. Space previously allocated for data blocks that no longer need to be stored can be reclaimed by the data storage system, and used to store other data. Accordingly, the techniques described herein provide for efficient use of available storage space (available on physical media).
Conclusion
[00117] From the foregoing, it will be appreciated that specific examples of data storage systems have been described herein for purposes of illustration, but that various modifications may be made without deviating from the spirit and scope of the invention. For example, although copy operations may have been described, the system may be used to perform many types of storage operations (e.g., backup operations, restore operations, archival operations, copy operations, Continuous Data Replication (CDR) operations, recovery operations, migration operations, HSM operations, etc.). As another example, although block-level single instancing has been described, the systems and methods detailed herein may be used to single instance files. As another example, the secondary storage computing device 165 may keep track of which blocks of data in container files 710 are not referenced, instead of keeping track of which blocks of data are referred to by links. As another example, non-SI data 244 may not be aligned into blocks of data. Accordingly, the invention is not limited except as by the appended claims.
[00118] Terms and phrases used in this document, and variations thereof, unless otherwise expressly stated, should be construed as open ended as opposed to limiting. As examples of the foregoing: the term "including" should be read as meaning "including, without limitation" or the like; the term "example" is used to provide exemplary instances of the item in discussion, not an exhaustive or limiting list thereof; the terms "a" or "an" should be read as meaning "at least one," "one or more" or the like; and adjectives such as "conventional," "traditional," "normal," "standard," "known" and terms of similar meaning should not be construed as limiting the item described to a given time period or to an item available as of a given time, but instead should be read to encompass conventional, traditional, normal, or standard technologies that may be available or known now or at any time in the future. Likewise, where this document refers to technologies that would be apparent or known to one of ordinary skill in the art, such technologies encompass those apparent or known to the skilled artisan now or at any time in the future.
[00119] The presence of broadening words and phrases such as "one or more," "at least," "but not limited to" or other like phrases in some instances shall not be read to mean that the narrower case is intended or required in instances where such broadening phrases may be absent. The use of the term "module" does not imply that the components or functionality described or claimed as part of the module are all configured in a common package. Indeed, any or all of the various components of a module, whether control logic or other components, can be combined in a single package or separately maintained and can further be distributed in multiple groupings or packages or across multiple locations.
[00120] If a synchronization process or synchronization processes are described herein, it is not intended to require that multiple synchronizations occur simultaneously or that multiple computing systems being synchronized each receive the same data. Although in some examples the data can be broadcast to all participating computing systems simultaneously (or close to simultaneously), in other examples the data can be sent to different computing systems or groups of computing systems at different times. Likewise, in some examples the same data, or the same subset of the data can be sent to all computing systems. However, in other examples, subsets of the data can be tailored for a given computing system or group of computing systems.
[00121] Unless the context clearly requires otherwise, throughout the description and the claims, the words "comprise," "comprising," and the like are to be construed in an inclusive sense, as opposed to an exclusive or exhaustive sense; that is to say, in the sense of "including, but not limited to." The word "coupled," as generally used herein, refers to two or more elements that may be either directly connected, or connected by way of one or more intermediate elements. Additionally, the words "herein," "above," "below," and words of similar import, when used in this application, shall refer to this application as a whole and not to any particular portions of this application. Where the context permits, words in the above Detailed Description using the singular or plural number may also include the plural or singular number respectively. The word "or" in reference to a list of two or more items, that word covers all of the following interpretations of the word: any of the items in the list, all of the items in the list, and any combination of the items in the list.
[00122] The above detailed description of embodiments of the invention is not intended to be exhaustive or to limit the invention to the precise form disclosed above. While specific embodiments of, and examples for, the invention are described above for illustrative purposes, various equivalent modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize. For example, while processes or blocks are presented in a given order, alternative embodiments may perform routines having steps, or employ systems having blocks, in a different order, and some processes or blocks may be deleted, moved, added, subdivided, combined, and/or modified. Each of these processes or blocks may be implemented in a variety of different ways. Also, while processes or blocks are at times shown as being performed in series, these processes or blocks may instead be performed in parallel, or may be performed at different times.
[00123] The teachings of the invention provided herein can be applied to other systems, not necessarily the system described above. The elements and acts of the various embodiments described above can be combined to provide further embodiments.
[00124] Any patents and applications and other references noted above, including any that may be listed in accompanying filing papers, are incorporated herein by reference. Aspects of the invention can be modified, if necessary, to employ the systems, functions, and concepts of the various references described above to provide yet further implementations of the invention.
[00125] These and other changes can be made to the invention in light of the above Detailed Description. While the above description details certain embodiments of the invention and describes the best mode contemplated, no matter how detailed the above appears in text, the invention can be practiced in many ways. Details of the system may vary considerably in implementation details, while still being encompassed by the invention disclosed herein. As noted above, particular terminology used when describing certain features or aspects of the invention should not be taken to imply that the terminology is being redefined herein to be restricted to any specific characteristics, features, or aspects of the invention with which that terminology is associated. In general, the terms used in the following claims should not be construed to limit the invention to the specific embodiments disclosed in the specification, unless the above Detailed Description section explicitly defines such terms. Accordingly, the actual scope of the invention encompasses not only the disclosed embodiments, but also all equivalent ways of practicing or implementing the invention under the claims.
[00126] While certain aspects of the invention are presented below in certain claim forms, the inventors contemplate the various aspects of the invention in any number of claim forms. For example, while only one aspect of the invention is recited as embodied in a computer-readable medium, other aspects may likewise be embodied in a computer-readable medium. As another example, while only one aspect of the invention is recited as a means-plus-function claim under 35 U. S. C. § 112, sixth paragraph, other aspects may likewise be embodied as a means-plus-function claim, or in other forms, such as being embodied in a computer-readable medium. (Any claims intended to be treated under 35 U. S. C. § 112, U 6 will begin with the words "means for.") Accordingly, the inventors reserve the right to add additional claims after filing the application to pursue such additional claim forms for other aspects of the invention.

Claims

CLAIMSWe claim:
1. A computing system for reclaiming storage space on one or more storage devices having native file systems, wherein the storage space is utilized by one or more logical containers to store deduplicated blocks of data, wherein locations of the deduplicated blocks of data in the logical containers are not tracked by the native file systems of the storage devices, the computing system comprising: one or more storage devices storing on physical media — one or more logical containers that include multiple deduplicated blocks of data that correspond to data objects; and one or more data structures that indicate whether the blocks of data are referred to; one or more databases storing information indicating whether the blocks of data are referred to; and a secondary storage computing device programmed to — receive an indication to remove a first set of blocks of data from a first logical container; for each of the blocks of data in the first set — determine, from the databases, whether the block of data is referred to; and if the block of data is not referred to, update the data structures to indicate that the block of data is not referred to; determine from the data structures that a threshold number of contiguous blocks of data in the first logical container that are not referred to has been reached; and make available for storage portions of the one or more physical media corresponding to the contiguous blocks of data in the first logical container, wherein the data structures and the databases are not part of the native file systems of the storage devices
2. The computing system of claim 1 , wherein the information stored by the databases includes, for a block of data, either a reference count of a number of references to the block of data, or information from which the reference count can be derived, and wherein the secondary storage computing device is further programmed to, for each of the blocks of data in the first set: determine the reference count for the block of data; and if the reference count of the block of data is zero, update the data structures to indicate that the block of data is not referred to.
3. The computing system of claim 1 , wherein the secondary storage computing device is further programmed to: either make available for storage an intermediate portion of the first logical container; or truncate the first logical container.
4. The computing system of claim 1 , wherein the one or more logical containers correspond to one or more container files and the one or more data structures correspond to one or more index files.
5. The computing system of claim 1 , wherein the secondary storage computing device is further programmed to: make available for storage portions of the one or more physical media corresponding to the first set of blocks of data in the first logical container; determine a first data object that is referred to by at least one of the blocks of data in the first set; and update the databases to indicate that the first data object is not referred to by any of the blocks of data in the first set.
6. The computing system of claim 1 , wherein the first set of blocks of data in the first logical container was created as a result of a first storage operation.
7. The computing system of claim 1 , wherein the first set of blocks of data corresponds to a file.
8. A method of reclaiming storage space on one or more storage devices, wherein the storage space is utilized by one or more logical containers to store deduplicated blocks of data, and wherein the method is performed by a computing system having a processor and memory, the method comprising: receiving an indication to remove a first data object, wherein the first data object is stored as multiple first blocks of data in at least a first logical container; accessing, by the computing system, a first data structure that indicates whether the first data object is referred to; determining that the first data object is not referred to; determining from a second data structure that a first number of multiple contiguous second blocks of data in the first logical container that are not referred to has been reached; and after determining that the first number has been reached, specifying as available for storage a portion of the first logical container corresponding to the multiple contiguous second blocks of data.
9. The method of claim 8, wherein the storage devices have native file systems, wherein locations of the deduplicated blocks of data in the logical containers are not tracked by the native file systems of the storage devices, and wherein the first and second data structures are not part of the native file systems of the storage devices.
10. The method of claim 8, wherein the first data structure includes, for each of the multiple first blocks of data, either a reference count of a number of references to each block of data, or information from which the reference count can be derived, and wherein the method further comprises, for each block of data of the multiple first blocks of data: determining the reference count for the block of data; and if the reference count of the block of data is zero, updating the second data structure to indicate that the block of data is not referred to.
11. The method of claim 8, wherein specifying as available for storage a portion of the first logical container corresponding to the multiple contiguous second blocks of data includes either specifying as available for storage an intermediate portion of the first logical container or truncating the first logical container.
12. The method of claim 8, further comprising storing an indication in the second data structure that the multiple first blocks of data are not referred to.
13. The method of claim 8, wherein the contiguous multiple second blocks of data include at least one of the multiple first blocks of data.
14. The method of claim 8, wherein specifying as available for storage a portion of the first logical container corresponding to the multiple contiguous second blocks of data includes removing the multiple first blocks of data from the first logical container, and wherein the method further comprises: determining a second data object that is referred to by at least one of the multiple first blocks of data; and updating the first data structure to indicate that the second data object is not referred to by any of the multiple first blocks of data.
15. The method of claim 8, wherein the first data object was created as a result of a first data storage operation.
16. The method of claim 8, wherein the first data object corresponds to a first file.
17. A computing system for reclaiming storage space on one or more means for storing, wherein the storage space is utilized by one or more logical containers to store deduplicated blocks of data, the computing system comprising: means for storing — one or more logical containers that include multiple deduplicated blocks of data that correspond to data objects; one or more first data structures that indicate whether the blocks of data are referred to by other data objects; and one or more second data structures that indicate whether the blocks of data are referred to by other data objects; means for receiving an indication to remove a first set of blocks of data from a first logical container; means for determining whether a block of data is referred to, for each of the blocks of data in the first set; means for updating the second data structures to indicate that a block of data is not referred to, for each of the blocks of data in the first set that is not referred to; means for determining from the second data structures that a first number of contiguous blocks of data in the first logical container that are not referred to has been reached; and means for specifying as available for storage portions of the one or more physical media corresponding to the contiguous blocks of data in the first logical container.
18. The computing system of claim 17, wherein the means for storing has a native file system, wherein locations of the deduplicated blocks of data in the logical containers are not tracked by the native file system of the means for storing, and wherein the first and second data structures do not form part of the native file system of the means for storing.
19. The computing system of claim 17, wherein: the one or more first data structures include either a reference count of a number of references to a given block of data, or information from which the reference count can be derived, the means for determining whether a block of data is referred to includes means for determining a reference count for the block of data, and the means for updating the second data structures updates the second data structures if the reference count for the block of data is zero.
20. The computing system of claim 17, wherein the means for specifying as available for storage portions of the one or more physical media includes: means for specifying as available for storage portions of the one or more physical media corresponding to the contiguous blocks of data located at an intermediate portion of the first logical container; and means for specifying as available for storage portions of the one or more physical media corresponding to the contiguous blocks of data located at an extremity of the first logical container.
21. The computing system of claim 17, wherein the first set of blocks of data was created as a result of a first storage operation.
22. The computer system of claim 17, wherein the first set of blocks of data corresponds to a first file.
23. A system for storing a single instance of a block of data in a data storage network, wherein the data storage network includes multiple storage devices coupled via a computer network, and wherein the computer network also couples to one or more computing devices having file systems on which files are stored, the system comprising: one or more storage devices storing multiple blocks of data in one or more container files, wherein a file stored on a file system of a computing device is comprised in one or more blocks of data; one or more single instance databases storing, for at least some of the multiple blocks of data, an identifier of the block of data and a location of the block of data in a container file; and a secondary storage computing device configured to — receive data corresponding to one or more data storage jobs from the one or more computing devices, wherein a data storage job is performed on one or more files stored on the file systems of the one or more computing devices, and wherein the received data includes — multiple blocks of data indicated to be eligible for single instancing; and multiple items of data not indicated to be eligible for single instancing; for at least some of the multiple blocks of data indicated to be eligible for single instancing — determine an identifier of the block of data; determine if the single instance database already stores the identifier; when the single instance database already stores the identifier, determine a location of the corresponding block of data in a container file and store a reference to the location; and when the single instance database does not already store the identifier, store the block of data in a container file; for at least some of the multiple items of data not indicated to be eligible for single instancing, store the items of data in a metadata file; and maintain an index file that indicates whether the multiple blocks of data in the one or more container files are referred to.
24. The system of claim 23, wherein the secondary storage computing device receives the data corresponding to the one or more data storage jobs from the one or more computing devices in one or more data streams, and wherein a data stream includes: multiple stream data items, wherein the multiple stream data items include multiple blocks of data indicated to be eligible for single instancing and multiple items of data not indicated to be eligible for single instancing; multiple stream header items, each stream header item containing metadata describing an associated stream data item, the metadata including an indication of whether the stream data item includes data indicated to be eligible for single instancing; and multiple identifiers of the multiple blocks of data indicated to be eligible for single instancing.
25. The system of claim 23, further comprising the one or more computing devices, wherein the one or more computing devices are configured to: access files on which the one or more data storage jobs are performed; determine a first set of files that are not eligible for single instancing and a second set of files that are eligible for single instancing, wherein the first set of files includes multiple items of data, and wherein the second set of files are comprised in multiple blocks of data; for at least some of the multiple blocks of data — generate the identifier for the block of data; and associate the identifier with the block of data; and provide the first and second sets of files to the secondary storage computing device.
26. The system of claim 23, wherein the one or more single instance databases either store a reference count of a number of references to locations of blocks of data or information from which the reference count can be derived, wherein the one or more storage devices store the multiple blocks of data in the one or more container files on one or more physical media, and wherein the secondary storage computing device is further configured to: receive an indication to delete one or more blocks of data stored in the one or more container files; for at least some of the blocks of data indicated to be deleted — determine the reference count of the block of data; and when the reference count of the block of data is zero, update the index file to indicate that the block of data is not referred to; and when a threshold number of contiguous blocks of data in a container file that are not referred to is reached, make available for storage portions of the one or more physical media corresponding to the threshold number of contiguous blocks of data.
27. The system of claim 23, wherein the one or more single instance databases either store a reference count of a number of references to locations of blocks of data or information from which the reference count can be derived, and wherein the secondary storage computing device is further configured to: receive an indication to delete one or more blocks of data stored in the one or more container files; for at least some of the blocks of data indicated to be deleted — determine the reference count of the block of data; and when the reference count of the block of data is zero, update the index file to indicate that the block of data is not referred to; and when a threshold number of contiguous blocks of data at an extremity of a container file that are not referred to is reached, delete a portion of the container file corresponding to the threshold number of contiguous blocks of data.
28. The system of claim 23, wherein the one or more single instance databases either store a reference count of a number of references to locations of blocks of data or information from which the reference count can be derived, and wherein the secondary storage computing device is further configured to: receive an indication to delete one or more blocks of data stored in the one or more container files; for at least some of the blocks of data indicated to be deleted — determine the reference count of the block of data; and when the reference count of the block of data is zero, update the index file to indicate that the block of data is not referred to; and when none of the blocks of data in a container file is referred to, delete the container file.
29. The system of claim 23, wherein the secondary storage computing device includes one or more memory buffers, wherein each of the memory buffers has a size that is greater than a size of a block of data but is less than ten times the size of the block of data, and wherein the secondary storage computing device is further configured to store multiple blocks of data indicated to be eligible for single instancing in the one or more memory buffers.
30. The system of claim 23, wherein the secondary storage computing device is further configured to, for at least some of the multiple blocks of data indicated to be eligible for single instancing, generate the identifier for the block of data.
31. The system of claim 23, wherein the one or more single instance databases maintain: a first data structure storing, for at least some of the blocks of data, the identifier of the block of data and the location of the block of data in a container file; and a second data structure storing, for at least some of the blocks of data, a location of a reference to the location of the block of data.
32. The system of claim 23, wherein the secondary storage computing device is further configured to: determine if a container file contains any referenced blocks of data; and when the container file does not contain any referenced blocks of data, delete the container file.
33. A method of single instancing multiple blocks of data, wherein the method is performed by a first computing device having a processor and memory, the method comprising: for at least some of multiple blocks of data indicated to be eligible for single instancing included in data received from a set of one or more computing devices distinct from the first computing device, wherein the one or more computing devices have file systems storing files that comprise one or more blocks of data — determining an identifier of the block of data; accessing, by the first computing device, a data structure that stores, for a set of blocks of data stored in one or more logical containers on one or more storage devices, an identifier of the block of data and a location of the block of data in a logical container; determining, based upon the identifier of the block of data and access to the data structure, if the block of data should be stored; when the block of data should not be stored, determining a location of an already stored instance of the block of data in a logical container and storing a reference to the location; and when the block of data should be stored, storing the block of data in a logical container; and maintaining a block reference data structure that indicates whether blocks of data in the one or more logical containers are referred to.
34. The method of claim 33, wherein the received data includes multiple items of data not indicated to be eligible for single instancing, and wherein the method further comprises, for at least some of the multiple items of data not indicated to be eligible for single instancing, storing the items of data in a metadata logical container.
35. The method of claim 33, wherein the data includes the identifiers for at least some of the multiple blocks of data indicated to be eligible for single instancing.
36. The method of claim 33, further comprising for at least some of the multiple blocks of data indicated to be eligible for single instancing: storing the block of data in a buffer; determining if the block of data should be stored; when the block of data should be stored, copying the block of data from the buffer to the logical container; and when the block of data should not be stored, making the buffer available for storage of another block of data.
37. The method of claim 33, further comprising: receiving an indication to delete one or more blocks of data stored in the one or more logical containers; for at least some of the blocks of data in the one or more logical containers — determining if the block of data is referred to; and when the block of data is not referred to, updating the block reference data structure to indicate that the block of data is not referred to; and upon reaching a threshold number of contiguous blocks of data in a logical container that are not referred to, making available storage space corresponding to the contiguous blocks of data in the logical container that are not referred to.
38. The method of claim 33, further comprising: receiving an indication to delete one or more blocks of data stored in the one or more logical containers; for at least some of the blocks of data in the one or more logical containers — determining if the block of data is referred to; and when the block of data is not referred to, updating the block reference data structure to indicate that the block of data is not referred to; and when none of the blocks of data in a logical container is referred to, deleting the logical container.
39. A computing system for single instancing blocks of data, the computing system comprising: means for determining an identifier for a block of data included in data received from a set of one or more computing devices, wherein the data includes multiple blocks of data eligible for single instancing; means for determining based on the identifier if an instance of the block of data is already stored on a storage device distinct from the set of computing devices; means for storing the block of data in a container file on the storage device if a sufficient number of instances of the block of data is not already stored on the storage device; and means for determining whether the block of data in the container file is referred to.
40. The computing system of claim 39, further comprising: means for storing the identifier and a location of the block of data in the container file; and means for storing a location of a reference to the block of data.
41. The computing system of claim 39, further comprising means for deleting blocks of data from the container file when the blocks of data are not referred to.
42. A computer-readable storage medium whose contents cause a computing system to perform a method of restoring a file, the method comprising: receiving a request for a file stored on a storage device, wherein the file is stored as one or more data blocks on the storage device; determining a first file and a first offset within the first file corresponding to the requested file; accessing the first file at the first offset; determining if a first data block beginning at the first offset includes at least a first portion of the requested file; when the first data block beginning at the first offset includes at least the first portion of the requested file, obtaining the first data block from the first file; when a second data block that includes at least a second portion of the requested file exists, obtaining the second data block; and providing the requested file to a computing device.
43. The computer-readable storage medium of claim 42, wherein the method further comprises: determining if data beginning at the first offset includes a reference to a third data block in a second file that includes at least a third portion of the requested file; and when the data beginning at the first offset includes the reference to the third data block in the second container file — accessing the second container file; obtaining the third data block from the second file; and when a fourth data block that includes at least a fourth portion of the requested file exists, obtaining the fourth data block.
44. The computer-readable storage medium of claim 42, wherein the second data block is included in a second file, and wherein the method further comprises determining a second offset within the second file at which the second data block is located.
45. The computer-readable storage medium of claim 42, wherein the first file is located within a first folder, the second data block is included in a second file located within a second folder, and wherein the method further comprises determining a second offset within the second file at which the second data block is located.
46. The computer-readable storage medium of claim 42, wherein the method further comprises: determining a logical container corresponding to the requested file; and analyzing the logical container to determine the first file.
47. The computer-readable storage medium of claim 42, wherein the method further comprises: accessing an index file associated with the first file; and analyzing the index file to determine the first offset within the first file.
PCT/US2010/034676 2009-05-22 2010-05-13 Block-level single instancing WO2010135136A2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US18079109P 2009-05-22 2009-05-22
US61/180,791 2009-05-22
US12/647,906 2009-12-28
US12/647,906 US8578120B2 (en) 2009-05-22 2009-12-28 Block-level single instancing

Publications (2)

Publication Number Publication Date
WO2010135136A2 true WO2010135136A2 (en) 2010-11-25
WO2010135136A3 WO2010135136A3 (en) 2011-02-17

Family

ID=43125335

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2010/034676 WO2010135136A2 (en) 2009-05-22 2010-05-13 Block-level single instancing

Country Status (2)

Country Link
US (7) US8578120B2 (en)
WO (1) WO2010135136A2 (en)

Families Citing this family (330)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7434219B2 (en) 2000-01-31 2008-10-07 Commvault Systems, Inc. Storage of application specific profiles correlating to document versions
JP2005505045A (en) 2001-09-28 2005-02-17 コムヴォールト・システムズ・インコーポレーテッド Method and apparatus for creating and managing a quick recovery volume
WO2003027891A1 (en) 2001-09-28 2003-04-03 Commvault Systems, Inc. System and method for archiving objects in an information store
CA2508089A1 (en) 2002-10-07 2004-04-22 Commvault Systems, Inc. System and method for managing stored data
WO2005050385A2 (en) 2003-11-13 2005-06-02 Commvault Systems, Inc. System and method for performing integrated storage operations
CA2564967C (en) 2004-04-30 2014-09-30 Commvault Systems, Inc. Hierarchical systems and methods for providing a unified view of storage information
US8775823B2 (en) 2006-12-29 2014-07-08 Commvault Systems, Inc. System and method for encrypting secondary copies of data
US8959299B2 (en) 2004-11-15 2015-02-17 Commvault Systems, Inc. Using a snapshot as a data source
US7747579B2 (en) 2005-11-28 2010-06-29 Commvault Systems, Inc. Metabase for facilitating data classification
US20110010518A1 (en) 2005-12-19 2011-01-13 Srinivas Kavuri Systems and Methods for Migrating Components in a Hierarchical Storage Network
US8572330B2 (en) 2005-12-19 2013-10-29 Commvault Systems, Inc. Systems and methods for granular resource management in a storage network
US20200257596A1 (en) 2005-12-19 2020-08-13 Commvault Systems, Inc. Systems and methods of unified reconstruction in storage systems
US8930496B2 (en) 2005-12-19 2015-01-06 Commvault Systems, Inc. Systems and methods of unified reconstruction in storage systems
US7882077B2 (en) 2006-10-17 2011-02-01 Commvault Systems, Inc. Method and system for offline indexing of content and classifying stored data
US8370442B2 (en) 2008-08-29 2013-02-05 Commvault Systems, Inc. Method and system for leveraging identified changes to a mail server
CA2705379C (en) 2006-12-04 2016-08-30 Commvault Systems, Inc. Systems and methods for creating copies of data, such as archive copies
US8677091B2 (en) 2006-12-18 2014-03-18 Commvault Systems, Inc. Writing data and storage system specific metadata to network attached storage device
US20080228771A1 (en) 2006-12-22 2008-09-18 Commvault Systems, Inc. Method and system for searching stored data
US7840537B2 (en) 2006-12-22 2010-11-23 Commvault Systems, Inc. System and method for storing redundant information
EP2195724B1 (en) 2007-08-28 2019-10-09 Commvault Systems, Inc. Power management of data processing resources, such as power adaptive management of data storage operations
US8769048B2 (en) 2008-06-18 2014-07-01 Commvault Systems, Inc. Data protection scheduling, such as providing a flexible backup window in a data protection system
US8352954B2 (en) 2008-06-19 2013-01-08 Commvault Systems, Inc. Data storage resource allocation by employing dynamic methods and blacklisting resource request pools
US9128883B2 (en) 2008-06-19 2015-09-08 Commvault Systems, Inc Data storage resource allocation by performing abbreviated resource checks based on relative chances of failure of the data storage resources to determine whether data storage requests would fail
US8484162B2 (en) 2008-06-24 2013-07-09 Commvault Systems, Inc. De-duplication systems and methods for application-specific data
US9098495B2 (en) 2008-06-24 2015-08-04 Commvault Systems, Inc. Application-aware and remote single instance data management
US8166263B2 (en) 2008-07-03 2012-04-24 Commvault Systems, Inc. Continuous data protection over intermittent connections, such as continuous data backup for laptops or wireless devices
US8307177B2 (en) 2008-09-05 2012-11-06 Commvault Systems, Inc. Systems and methods for management of virtualization data
US8725688B2 (en) 2008-09-05 2014-05-13 Commvault Systems, Inc. Image level copy or restore, such as image level restore without knowledge of data object metadata
US20100070474A1 (en) 2008-09-12 2010-03-18 Lad Kamleshkumar K Transferring or migrating portions of data objects, such as block-level data migration or chunk-based data migration
US20100070466A1 (en) 2008-09-15 2010-03-18 Anand Prahlad Data transfer techniques within data storage devices, such as network attached storage performing data migration
WO2010036754A1 (en) 2008-09-26 2010-04-01 Commvault Systems, Inc. Systems and methods for managing single instancing data
US9015181B2 (en) 2008-09-26 2015-04-21 Commvault Systems, Inc. Systems and methods for managing single instancing data
US9178842B2 (en) 2008-11-05 2015-11-03 Commvault Systems, Inc. Systems and methods for monitoring messaging applications for compliance with a policy
US8412677B2 (en) 2008-11-26 2013-04-02 Commvault Systems, Inc. Systems and methods for byte-level or quasi byte-level single instancing
US8401996B2 (en) 2009-03-30 2013-03-19 Commvault Systems, Inc. Storing a variable number of instances of data objects
US8578120B2 (en) 2009-05-22 2013-11-05 Commvault Systems, Inc. Block-level single instancing
US20100333116A1 (en) 2009-06-30 2010-12-30 Anand Prahlad Cloud gateway system for managing data storage to cloud storage sites
US8930306B1 (en) 2009-07-08 2015-01-06 Commvault Systems, Inc. Synchronized data deduplication
US8204867B2 (en) * 2009-07-29 2012-06-19 International Business Machines Corporation Apparatus, system, and method for enhanced block-level deduplication
US8719767B2 (en) 2011-03-31 2014-05-06 Commvault Systems, Inc. Utilizing snapshots to provide builds to developer computing devices
US9092500B2 (en) 2009-09-03 2015-07-28 Commvault Systems, Inc. Utilizing snapshots for access to databases and other applications
US8433682B2 (en) 2009-12-31 2013-04-30 Commvault Systems, Inc. Systems and methods for analyzing snapshots
US8595191B2 (en) 2009-12-31 2013-11-26 Commvault Systems, Inc. Systems and methods for performing data management operations using snapshots
US8442983B2 (en) 2009-12-31 2013-05-14 Commvault Systems, Inc. Asynchronous methods of data classification using change journals and other data structures
US9047351B2 (en) * 2010-04-12 2015-06-02 Sandisk Enterprise Ip Llc Cluster of processing nodes with distributed global flash memory using commodity server technology
US8856593B2 (en) 2010-04-12 2014-10-07 Sandisk Enterprise Ip Llc Failure recovery using consensus replication in a distributed flash memory system
US9164554B2 (en) 2010-04-12 2015-10-20 Sandisk Enterprise Ip Llc Non-volatile solid-state storage system supporting high bandwidth and random access
US8868487B2 (en) 2010-04-12 2014-10-21 Sandisk Enterprise Ip Llc Event processing in a flash memory-based object store
US8504526B2 (en) 2010-06-04 2013-08-06 Commvault Systems, Inc. Failover systems and methods for performing backup operations
US11449394B2 (en) 2010-06-04 2022-09-20 Commvault Systems, Inc. Failover systems and methods for performing backup operations, including heterogeneous indexing and load balancing of backup and indexing resources
US8666939B2 (en) 2010-06-28 2014-03-04 Sandisk Enterprise Ip Llc Approaches for the replication of write sets
US8572340B2 (en) 2010-09-30 2013-10-29 Commvault Systems, Inc. Systems and methods for retaining and using data block signatures in data protection operations
US8364652B2 (en) * 2010-09-30 2013-01-29 Commvault Systems, Inc. Content aligned block-based deduplication
US8620870B2 (en) 2010-09-30 2013-12-31 Commvault Systems, Inc. Efficient data management improvements, such as docking limited-feature data management modules to a full-featured data management system
US9244779B2 (en) 2010-09-30 2016-01-26 Commvault Systems, Inc. Data recovery operations, such as recovery from modified network data management protocol data
WO2012045023A2 (en) 2010-09-30 2012-04-05 Commvault Systems, Inc. Archiving data objects using secondary copies
US8954446B2 (en) 2010-12-14 2015-02-10 Comm Vault Systems, Inc. Client-side repository in a networked deduplicated storage system
US9020900B2 (en) 2010-12-14 2015-04-28 Commvault Systems, Inc. Distributed deduplicated storage system
US8719264B2 (en) 2011-03-31 2014-05-06 Commvault Systems, Inc. Creating secondary copies of data based on searches for content
US8849762B2 (en) 2011-03-31 2014-09-30 Commvault Systems, Inc. Restoring computing environments, such as autorecovery of file systems at certain points in time
US8874515B2 (en) 2011-04-11 2014-10-28 Sandisk Enterprise Ip Llc Low level object version tracking using non-volatile memory write generations
US9323466B2 (en) 2011-04-27 2016-04-26 Commvault Systems, Inc. System and method for client policy assignment in a data storage system
US8656253B2 (en) * 2011-06-06 2014-02-18 Cleversafe, Inc. Storing portions of data in a dispersed storage network
US8904128B2 (en) * 2011-06-08 2014-12-02 Hewlett-Packard Development Company, L.P. Processing a request to restore deduplicated data
US9069477B1 (en) * 2011-06-16 2015-06-30 Amazon Technologies, Inc. Reuse of dynamically allocated memory
US9116633B2 (en) 2011-09-30 2015-08-25 Commvault Systems, Inc. Information management of virtual machines having mapped storage devices
US9461881B2 (en) 2011-09-30 2016-10-04 Commvault Systems, Inc. Migration of existing computing systems to cloud computing sites or virtual machines
US9749132B1 (en) * 2011-11-28 2017-08-29 Amazon Technologies, Inc. System and method for secure deletion of data
US20130227352A1 (en) 2012-02-24 2013-08-29 Commvault Systems, Inc. Log monitoring
US9135064B2 (en) 2012-03-07 2015-09-15 Sandisk Enterprise Ip Llc Fine grained adaptive throttling of background processes
US8868505B1 (en) * 2012-03-20 2014-10-21 Emc Corporation Systems and methods for protecting data in a network host environment
US9292815B2 (en) 2012-03-23 2016-03-22 Commvault Systems, Inc. Automation of data storage activities
US8950009B2 (en) 2012-03-30 2015-02-03 Commvault Systems, Inc. Information management of data associated with multiple cloud services
AU2013202553B2 (en) 2012-03-30 2015-10-01 Commvault Systems, Inc. Information management of mobile device data
US9020890B2 (en) 2012-03-30 2015-04-28 Commvault Systems, Inc. Smart archiving and data previewing for mobile devices
US10157184B2 (en) 2012-03-30 2018-12-18 Commvault Systems, Inc. Data previewing before recalling large data files
US9286327B2 (en) 2012-03-30 2016-03-15 Commvault Systems, Inc. Data storage recovery automation
US9262496B2 (en) 2012-03-30 2016-02-16 Commvault Systems, Inc. Unified access to personal data
US9342537B2 (en) 2012-04-23 2016-05-17 Commvault Systems, Inc. Integrated snapshot interface for a data storage system
US9021203B2 (en) 2012-05-07 2015-04-28 International Business Machines Corporation Enhancing tiering storage performance
US9645944B2 (en) 2012-05-07 2017-05-09 International Business Machines Corporation Enhancing data caching performance
US9110815B2 (en) * 2012-05-07 2015-08-18 International Business Machines Corporation Enhancing data processing performance by cache management of fingerprint index
US9189167B2 (en) 2012-05-31 2015-11-17 Commvault Systems, Inc. Shared library in a data storage system
US8977672B2 (en) 2012-06-08 2015-03-10 Commvault Systems, Inc. Intelligent scheduling for remote computers
US8892523B2 (en) 2012-06-08 2014-11-18 Commvault Systems, Inc. Auto summarization of content
US9189170B2 (en) 2012-06-12 2015-11-17 Commvault Systems, Inc. External storage manager for a data storage cell
US20130339310A1 (en) 2012-06-13 2013-12-19 Commvault Systems, Inc. Restore using a client side signature repository in a networked storage system
US20140025796A1 (en) 2012-07-19 2014-01-23 Commvault Systems, Inc. Automated grouping of computing devices in a networked data storage system
US9275086B2 (en) 2012-07-20 2016-03-01 Commvault Systems, Inc. Systems and methods for database archiving
US20140040580A1 (en) 2012-07-31 2014-02-06 Commvault Systems, Inc. Administering a shared, on-line pool of data storage resources for performing data storage operations
US8938481B2 (en) 2012-08-13 2015-01-20 Commvault Systems, Inc. Generic file level restore from a block-level secondary copy
US9026498B2 (en) 2012-08-13 2015-05-05 Commvault Systems, Inc. Lightweight mounting of a secondary copy of file system data
WO2014054065A1 (en) * 2012-10-01 2014-04-10 Hitachi, Ltd. Backup and restore system for a deduplicated file system and corresponding server and method
US9201906B2 (en) 2012-12-21 2015-12-01 Commvault Systems, Inc. Systems and methods to perform data backup in data storage systems
US20140181038A1 (en) 2012-12-21 2014-06-26 Commvault Systems, Inc. Systems and methods to categorize unprotected virtual machines
US20140181047A1 (en) 2012-12-21 2014-06-26 Commvault Systems, Inc. Systems and methods to protect deleted files
US9223597B2 (en) 2012-12-21 2015-12-29 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US9747169B2 (en) 2012-12-21 2017-08-29 Commvault Systems, Inc. Reporting using data obtained during backup of primary storage
US9633216B2 (en) 2012-12-27 2017-04-25 Commvault Systems, Inc. Application of information management policies based on operation with a geographic entity
US9069799B2 (en) 2012-12-27 2015-06-30 Commvault Systems, Inc. Restoration of centralized data storage manager, such as data storage manager in a hierarchical data storage system
US9378035B2 (en) 2012-12-28 2016-06-28 Commvault Systems, Inc. Systems and methods for repurposing virtual machines
EP3425493A1 (en) * 2012-12-28 2019-01-09 Huawei Technologies Co., Ltd. Data processing method and apparatus
US10346259B2 (en) 2012-12-28 2019-07-09 Commvault Systems, Inc. Data recovery using a cloud-based remote data recovery center
US9633022B2 (en) 2012-12-28 2017-04-25 Commvault Systems, Inc. Backup and restoration for a deduplicated file system
US20140196039A1 (en) 2013-01-08 2014-07-10 Commvault Systems, Inc. Virtual machine categorization system and method
US9760444B2 (en) 2013-01-11 2017-09-12 Commvault Systems, Inc. Sharing of secondary storage data
US9804930B2 (en) 2013-01-11 2017-10-31 Commvault Systems, Inc. Partial file restore in a data storage system
US20140201151A1 (en) 2013-01-11 2014-07-17 Commvault Systems, Inc. Systems and methods to select files for restoration from block-level backup for virtual machines
US9766987B2 (en) 2013-01-11 2017-09-19 Commvault Systems, Inc. Table level database restore in a data storage system
US9633033B2 (en) 2013-01-11 2017-04-25 Commvault Systems, Inc. High availability distributed deduplicated storage system
US9886346B2 (en) 2013-01-11 2018-02-06 Commvault Systems, Inc. Single snapshot for multiple agents
US9336226B2 (en) 2013-01-11 2016-05-10 Commvault Systems, Inc. Criteria-based data synchronization management
US9286110B2 (en) 2013-01-14 2016-03-15 Commvault Systems, Inc. Seamless virtual machine recall in a data storage system
US9483489B2 (en) 2013-01-14 2016-11-01 Commvault Systems, Inc. Partial sharing of secondary storage files in a data storage system
US9645950B2 (en) * 2013-01-31 2017-05-09 Vmware, Inc. Low-cost backup and edge caching using unused disk blocks
US10275397B2 (en) * 2013-02-22 2019-04-30 Veritas Technologies Llc Deduplication storage system with efficient reference updating and space reclamation
US9459968B2 (en) 2013-03-11 2016-10-04 Commvault Systems, Inc. Single index to query multiple backup formats
US9483655B2 (en) 2013-03-12 2016-11-01 Commvault Systems, Inc. File backup with selective encryption
US9239844B2 (en) * 2013-03-29 2016-01-19 Hitachi, Ltd. Data duplication detection system and method for controlling data duplication detection system
US9405635B2 (en) 2013-04-16 2016-08-02 Commvault Systems, Inc. Multi-source restore in an information management system
US9483363B2 (en) 2013-05-08 2016-11-01 Commvault Systems, Inc. Use of temporary secondary copies in failover operations
US9483558B2 (en) 2013-05-29 2016-11-01 Commvault Systems, Inc. Assessing user performance in a community of users of data storage resources
US20150074536A1 (en) 2013-09-12 2015-03-12 Commvault Systems, Inc. File manager integration with virtualization in an information management system, including user control and storage management of virtual machines
CN103605479B (en) * 2013-10-16 2016-06-01 北京奇虎科技有限公司 Data file wiring method and system, data file read method and system
US9590886B2 (en) 2013-11-01 2017-03-07 Commvault Systems, Inc. Systems and methods for differential health checking of an information management system
US9495251B2 (en) 2014-01-24 2016-11-15 Commvault Systems, Inc. Snapshot readiness checking and reporting
US9639426B2 (en) 2014-01-24 2017-05-02 Commvault Systems, Inc. Single snapshot for multiple applications
US10324897B2 (en) 2014-01-27 2019-06-18 Commvault Systems, Inc. Techniques for serving archived electronic mail
US10169121B2 (en) 2014-02-27 2019-01-01 Commvault Systems, Inc. Work flow management for an information management system
US9648100B2 (en) * 2014-03-05 2017-05-09 Commvault Systems, Inc. Cross-system storage management for transferring data across autonomous information management systems
US9633026B2 (en) 2014-03-13 2017-04-25 Commvault Systems, Inc. Systems and methods for protecting email data
US10380072B2 (en) 2014-03-17 2019-08-13 Commvault Systems, Inc. Managing deletions from a deduplication database
US9633056B2 (en) 2014-03-17 2017-04-25 Commvault Systems, Inc. Maintaining a deduplication database
US9668298B1 (en) * 2014-03-27 2017-05-30 Amazon Technologies, Inc. Network-enabled user device charger
US9811427B2 (en) 2014-04-02 2017-11-07 Commvault Systems, Inc. Information management by a media agent in the absence of communications with a storage manager
US9823978B2 (en) 2014-04-16 2017-11-21 Commvault Systems, Inc. User-level quota management of data objects stored in information management systems
US9740574B2 (en) 2014-05-09 2017-08-22 Commvault Systems, Inc. Load balancing across multiple data paths
US20160004605A1 (en) 2014-07-01 2016-01-07 Commvault Systems, Inc. Lightweight data reconstruction based on backup data
US20160019317A1 (en) 2014-07-16 2016-01-21 Commvault Systems, Inc. Volume or virtual machine level backup and generating placeholders for virtual machine files
US20160019224A1 (en) 2014-07-18 2016-01-21 Commvault Systems, Inc. File system content archiving based on third-party application archiving rules and metadata
US10031917B2 (en) 2014-07-29 2018-07-24 Commvault Systems, Inc. Efficient volume-level replication of data via snapshots in an information management system
US9641388B2 (en) 2014-07-29 2017-05-02 Commvault Systems, Inc. Customized deployment in information management systems
US11249858B2 (en) * 2014-08-06 2022-02-15 Commvault Systems, Inc. Point-in-time backups of a production application made accessible over fibre channel and/or ISCSI as data sources to a remote application by representing the backups as pseudo-disks operating apart from the production application and its host
US10360110B2 (en) 2014-08-06 2019-07-23 Commvault Systems, Inc. Point-in-time backups of a production application made accessible over fibre channel and/or iSCSI as data sources to a remote application by representing the backups as pseudo-disks operating apart from the production application and its host
US9852026B2 (en) 2014-08-06 2017-12-26 Commvault Systems, Inc. Efficient application recovery in an information management system based on a pseudo-storage-device driver
US10042716B2 (en) 2014-09-03 2018-08-07 Commvault Systems, Inc. Consolidated processing of storage-array commands using a forwarder media agent in conjunction with a snapshot-control media agent
US9774672B2 (en) 2014-09-03 2017-09-26 Commvault Systems, Inc. Consolidated processing of storage-array commands by a snapshot-control media agent
US9405928B2 (en) 2014-09-17 2016-08-02 Commvault Systems, Inc. Deriving encryption rules based on file content
US9417968B2 (en) 2014-09-22 2016-08-16 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US9710465B2 (en) 2014-09-22 2017-07-18 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US9436555B2 (en) 2014-09-22 2016-09-06 Commvault Systems, Inc. Efficient live-mount of a backed up virtual machine in a storage management system
US10204010B2 (en) 2014-10-03 2019-02-12 Commvault Systems, Inc. Intelligent protection of off-line mail data
US9444811B2 (en) 2014-10-21 2016-09-13 Commvault Systems, Inc. Using an enhanced data agent to restore backed up data across autonomous storage management systems
US9575673B2 (en) 2014-10-29 2017-02-21 Commvault Systems, Inc. Accessing a file system using tiered deduplication
US10776209B2 (en) 2014-11-10 2020-09-15 Commvault Systems, Inc. Cross-platform virtual machine backup and replication
US9848046B2 (en) 2014-11-13 2017-12-19 Commvault Systems, Inc. Archiving applications in information management systems
US9648105B2 (en) 2014-11-14 2017-05-09 Commvault Systems, Inc. Unified snapshot storage management, using an enhanced storage manager and enhanced media agents
US9448731B2 (en) 2014-11-14 2016-09-20 Commvault Systems, Inc. Unified snapshot storage management
US9912625B2 (en) 2014-11-18 2018-03-06 Commvault Systems, Inc. Storage and management of mail attachments
US9983936B2 (en) 2014-11-20 2018-05-29 Commvault Systems, Inc. Virtual machine change block tracking
US20160162364A1 (en) 2014-12-03 2016-06-09 Commvault Systems, Inc. Secondary storage pruning
US9632713B2 (en) 2014-12-03 2017-04-25 Commvault Systems, Inc. Secondary storage editor
US9645891B2 (en) 2014-12-04 2017-05-09 Commvault Systems, Inc. Opportunistic execution of secondary copy operations
US9753816B2 (en) 2014-12-05 2017-09-05 Commvault Systems, Inc. Synchronization based on filtered browsing
US20160210306A1 (en) 2015-01-15 2016-07-21 Commvault Systems, Inc. Managing structured data in a data storage system
US20160210044A1 (en) 2015-01-15 2016-07-21 Commvault Systems, Inc. Intelligent hybrid drive caching
US9952934B2 (en) * 2015-01-20 2018-04-24 Commvault Systems, Inc. Synchronizing selected portions of data in a storage management system
US9588849B2 (en) 2015-01-20 2017-03-07 Commvault Systems, Inc. Synchronizing selected portions of data in a storage management system
US10108687B2 (en) 2015-01-21 2018-10-23 Commvault Systems, Inc. Database protection using block-level mapping
US9575804B2 (en) 2015-03-27 2017-02-21 Commvault Systems, Inc. Job management and resource allocation
US9928144B2 (en) 2015-03-30 2018-03-27 Commvault Systems, Inc. Storage management of data using an open-archive architecture, including streamlined access to primary data originally stored on network-attached storage and archived to secondary storage
US9934265B2 (en) 2015-04-09 2018-04-03 Commvault Systems, Inc. Management of log data
US10339106B2 (en) 2015-04-09 2019-07-02 Commvault Systems, Inc. Highly reusable deduplication database after disaster recovery
US10311150B2 (en) 2015-04-10 2019-06-04 Commvault Systems, Inc. Using a Unix-based file system to manage and serve clones to windows-based computing clients
US9904598B2 (en) 2015-04-21 2018-02-27 Commvault Systems, Inc. Content-independent and database management system-independent synthetic full backup of a database based on snapshot technology
US9639286B2 (en) 2015-05-14 2017-05-02 Commvault Systems, Inc. Restore of secondary data using thread pooling
US10324914B2 (en) 2015-05-20 2019-06-18 Commvalut Systems, Inc. Handling user queries against production and archive storage systems, such as for enterprise customers having large and/or numerous files
US20160350391A1 (en) 2015-05-26 2016-12-01 Commvault Systems, Inc. Replication using deduplicated secondary copy data
US10769102B2 (en) 2015-06-12 2020-09-08 Hewlett Packard Enterprise Development Lp Disk storage allocation
US9563514B2 (en) 2015-06-19 2017-02-07 Commvault Systems, Inc. Assignment of proxies for virtual-machine secondary copy operations including streaming backup jobs
US10084873B2 (en) 2015-06-19 2018-09-25 Commvault Systems, Inc. Assignment of data agent proxies for executing virtual-machine secondary copy operations including streaming backup jobs
US10275320B2 (en) 2015-06-26 2019-04-30 Commvault Systems, Inc. Incrementally accumulating in-process performance data and hierarchical reporting thereof for a data stream in a secondary copy operation
US9766825B2 (en) 2015-07-22 2017-09-19 Commvault Systems, Inc. Browse and restore for block-level backups
US10192065B2 (en) 2015-08-31 2019-01-29 Commvault Systems, Inc. Automated intelligent provisioning of data storage resources in response to user requests in a data storage management system
US10311042B1 (en) 2015-08-31 2019-06-04 Commvault Systems, Inc. Organically managing primary and secondary storage of a data object based on expiry timeframe supplied by a user of the data object
US10101913B2 (en) 2015-09-02 2018-10-16 Commvault Systems, Inc. Migrating data to disk without interrupting running backup operations
EP3350728B1 (en) 2015-09-18 2022-04-06 Commvault Systems, Inc. Data storage management operations in a secondary storage subsystem using image recognition and image-based criteria
US10459801B2 (en) 2015-09-30 2019-10-29 Commvault Systems, Inc. Dynamic triggering of block-level backups based on block change thresholds and corresponding file identities using indexing in a data storage management system
US10248494B2 (en) 2015-10-29 2019-04-02 Commvault Systems, Inc. Monitoring, diagnosing, and repairing a management database in a data storage management system
US10353994B2 (en) 2015-11-03 2019-07-16 Commvault Systems, Inc. Summarization of email on a client computing device based on content contribution to an email thread using classification and word frequency considerations
US10228962B2 (en) 2015-12-09 2019-03-12 Commvault Systems, Inc. Live synchronization and management of virtual machines across computing and virtualization platforms and using live synchronization to support disaster recovery
US10387266B2 (en) 2015-12-23 2019-08-20 Commvault Systems, Inc. Application-level live synchronization across computing platforms including synchronizing co-resident applications to disparate standby destinations and selectively synchronizing some applications and not others
US10592357B2 (en) 2015-12-30 2020-03-17 Commvault Systems, Inc. Distributed file system in a distributed deduplication data storage system
US10684924B2 (en) 2016-02-18 2020-06-16 Commvault Systems, Inc. Data restoration operations based on network path information
US10565067B2 (en) 2016-03-09 2020-02-18 Commvault Systems, Inc. Virtual server cloud file system for virtual machine backup from cloud operations
US10452490B2 (en) 2016-03-09 2019-10-22 Commvault Systems, Inc. Data management and backup of distributed storage environment
US10296368B2 (en) 2016-03-09 2019-05-21 Commvault Systems, Inc. Hypervisor-independent block-level live browse for access to backed up virtual machine (VM) data and hypervisor-free file-level recovery (block-level pseudo-mount)
US10503753B2 (en) 2016-03-10 2019-12-10 Commvault Systems, Inc. Snapshot replication operations based on incremental block change tracking
US10248174B2 (en) 2016-05-24 2019-04-02 Hedvig, Inc. Persistent reservations for virtual disk using multiple targets
US10664447B2 (en) 2016-07-12 2020-05-26 Commvault Systems, Inc. Dynamic management of expandable cache storage for multiple network shares configured in a file server
US10417182B2 (en) * 2016-08-19 2019-09-17 Charles Nicholls System, method and apparatus for data management with rules on containers for collections of data
US10747630B2 (en) 2016-09-30 2020-08-18 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, including operations by a master monitor node
US10540516B2 (en) 2016-10-13 2020-01-21 Commvault Systems, Inc. Data protection within an unsecured storage environment
US10162528B2 (en) 2016-10-25 2018-12-25 Commvault Systems, Inc. Targeted snapshot based on virtual machine location
US10152251B2 (en) 2016-10-25 2018-12-11 Commvault Systems, Inc. Targeted backup of virtual machine
US10210048B2 (en) 2016-10-25 2019-02-19 Commvault Systems, Inc. Selective snapshot and backup copy operations for individual virtual machines in a shared storage
US10922189B2 (en) 2016-11-02 2021-02-16 Commvault Systems, Inc. Historical network data-based scanning thread generation
US10389810B2 (en) 2016-11-02 2019-08-20 Commvault Systems, Inc. Multi-threaded scanning of distributed file systems
US11106632B2 (en) 2016-11-16 2021-08-31 Commvault Systems, Inc. Dynamically configuring a proxy server using containerization for concurrent and/or overlapping backup, restore, and/or test operations
US10678758B2 (en) 2016-11-21 2020-06-09 Commvault Systems, Inc. Cross-platform virtual machine data and memory backup and replication
US10157102B2 (en) * 2016-12-29 2018-12-18 Whatsapp Inc. Techniques to scan and reorganize media files to remove gaps
US10838821B2 (en) 2017-02-08 2020-11-17 Commvault Systems, Inc. Migrating content and metadata from a backup system
US10740193B2 (en) 2017-02-27 2020-08-11 Commvault Systems, Inc. Hypervisor-independent reference copies of virtual machine payload data based on block-level pseudo-mount
US10459666B2 (en) 2017-03-03 2019-10-29 Commvault Systems, Inc. Using storage managers in respective data storage management systems for license distribution, compliance, and updates
US10949308B2 (en) 2017-03-15 2021-03-16 Commvault Systems, Inc. Application aware backup of virtual machines
US20180276085A1 (en) 2017-03-24 2018-09-27 Commvault Systems, Inc. Virtual machine recovery point generation
US10891069B2 (en) 2017-03-27 2021-01-12 Commvault Systems, Inc. Creating local copies of data stored in online data repositories
US11108858B2 (en) 2017-03-28 2021-08-31 Commvault Systems, Inc. Archiving mail servers via a simple mail transfer protocol (SMTP) server
US10776329B2 (en) 2017-03-28 2020-09-15 Commvault Systems, Inc. Migration of a database management system to cloud storage
US10613939B2 (en) 2017-03-28 2020-04-07 Commvault Systems, Inc. Backup index generation process
US10599527B2 (en) 2017-03-29 2020-03-24 Commvault Systems, Inc. Information management cell health monitoring system
US11074140B2 (en) 2017-03-29 2021-07-27 Commvault Systems, Inc. Live browsing of granular mailbox data
US11074138B2 (en) 2017-03-29 2021-07-27 Commvault Systems, Inc. Multi-streaming backup operations for mailboxes
US10949398B2 (en) 2017-03-29 2021-03-16 Commvault Systems, Inc. Synchronization operations for network-accessible folders
US10387073B2 (en) 2017-03-29 2019-08-20 Commvault Systems, Inc. External dynamic virtual machine synchronization
US10552294B2 (en) 2017-03-31 2020-02-04 Commvault Systems, Inc. Management of internet of things devices
US10853195B2 (en) 2017-03-31 2020-12-01 Commvault Systems, Inc. Granular restoration of virtual machine application data
US11221939B2 (en) 2017-03-31 2022-01-11 Commvault Systems, Inc. Managing data from internet of things devices in a vehicle
US11294786B2 (en) 2017-03-31 2022-04-05 Commvault Systems, Inc. Management of internet of things devices
US10984041B2 (en) 2017-05-11 2021-04-20 Commvault Systems, Inc. Natural language processing integrated with database and data storage management
US10664352B2 (en) 2017-06-14 2020-05-26 Commvault Systems, Inc. Live browsing of backed up data residing on cloned disks
US10664442B1 (en) * 2017-07-14 2020-05-26 EMC IP Holding Company LLC Method and system for data consistency verification in a storage system
US11086834B2 (en) 2017-09-14 2021-08-10 Commvault Systems, Inc. Distributed framework for data proximity-based task splitting in a content indexing system
US10846266B2 (en) 2017-09-14 2020-11-24 Commvault Systems, Inc. Distributed architecture for content indexing emails
US11263088B2 (en) 2017-09-14 2022-03-01 Commvault Systems, Inc. Distributed architecture for tracking content indexing
US11036592B2 (en) 2017-09-14 2021-06-15 Commvault Systems, Inc. Distributed content indexing architecture with separately stored file previews
US10846180B2 (en) 2017-09-14 2020-11-24 Commvault Systems, Inc. Distributed framework for task splitting and task assignments in a content indexing system
US10635334B1 (en) * 2017-09-28 2020-04-28 EMC IP Holding Company LLC Rule based data transfer model to cloud
US10942779B1 (en) 2017-10-27 2021-03-09 EMC IP Holding Company LLC Method and system for compliance map engine
US10754368B1 (en) 2017-10-27 2020-08-25 EMC IP Holding Company LLC Method and system for load balancing backup resources
US10740300B1 (en) 2017-12-07 2020-08-11 Commvault Systems, Inc. Synchronization of metadata in a distributed storage system
US10635546B2 (en) 2017-12-07 2020-04-28 Commvault Systems, Inc. Synthesizing a restore image from one or more secondary copies to facilitate data restore operations to a file server
US10742735B2 (en) 2017-12-12 2020-08-11 Commvault Systems, Inc. Enhanced network attached storage (NAS) services interfacing to cloud storage
US10834189B1 (en) 2018-01-10 2020-11-10 EMC IP Holding Company LLC System and method for managing workload in a pooled environment
US10831591B2 (en) 2018-01-11 2020-11-10 Commvault Systems, Inc. Remedial action based on maintaining process awareness in data storage management
US10795927B2 (en) 2018-02-05 2020-10-06 Commvault Systems, Inc. On-demand metadata extraction of clinical image data
US10732885B2 (en) 2018-02-14 2020-08-04 Commvault Systems, Inc. Block-level live browsing and private writable snapshots using an ISCSI server
US10642886B2 (en) 2018-02-14 2020-05-05 Commvault Systems, Inc. Targeted search of backup data using facial recognition
US10592145B2 (en) 2018-02-14 2020-03-17 Commvault Systems, Inc. Machine learning-based data object storage
US10848468B1 (en) 2018-03-05 2020-11-24 Commvault Systems, Inc. In-flight data encryption/decryption for a distributed storage platform
US10877928B2 (en) 2018-03-07 2020-12-29 Commvault Systems, Inc. Using utilities injected into cloud-based virtual machines for speeding up virtual machine backup operations
US10761942B2 (en) 2018-03-12 2020-09-01 Commvault Systems, Inc. Recovery point objective (RPO) driven backup scheduling in a data storage management system using an enhanced data agent
US10789387B2 (en) 2018-03-13 2020-09-29 Commvault Systems, Inc. Graphical representation of an information management system
US10509587B2 (en) 2018-04-24 2019-12-17 EMC IP Holding Company LLC System and method for high priority backup
US10769030B2 (en) 2018-04-25 2020-09-08 EMC IP Holding Company LLC System and method for improved cache performance
US10673943B2 (en) 2018-05-02 2020-06-02 Commvault Systems, Inc. Network storage backup using distributed media agents
US11249863B2 (en) 2018-05-02 2022-02-15 Commvault Systems, Inc. Backup-based media agent configuration
US10628267B2 (en) 2018-05-02 2020-04-21 Commvault Systems, Inc. Client managed data backup process within an enterprise information management system
US11321183B2 (en) 2018-05-02 2022-05-03 Commvault Systems, Inc. Multi-tiered backup indexing
US10891304B2 (en) 2018-07-25 2021-01-12 Commvault Systems, Inc. Distributed and scalable client-based storage management
US10893103B2 (en) 2018-07-25 2021-01-12 Commvault Systems, Inc. Distributed and scalable storage management using a storage-microservices server
US10891198B2 (en) 2018-07-30 2021-01-12 Commvault Systems, Inc. Storing data to cloud libraries in cloud native formats
US11159469B2 (en) 2018-09-12 2021-10-26 Commvault Systems, Inc. Using machine learning to modify presentation of mailbox objects
US11016696B2 (en) 2018-09-14 2021-05-25 Commvault Systems, Inc. Redundant distributed data storage system
US11010258B2 (en) 2018-11-27 2021-05-18 Commvault Systems, Inc. Generating backup copies through interoperability between components of a data storage management system and appliances for data storage and deduplication
US11200124B2 (en) 2018-12-06 2021-12-14 Commvault Systems, Inc. Assigning backup resources based on failover of partnered data storage servers in a data storage management system
US10860443B2 (en) 2018-12-10 2020-12-08 Commvault Systems, Inc. Evaluation and reporting of recovery readiness in a data storage management system
US20200192572A1 (en) 2018-12-14 2020-06-18 Commvault Systems, Inc. Disk usage growth prediction system
US11012508B2 (en) 2018-12-14 2021-05-18 Commvault Systems, Inc. Region-based distributed information management system
US11698727B2 (en) 2018-12-14 2023-07-11 Commvault Systems, Inc. Performing secondary copy operations based on deduplication performance
US10872069B2 (en) 2019-01-22 2020-12-22 Commvault Systems, Inc. File indexing for virtual machine backups in a data storage management system
US11347707B2 (en) 2019-01-22 2022-05-31 Commvault Systems, Inc. File indexing for virtual machine backups based on using live browse features
US10768971B2 (en) 2019-01-30 2020-09-08 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data
US10996974B2 (en) 2019-01-30 2021-05-04 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data, including management of cache storage for virtual machine data
US11126365B2 (en) 2019-03-11 2021-09-21 Commvault Systems, Inc. Skipping data backed up in prior backup operations
US11269732B2 (en) 2019-03-12 2022-03-08 Commvault Systems, Inc. Managing structured data in a data storage system
US11321184B2 (en) 2019-03-26 2022-05-03 Commvault Systems, Inc. Streamlined secondary copy operations for data stored on shared file storage
US11579681B2 (en) 2019-04-08 2023-02-14 Commvault Systems, Inc. Power management of components within a storage management system
US20200327017A1 (en) 2019-04-10 2020-10-15 Commvault Systems, Inc. Restore using deduplicated secondary copy data
US11366723B2 (en) 2019-04-30 2022-06-21 Commvault Systems, Inc. Data storage management system for holistic protection and migration of serverless applications across multi-cloud computing environments
US11100064B2 (en) 2019-04-30 2021-08-24 Commvault Systems, Inc. Automated log-based remediation of an information management system
US11442623B2 (en) 2019-05-02 2022-09-13 Commvault Systems, Inc. Faster browse of secondary copies of block-level data volumes
US11463264B2 (en) 2019-05-08 2022-10-04 Commvault Systems, Inc. Use of data block signatures for monitoring in an information management system
CN110221782A (en) * 2019-06-06 2019-09-10 重庆紫光华山智安科技有限公司 Video file processing method and processing device
US11269734B2 (en) 2019-06-17 2022-03-08 Commvault Systems, Inc. Data storage management system for multi-cloud protection, recovery, and migration of databases-as-a-service and/or serverless database management systems
US11308034B2 (en) 2019-06-27 2022-04-19 Commvault Systems, Inc. Continuously run log backup with minimal configuration and resource usage from the source machine
US11561866B2 (en) 2019-07-10 2023-01-24 Commvault Systems, Inc. Preparing containerized applications for backup using a backup services container and a backup services container-orchestration pod
US11042318B2 (en) 2019-07-29 2021-06-22 Commvault Systems, Inc. Block-level data replication
US20210037112A1 (en) 2019-07-29 2021-02-04 Commvault Systems, Inc. Data storage system with rapid restore capability
US11263173B2 (en) 2019-07-30 2022-03-01 Commvault Systems, Inc. Transaction log index generation in an enterprise backup system
US11513922B2 (en) 2019-08-14 2022-11-29 Commvault Systems, Inc. Systems and methods for change block tracking for backing up changed data
US11775484B2 (en) 2019-08-27 2023-10-03 Vmware, Inc. Fast algorithm to find file system difference for deduplication
US11461229B2 (en) * 2019-08-27 2022-10-04 Vmware, Inc. Efficient garbage collection of variable size chunking deduplication
US11218450B2 (en) 2019-09-11 2022-01-04 Commvault Systems, Inc. Data protection component scaling in a cloud-based data storage system
US11256673B2 (en) 2019-09-11 2022-02-22 Commvault Systems, Inc. Anomaly detection in deduplication pruning operations
US11237935B2 (en) 2019-09-11 2022-02-01 Commvault Systems, Inc. Anomaly detection in data protection operations
US11409727B2 (en) * 2019-09-18 2022-08-09 International Business Machines Corporation Concurrent execution of database operations
US11663178B2 (en) 2019-11-04 2023-05-30 Commvault Systems, Inc. Efficient implementation of multiple deduplication databases in a heterogeneous data storage system
US20210173811A1 (en) 2019-12-04 2021-06-10 Commvault Systems, Inc. Optimizing the restoration of deduplicated data stored in multi-node replicated file systems
US11237924B2 (en) 2019-12-11 2022-02-01 Commvault Systems, Inc. Dynamic resizing and re-distribution of destination data storage resources for bare metal restore operations in a data storage management system
US11223535B2 (en) 2019-12-31 2022-01-11 Commvault Systems, Inc. Smart network topology systems and methods
US11467753B2 (en) 2020-02-14 2022-10-11 Commvault Systems, Inc. On-demand restore of virtual machine data
US11422900B2 (en) 2020-03-02 2022-08-23 Commvault Systems, Inc. Platform-agnostic containerized application data protection
US11321188B2 (en) 2020-03-02 2022-05-03 Commvault Systems, Inc. Platform-agnostic containerized application data protection
US11334442B2 (en) 2020-03-10 2022-05-17 Commvault Systems, Inc. Management database long-term archiving to a recovery manager
US11005935B1 (en) 2020-03-10 2021-05-11 Commvault Systems, Inc. Using multiple streams with network data management protocol to improve performance and granularity of backup and restore operations from/to a file server
US11442768B2 (en) 2020-03-12 2022-09-13 Commvault Systems, Inc. Cross-hypervisor live recovery of virtual machines
US11099956B1 (en) 2020-03-26 2021-08-24 Commvault Systems, Inc. Snapshot-based disaster recovery orchestration of virtual machine failover and failback operations
US11704035B2 (en) 2020-03-30 2023-07-18 Pure Storage, Inc. Unified storage on block containers
US11748143B2 (en) 2020-05-15 2023-09-05 Commvault Systems, Inc. Live mount of virtual machines in a public cloud computing environment
US11687424B2 (en) 2020-05-28 2023-06-27 Commvault Systems, Inc. Automated media agent state management
US11327663B2 (en) 2020-06-09 2022-05-10 Commvault Systems, Inc. Ensuring the integrity of data storage volumes used in block-level live synchronization operations in a data storage management system
US11537959B2 (en) 2020-06-16 2022-12-27 Commvault Systems, Inc. Dynamic computing progress tracker
US11487468B2 (en) 2020-07-17 2022-11-01 Commvault Systems, Inc. Healing failed erasure-coded write attempts in a distributed data storage system configured with fewer storage nodes than data plus parity fragments
US11494417B2 (en) 2020-08-07 2022-11-08 Commvault Systems, Inc. Automated email classification in an information management system
US11500566B2 (en) 2020-08-25 2022-11-15 Commvault Systems, Inc. Cloud-based distributed data storage system using block-level deduplication based on backup frequencies of incoming backup copies
US11108861B1 (en) 2020-08-26 2021-08-31 Commvault Systems, Inc. System for managing multiple information management cells
US11570243B2 (en) 2020-09-22 2023-01-31 Commvault Systems, Inc. Decommissioning, re-commissioning, and commissioning new metadata nodes in a working distributed data storage system
US11789830B2 (en) 2020-09-22 2023-10-17 Commvault Systems, Inc. Anti-entropy-based metadata recovery in a strongly consistent distributed data storage system
US11314687B2 (en) 2020-09-24 2022-04-26 Commvault Systems, Inc. Container data mover for migrating data between distributed data storage systems integrated with application orchestrators
US11656951B2 (en) 2020-10-28 2023-05-23 Commvault Systems, Inc. Data loss vulnerability detection
US11474753B2 (en) 2020-10-30 2022-10-18 Commvault Systems, Inc. Systems and methods for backing up to a virtual tape library
US11323513B1 (en) 2021-01-22 2022-05-03 Commvault Systems, Inc. Geo-location-based data replication
US11604706B2 (en) 2021-02-02 2023-03-14 Commvault Systems, Inc. Back up and restore related data on different cloud storage tiers
US11645175B2 (en) 2021-02-12 2023-05-09 Commvault Systems, Inc. Automatic failover of a storage manager
US11574050B2 (en) 2021-03-12 2023-02-07 Commvault Systems, Inc. Media agent hardening against ransomware attacks
US11561978B2 (en) 2021-06-29 2023-01-24 Commvault Systems, Inc. Intelligent cache management for mounted snapshots based on a behavior model
US11615147B2 (en) 2021-08-23 2023-03-28 Commvault Systems, Inc. Mobile storage manager control application for managing a storage manager of an information management system
US11593223B1 (en) 2021-09-02 2023-02-28 Commvault Systems, Inc. Using resource pool administrative entities in a data storage management system to provide shared infrastructure to tenants
US11809285B2 (en) 2022-02-09 2023-11-07 Commvault Systems, Inc. Protecting a management database of a data storage management system to meet a recovery point objective (RPO)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020112022A1 (en) * 2000-12-18 2002-08-15 Spinnaker Networks, Inc. Mechanism for handling file level and block level remote file accesses using the same server
US6928459B1 (en) * 2000-07-18 2005-08-09 International Business Machines Corporation Plurality of file systems using weighted allocation to allocate space on one or more storage devices
US20050210460A1 (en) * 2004-03-22 2005-09-22 Microsoft Corporation Computing device with relatively limited storage space and operating/file system thereof
US20090106480A1 (en) * 2007-10-23 2009-04-23 Keicy Chung Computer storage device having separate read-only space and read-write space, removable media component, system management interface, and network interface

Family Cites Families (430)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US675779A (en) * 1900-07-30 1901-06-04 George Theodore Temple Bottle-decanting apparatus.
US4686620A (en) 1984-07-26 1987-08-11 American Telephone And Telegraph Company, At&T Bell Laboratories Database backup method
US4713755A (en) * 1985-06-28 1987-12-15 Hewlett-Packard Company Cache memory consistency control with explicit software instructions
GB8622010D0 (en) 1986-09-12 1986-10-22 Hewlett Packard Ltd File backup facility
US5193154A (en) 1987-07-10 1993-03-09 Hitachi, Ltd. Buffered peripheral system and method for backing up and retrieving data to and from backup memory device
US5005122A (en) * 1987-09-08 1991-04-02 Digital Equipment Corporation Arrangement with cooperating management server node and network service node
JPH0743676B2 (en) 1988-03-11 1995-05-15 株式会社日立製作所 Back-up data dump control method and device
US4995035A (en) * 1988-10-31 1991-02-19 International Business Machines Corporation Centralized management in a computer network
US5093912A (en) 1989-06-26 1992-03-03 International Business Machines Corporation Dynamic resource pool expansion and contraction in multiprocessing environments
ATE145998T1 (en) 1989-06-30 1996-12-15 Digital Equipment Corp METHOD AND ARRANGEMENT FOR CONTROLLING SHADOW STORAGE
US5133065A (en) 1989-07-27 1992-07-21 Personal Computer Peripherals Corporation Backup computer program for networks
US5321816A (en) * 1989-10-10 1994-06-14 Unisys Corporation Local-remote apparatus with specialized image storage modules
US5504873A (en) * 1989-11-01 1996-04-02 E-Systems, Inc. Mass data storage and retrieval system
US5276880A (en) * 1989-12-15 1994-01-04 Siemens Corporate Research, Inc. Method for parsing and representing multi-versioned computer programs, for simultaneous and synchronous processing of the plural parses
US5276860A (en) 1989-12-19 1994-01-04 Epoch Systems, Inc. Digital data processor with improved backup storage
US5276867A (en) * 1989-12-19 1994-01-04 Epoch Systems, Inc. Digital data storage system with improved data migration
GB2246218B (en) 1990-07-18 1994-02-09 Stc Plc Distributed data processing systems
US5239647A (en) 1990-09-07 1993-08-24 International Business Machines Corporation Data storage hierarchy with shared storage level
US5544347A (en) * 1990-09-24 1996-08-06 Emc Corporation Data storage system controlled remote data mirroring with respectively maintained data indices
US5212772A (en) * 1991-02-11 1993-05-18 Gigatrend Incorporated System for storing data in backup tape device
US5287500A (en) * 1991-06-03 1994-02-15 Digital Equipment Corporation System for allocating storage spaces based upon required and optional service attributes having assigned piorities
US5333315A (en) 1991-06-27 1994-07-26 Digital Equipment Corporation System of device independent file directories using a tag between the directories and file descriptors that migrate with the files
US5347653A (en) 1991-06-28 1994-09-13 Digital Equipment Corporation System for reconstructing prior versions of indexes using records indicating changes between successive versions of the indexes
US5410700A (en) * 1991-09-04 1995-04-25 International Business Machines Corporation Computer system which supports asynchronous commitment of data
US5263154A (en) 1992-04-20 1993-11-16 International Business Machines Corporation Method and system for incremental time zero backup copying of data
US5241670A (en) 1992-04-20 1993-08-31 International Business Machines Corporation Method and system for automated backup copy ordering in a time zero backup copy session
US5241668A (en) 1992-04-20 1993-08-31 International Business Machines Corporation Method and system for automated termination and resumption in a time zero backup copy process
EP0680634B1 (en) * 1993-01-21 1997-05-14 Apple Computer, Inc. Apparatus and method for backing up data from networked computer storage devices
EP0681721B1 (en) * 1993-02-01 2005-03-23 Sun Microsystems, Inc. Archiving file system for data servers in a distributed network environment
US5794229A (en) * 1993-04-16 1998-08-11 Sybase, Inc. Database system with methodology for storing a database table by vertically partitioning all columns of the table
US5437012A (en) 1993-04-19 1995-07-25 Canon Information Systems, Inc. System for updating directory information and data on write once media such as an optical memory card
JPH0721135A (en) 1993-07-02 1995-01-24 Fujitsu Ltd Data processing system with duplex monitor function
JP3200757B2 (en) * 1993-10-22 2001-08-20 株式会社日立製作所 Storage control method for parallel computer and parallel computer
US5544345A (en) 1993-11-08 1996-08-06 International Business Machines Corporation Coherence controls for store-multiple shared data coordinated by cache directory entries in a shared electronic storage
WO1995013580A1 (en) 1993-11-09 1995-05-18 Arcada Software Data backup and restore system for a computer network
US5495607A (en) * 1993-11-15 1996-02-27 Conner Peripherals, Inc. Network management system having virtual catalog overview of files distributively stored across network domain
US5491810A (en) * 1994-03-01 1996-02-13 International Business Machines Corporation Method and system for automated data storage system space allocation utilizing prioritized data set parameters
US5673381A (en) 1994-05-27 1997-09-30 Cheyenne Software International Sales Corp. System and parallel streaming and data stripping to back-up a network
US5638509A (en) * 1994-06-10 1997-06-10 Exabyte Corporation Data storage and protection system
US5574906A (en) 1994-10-24 1996-11-12 International Business Machines Corporation System and method for reducing storage requirement in backup subsystems utilizing segmented compression and differencing
US5634052A (en) * 1994-10-24 1997-05-27 International Business Machines Corporation System for reducing storage requirements and transmission loads in a backup subsystem in client-server environment by transmitting only delta files from client to server
US5628004A (en) * 1994-11-04 1997-05-06 Optima Direct, Inc. System for managing database of communication of recipients
US5990810A (en) 1995-02-17 1999-11-23 Williams; Ross Neil Method for partitioning a block of data into subblocks and for storing and communcating such subblocks
US5604862A (en) * 1995-03-14 1997-02-18 Network Integrity, Inc. Continuously-snapshotted protection of computer files
US5559957A (en) 1995-05-31 1996-09-24 Lucent Technologies Inc. File system for a data storage device having a power fail recovery mechanism for write/replace operations
US5699361A (en) 1995-07-18 1997-12-16 Industrial Technology Research Institute Multimedia channel formulation mechanism
US5813009A (en) 1995-07-28 1998-09-22 Univirtual Corp. Computer based records management system method
US5619644A (en) * 1995-09-18 1997-04-08 International Business Machines Corporation Software directed microcode state save for distributed storage controller
US5819020A (en) 1995-10-16 1998-10-06 Network Specialists, Inc. Real time backup system
US5778395A (en) 1995-10-23 1998-07-07 Stac, Inc. System for backing up files from disk volumes on multiple nodes of a computer network
US5729743A (en) * 1995-11-17 1998-03-17 Deltatech Research, Inc. Computer apparatus and method for merging system deltas
US5761677A (en) * 1996-01-03 1998-06-02 Sun Microsystems, Inc. Computer system method and apparatus providing for various versions of a file without requiring data copy or log operations
US5862325A (en) 1996-02-29 1999-01-19 Intermind Corporation Computer-based communication system and method using metadata defining a control structure
KR970076238A (en) 1996-05-23 1997-12-12 포만 제프리 엘 Servers, methods and program products thereof for creating and managing multiple copies of client data files
US5889935A (en) * 1996-05-28 1999-03-30 Emc Corporation Disaster control features for remote data mirroring
US5812398A (en) 1996-06-10 1998-09-22 Sun Microsystems, Inc. Method and system for escrowed backup of hotelled world wide web sites
US5813008A (en) 1996-07-12 1998-09-22 Microsoft Corporation Single instance storage of information
US5940833A (en) 1996-07-12 1999-08-17 Microsoft Corporation Compressing sets of integers
US5758359A (en) * 1996-10-24 1998-05-26 Digital Equipment Corporation Method and apparatus for performing retroactive backups in a computer system
US5875478A (en) * 1996-12-03 1999-02-23 Emc Corporation Computer backup using a file system, network, disk, tape and remote archiving repository media system
US6131095A (en) 1996-12-11 2000-10-10 Hewlett-Packard Company Method of accessing a target entity over a communications network
US5822780A (en) 1996-12-31 1998-10-13 Emc Corporation Method and apparatus for hierarchical storage management for data base management systems
WO1998033113A1 (en) 1997-01-23 1998-07-30 Overland Data, Inc. Virtual media library
US6658526B2 (en) 1997-03-12 2003-12-02 Storage Technology Corporation Network attached virtual data storage subsystem
US5924102A (en) 1997-05-07 1999-07-13 International Business Machines Corporation System and method for managing critical files
US6094416A (en) 1997-05-09 2000-07-25 I/O Control Corporation Multi-tier architecture for control network
US6311252B1 (en) 1997-06-30 2001-10-30 Emc Corporation Method and apparatus for moving data between storage levels of a hierarchically arranged data storage system
US5887134A (en) * 1997-06-30 1999-03-23 Sun Microsystems System and method for preserving message order while employing both programmed I/O and DMA operations
EP0899662A1 (en) 1997-08-29 1999-03-03 Hewlett-Packard Company Backup and restore system for a computer network
DE69802294T2 (en) 1997-08-29 2002-05-16 Hewlett Packard Co SYSTEMS FOR DATA BACKUP AND RECOVERY
US5950205A (en) 1997-09-25 1999-09-07 Cisco Technology, Inc. Data transmission over the internet using a cache memory file system
US6275953B1 (en) 1997-09-26 2001-08-14 Emc Corporation Recovery from failure of a data processor in a network server
US6173291B1 (en) * 1997-09-26 2001-01-09 Powerquest Corporation Method and apparatus for recovering data from damaged or corrupted file storage media
US6125369A (en) 1997-10-02 2000-09-26 Microsoft Corporation Continuous object sychronization between object stores on different computers
US6052735A (en) * 1997-10-24 2000-04-18 Microsoft Corporation Electronic mail object synchronization between a desktop computer and mobile device
US6021415A (en) * 1997-10-29 2000-02-01 International Business Machines Corporation Storage management system with file aggregation and space reclamation within aggregated files
US6418478B1 (en) 1997-10-30 2002-07-09 Commvault Systems, Inc. Pipelined high speed data transfer mechanism
JPH11143754A (en) 1997-11-05 1999-05-28 Hitachi Ltd Version information and constitution information display method and device therefor, and computer readable recording medium for recording version information and constitution information display program
US6131190A (en) 1997-12-18 2000-10-10 Sidwell; Leland P. System for modifying JCL parameters to optimize data storage allocations
US6076148A (en) * 1997-12-26 2000-06-13 Emc Corporation Mass storage subsystem and backup arrangement for digital data processing system which permits information to be backed up while host computer(s) continue(s) operating in connection with information stored on mass storage subsystem
US6154787A (en) 1998-01-21 2000-11-28 Unisys Corporation Grouping shared resources into one or more pools and automatically re-assigning shared resources from where they are not currently needed to where they are needed
US6807632B1 (en) 1999-01-21 2004-10-19 Emc Corporation Content addressable information encapsulation, representation, and transfer
US6260069B1 (en) 1998-02-10 2001-07-10 International Business Machines Corporation Direct data retrieval in a distributed computing system
US6901493B1 (en) 1998-02-24 2005-05-31 Adaptec, Inc. Method for protecting data of a computer system
EP0945800B1 (en) 1998-03-02 2003-07-16 Hewlett-Packard Company, A Delaware Corporation Data backup system
US6026414A (en) * 1998-03-05 2000-02-15 International Business Machines Corporation System including a proxy client to backup files in a distributed computing environment
US6161111A (en) 1998-03-31 2000-12-12 Emc Corporation System and method for performing file-handling operations in a digital data processing system using an operating system-independent file map
US6167402A (en) 1998-04-27 2000-12-26 Sun Microsystems, Inc. High performance message store
US6073133A (en) * 1998-05-15 2000-06-06 Micron Electronics Inc. Electronic mail attachment verifier
US7035943B2 (en) * 1998-05-29 2006-04-25 Yahoo! Inc. Web server content replication
US6421711B1 (en) 1998-06-29 2002-07-16 Emc Corporation Virtual ports for data transferring of a data storage system
US6269431B1 (en) 1998-08-13 2001-07-31 Emc Corporation Virtual storage and block level direct access of secondary storage for recovery of backup data
GB2341249A (en) 1998-08-17 2000-03-08 Connected Place Limited A method of generating a difference file defining differences between an updated file and a base file
US7111173B1 (en) 1998-09-01 2006-09-19 Tecsec, Inc. Encryption process including a biometric unit
US6304914B1 (en) 1998-09-22 2001-10-16 Microsoft Corporation Method and apparatus for pre-compression packaging
GB2342198B (en) 1998-10-02 2003-10-22 Ibm Composite locking of objects in a database
US6324544B1 (en) 1998-10-21 2001-11-27 Microsoft Corporation File object synchronization between a desktop computer and a mobile device
US6487561B1 (en) 1998-12-31 2002-11-26 Emc Corporation Apparatus and methods for copying, backing up, and restoring data using a backup segment size larger than the storage block size
US6212512B1 (en) * 1999-01-06 2001-04-03 Hewlett-Packard Company Integration of a database into file management software for protecting, tracking and retrieving data
CN1293491C (en) 1999-02-17 2007-01-03 索尼公司 Information processing appliance and method, and program storage medium
US6356915B1 (en) * 1999-02-22 2002-03-12 Starbase Corp. Installable file system having virtual file system drive, virtual device driver, and virtual disks
US6397307B2 (en) 1999-02-23 2002-05-28 Legato Systems, Inc. Method and system for mirroring and archiving mass storage
US6324581B1 (en) 1999-03-03 2001-11-27 Emc Corporation File server system using file system storage, data movers, and an exchange of meta data among data movers for file locking and direct access to shared file systems
US6389432B1 (en) * 1999-04-05 2002-05-14 Auspex Systems, Inc. Intelligent virtual volume access
US6519679B2 (en) * 1999-06-11 2003-02-11 Dell Usa, L.P. Policy based storage configuration
US6959368B1 (en) 1999-06-29 2005-10-25 Emc Corporation Method and apparatus for duplicating computer backup data
US6609187B1 (en) 1999-07-06 2003-08-19 Dell Products L.P. Method and apparatus for supporting resizing of file system partitions
US7035880B1 (en) 1999-07-14 2006-04-25 Commvault Systems, Inc. Modular backup and retrieval system used in conjunction with a storage area network
US6538669B1 (en) * 1999-07-15 2003-03-25 Dell Products L.P. Graphical user interface for configuration of a storage system
US7395282B1 (en) 1999-07-15 2008-07-01 Commvault Systems, Inc. Hierarchical backup and retrieval system
US6477544B1 (en) * 1999-07-16 2002-11-05 Microsoft Corporation Single instance store for file systems
US6513051B1 (en) * 1999-07-16 2003-01-28 Microsoft Corporation Method and system for backing up and restoring files stored in a single instance store
US6490666B1 (en) 1999-08-20 2002-12-03 Microsoft Corporation Buffering data in a hierarchical data storage environment
US6343324B1 (en) * 1999-09-13 2002-01-29 International Business Machines Corporation Method and system for controlling access share storage devices in a network environment by configuring host-to-volume mapping data structures in the controller memory for granting and denying access to the devices
US6564228B1 (en) * 2000-01-14 2003-05-13 Sun Microsystems, Inc. Method of enabling heterogeneous platforms to utilize a universal file system in a storage area network
US6973187B2 (en) * 2000-01-31 2005-12-06 Vdg, Inc. Block encryption method and schemes for data confidentiality and integrity protection
US7412462B2 (en) 2000-02-18 2008-08-12 Burnside Acquisition, Llc Data repository and method for promoting network storage of data
US6704730B2 (en) 2000-02-18 2004-03-09 Avamar Technologies, Inc. Hash file system and method for use in a commonality factoring system
US7117246B2 (en) 2000-02-22 2006-10-03 Sendmail, Inc. Electronic mail system with methodology providing distributed message store
CA2400943C (en) * 2000-02-24 2009-06-09 Alcoa, Inc. Method of converting hall-heroult cells to inert anode cells for aluminum production
US7143108B1 (en) 2000-04-06 2006-11-28 International Business Machines Corporation Apparatus and method for deletion of objects from an object-relational system in a customizable and database independent manner
JP2001356945A (en) 2000-04-12 2001-12-26 Anetsukusu Syst Kk Data backup recovery system
US20020055972A1 (en) 2000-05-08 2002-05-09 Weinman Joseph Bernard Dynamic content distribution and data continuity architecture
US6356801B1 (en) * 2000-05-19 2002-03-12 International Business Machines Corporation High availability work queuing in an automated data storage library
US6675177B1 (en) 2000-06-21 2004-01-06 Teradactyl, Llc Method and system for backing up digital data
US6330642B1 (en) 2000-06-29 2001-12-11 Bull Hn Informatin Systems Inc. Three interconnected raid disk controller data processing system architecture
US6502111B1 (en) * 2000-07-31 2002-12-31 Microsoft Corporation Method and system for concurrent garbage collection
DE60123442D1 (en) 2000-08-31 2006-11-09 Ontrack Data Internat Inc SYSTEM AND METHOD FOR DATA MANAGEMENT
US6757699B2 (en) * 2000-10-06 2004-06-29 Franciscan University Of Steubenville Method and system for fragmenting and reconstituting data
US6973553B1 (en) 2000-10-20 2005-12-06 International Business Machines Corporation Method and apparatus for using extended disk sector formatting to assist in backup and hierarchical storage management
US8996698B1 (en) 2000-11-03 2015-03-31 Truphone Limited Cooperative network for mobile internet access
US6810398B2 (en) 2000-11-06 2004-10-26 Avamar Technologies, Inc. System and method for unorchestrated determination of data sequences using sticky byte factoring to determine breakpoints in digital sequences
US20020099806A1 (en) 2000-11-30 2002-07-25 Phillip Balsamo Processing node for eliminating duplicate network usage data
US7003551B2 (en) 2000-11-30 2006-02-21 Bellsouth Intellectual Property Corp. Method and apparatus for minimizing storage of common attachment files in an e-mail communications server
FR2820849B1 (en) * 2001-02-15 2003-05-16 Cit Alcatel COMPUTER DATA STORAGE METHOD AND CORRESPONDING STORAGE DEVICE
US6820081B1 (en) * 2001-03-19 2004-11-16 Attenex Corporation System and method for evaluating a structured message store for message redundancy
EP1244221A1 (en) 2001-03-23 2002-09-25 Sun Microsystems, Inc. Method and system for eliminating data redundancies
US7856414B2 (en) 2001-03-29 2010-12-21 Christopher Zee Assured archival and retrieval system for digital intellectual property
US6976039B2 (en) 2001-05-25 2005-12-13 International Business Machines Corporation Method and system for processing backup data associated with application, querying metadata files describing files accessed by the application
US6915397B2 (en) 2001-06-01 2005-07-05 Hewlett-Packard Development Company, L.P. System and method for generating point in time storage copy
US6988124B2 (en) 2001-06-06 2006-01-17 Microsoft Corporation Locating potentially identical objects across multiple computers based on stochastic partitioning of workload
US6993162B2 (en) 2001-06-15 2006-01-31 Eastman Kodak Company Method for authenticating animation
US20030004922A1 (en) * 2001-06-27 2003-01-02 Ontrack Data International, Inc. System and method for data management
US20050134578A1 (en) 2001-07-13 2005-06-23 Universal Electronics Inc. System and methods for interacting with a control environment
US6912645B2 (en) 2001-07-19 2005-06-28 Lucent Technologies Inc. Method and apparatus for archival data storage
US7516208B1 (en) 2001-07-20 2009-04-07 International Business Machines Corporation Event database management method and system for network event reporting system
US7685126B2 (en) * 2001-08-03 2010-03-23 Isilon Systems, Inc. System and methods for providing a distributed file system utilizing metadata to track information about data stored throughout the system
WO2003027891A1 (en) 2001-09-28 2003-04-03 Commvault Systems, Inc. System and method for archiving objects in an information store
US20030167318A1 (en) 2001-10-22 2003-09-04 Apple Computer, Inc. Intelligent synchronization of media player with host computer
US7092956B2 (en) * 2001-11-02 2006-08-15 General Electric Capital Corporation Deduplication system
ES2359399T3 (en) * 2001-11-23 2011-05-23 Commvault Systems, Inc. SYSTEM AND PROCEDURE FOR SELECTIVE DATA DUPLICATION.
US7496604B2 (en) * 2001-12-03 2009-02-24 Aol Llc Reducing duplication of files on a network
US20030110190A1 (en) * 2001-12-10 2003-06-12 Hitachi, Ltd. Method and system for file space management
US20030172368A1 (en) 2001-12-26 2003-09-11 Elizabeth Alumbaugh System and method for autonomously generating heterogeneous data source interoperability bridges based on semantic modeling derived from self adapting ontology
US6839819B2 (en) 2001-12-28 2005-01-04 Storage Technology Corporation Data management appliance
US20030135480A1 (en) 2002-01-14 2003-07-17 Van Arsdale Robert S. System for updating a database
US6795903B2 (en) 2002-01-17 2004-09-21 Thomas Licensing S.A. System and method for searching for duplicate data
US7017113B2 (en) * 2002-01-25 2006-03-21 The United States Of America As Represented By The Secretary Of The Air Force Method and apparatus for removing redundant information from digital documents
WO2003067497A1 (en) 2002-02-04 2003-08-14 Cataphora, Inc A method and apparatus to visually present discussions for data mining purposes
JP3943949B2 (en) * 2002-02-12 2007-07-11 キヤノン株式会社 E-mail processing system, method, program, and storage medium
US20030177149A1 (en) 2002-03-18 2003-09-18 Coombs David Lawrence System and method for data backup
AU2002244892A1 (en) 2002-03-26 2003-10-08 Nokia Corporation Method and apparatus for compressing log record information
US7139808B2 (en) 2002-04-30 2006-11-21 Intel Corporation Method and apparatus for bandwidth-efficient and storage-efficient backups
US20060089954A1 (en) 2002-05-13 2006-04-27 Anschutz Thomas A Scalable common access back-up architecture
US6862674B2 (en) * 2002-06-06 2005-03-01 Sun Microsystems Methods and apparatus for performing a memory management technique
US20030236763A1 (en) 2002-06-25 2003-12-25 Alan Kilduff Electronic message filing system
US7275063B2 (en) 2002-07-16 2007-09-25 Horn Bruce L Computer system for automatic organization, indexing and viewing of information from multiple sources
US6865655B1 (en) * 2002-07-30 2005-03-08 Sun Microsystems, Inc. Methods and apparatus for backing up and restoring data portions stored in client computer systems
US6952758B2 (en) 2002-07-31 2005-10-04 International Business Machines Corporation Method and system for providing consistent data modification information to clients in a storage system
US7171469B2 (en) * 2002-09-16 2007-01-30 Network Appliance, Inc. Apparatus and method for storing data in a proxy cache in a network
US7287252B2 (en) 2002-09-27 2007-10-23 The United States Of America Represented By The Secretary Of The Navy Universal client and consumer
US7089395B2 (en) 2002-10-03 2006-08-08 Hewlett-Packard Development Company, L.P. Computer systems, virtual storage systems and virtual storage system operational methods
CA2508089A1 (en) 2002-10-07 2004-04-22 Commvault Systems, Inc. System and method for managing stored data
WO2004034286A1 (en) 2002-10-10 2004-04-22 Action Engine Corporation Backing up a wireless computing device
US7117324B2 (en) 2002-10-18 2006-10-03 International Business Machines Corporation Simultaneous data backup in a computer system
US7370003B2 (en) 2002-11-08 2008-05-06 Amdocs Software Systems Ltd. Method and apparatus for implied attribution of responses to promotional contacts
US7136850B2 (en) 2002-12-20 2006-11-14 International Business Machines Corporation Self tuning database retrieval optimization using regression functions
WO2004061590A2 (en) * 2003-01-02 2004-07-22 Cricket Technologies Llc Electronic archive filter and profiling apparatus, system, method, and electronically stored computer program product
US20040187075A1 (en) 2003-01-08 2004-09-23 Maxham Jason G. Document management apparatus, system and method
US8375008B1 (en) 2003-01-17 2013-02-12 Robert Gomes Method and system for enterprise-wide retention of digital or electronic data
US7403942B1 (en) * 2003-02-04 2008-07-22 Seisint, Inc. Method and system for processing data records
US7830889B1 (en) 2003-02-06 2010-11-09 Juniper Networks, Inc. Systems for scheduling the transmission of data in a network device
JP4991283B2 (en) 2003-02-21 2012-08-01 カリンゴ・インコーポレーテッド Additional hash functions in content-based addressing
US7478096B2 (en) 2003-02-26 2009-01-13 Burnside Acquisition, Llc History preservation in a computer storage system
US7389345B1 (en) 2003-03-26 2008-06-17 Sprint Communications Company L.P. Filtering approach for network system alarms
WO2004090740A1 (en) * 2003-04-03 2004-10-21 Commvault Systems, Inc. System and method for dynamically sharing media in a computer network
US20040230817A1 (en) 2003-05-14 2004-11-18 Kenneth Ma Method and system for disaster recovery of data from a storage device
WO2004104739A2 (en) 2003-05-14 2004-12-02 Rhysome, Inc. Method and system for reducing information latency in a business enterprise
US7089383B2 (en) 2003-06-06 2006-08-08 Hewlett-Packard Development Company, L.P. State machine and system for data redundancy
US7412489B2 (en) 2003-08-21 2008-08-12 Ameriprise Financial, Inc. Method and system for electronic archival and retrieval of electronic communications
US20050060643A1 (en) 2003-08-25 2005-03-17 Miavia, Inc. Document similarity detection and classification system
US7814142B2 (en) 2003-08-27 2010-10-12 International Business Machines Corporation User interface service for a services oriented architecture in a data integration platform
US20050223109A1 (en) 2003-08-27 2005-10-06 Ascential Software Corporation Data integration through a services oriented architecture
US20050262193A1 (en) 2003-08-27 2005-11-24 Ascential Software Corporation Logging service for a services oriented architecture in a data integration platform
US7536440B2 (en) 2003-09-18 2009-05-19 Vulcan Portals Inc. Method and system for email synchronization for an electronic device
US7631120B2 (en) 2004-08-24 2009-12-08 Symantec Operating Corporation Methods and apparatus for optimally selecting a storage buffer for the storage of data
US7143117B2 (en) * 2003-09-25 2006-11-28 International Business Machines Corporation Method, system, and program for data synchronization by determining whether a first identifier for a portion of data at a first source and a second identifier for a portion of corresponding data at a second source match
JP4267420B2 (en) 2003-10-20 2009-05-27 株式会社日立製作所 Storage apparatus and backup acquisition method
GB0325626D0 (en) 2003-11-03 2003-12-10 Infoshare Ltd Data aggregation
US7613748B2 (en) 2003-11-13 2009-11-03 Commvault Systems, Inc. Stored data reverification management system and method
US7272606B2 (en) 2003-11-26 2007-09-18 Veritas Operating Corporation System and method for detecting and storing file content access information within a file system
US7519726B2 (en) * 2003-12-12 2009-04-14 International Business Machines Corporation Methods, apparatus and computer programs for enhanced access to resources within a network
US7200621B2 (en) 2003-12-17 2007-04-03 International Business Machines Corporation System to automate schema creation for table restore
JP4581518B2 (en) * 2003-12-19 2010-11-17 株式会社日立製作所 How to get a snapshot
US7103740B1 (en) 2003-12-31 2006-09-05 Veritas Operating Corporation Backup mechanism for a multi-class file system
US7747659B2 (en) * 2004-01-05 2010-06-29 International Business Machines Corporation Garbage collector with eager read barrier
US7246272B2 (en) 2004-01-16 2007-07-17 International Business Machines Corporation Duplicate network address detection
US7685384B2 (en) 2004-02-06 2010-03-23 Globalscape, Inc. System and method for replicating files in a computer network
US7627617B2 (en) 2004-02-11 2009-12-01 Storage Technology Corporation Clustered hierarchical file services
US7200604B2 (en) 2004-02-17 2007-04-03 Hewlett-Packard Development Company, L.P. Data de-duplication
GB0404444D0 (en) 2004-02-27 2004-09-01 Bae Sys Defence Sys Ltd Secure computer communication
US20080082736A1 (en) * 2004-03-11 2008-04-03 Chow David Q Managing bad blocks in various flash memory cells for electronic data flash card
US20050203887A1 (en) 2004-03-12 2005-09-15 Solix Technologies, Inc. System and method for seamless access to multiple data sources
WO2005097439A2 (en) 2004-04-01 2005-10-20 Christian Schmidt Manufacturing and use of microperforated substrates
US20050234823A1 (en) 2004-04-20 2005-10-20 Rainer Schimpf Systems and methods to prevent products from counterfeiting and surplus production also of tracking their way of distribution.
US7590639B1 (en) 2004-04-29 2009-09-15 Sap Ag System and method for ordering a database flush sequence at transaction commit
CA2564967C (en) 2004-04-30 2014-09-30 Commvault Systems, Inc. Hierarchical systems and methods for providing a unified view of storage information
US7778962B2 (en) 2004-04-30 2010-08-17 Microsoft Corporation Client store synchronization through intermediary store change packets
US7676590B2 (en) 2004-05-03 2010-03-09 Microsoft Corporation Background transcoding
US8108429B2 (en) 2004-05-07 2012-01-31 Quest Software, Inc. System for moving real-time data events across a plurality of devices in a network for simultaneous data protection, replication, and access services
US20050254072A1 (en) 2004-05-12 2005-11-17 Canon Kabushiki Kaisha Image data processing method, client terminal, image processing program, image data management method and image management system
US7734581B2 (en) 2004-05-18 2010-06-08 Oracle International Corporation Vector reads for array updates
US8055745B2 (en) * 2004-06-01 2011-11-08 Inmage Systems, Inc. Methods and apparatus for accessing data from a primary data storage system for secondary storage
DE502004001164D1 (en) 2004-06-02 2006-09-21 Ixos Software Ag Method and device for managing electronic messages
US7330997B1 (en) 2004-06-03 2008-02-12 Gary Odom Selective reciprocal backup
US7383462B2 (en) * 2004-07-02 2008-06-03 Hitachi, Ltd. Method and apparatus for encrypted remote copy for secure data backup and restoration
JP4477950B2 (en) * 2004-07-07 2010-06-09 株式会社日立製作所 Remote copy system and storage device system
US7617297B2 (en) 2004-07-26 2009-11-10 International Business Machines Corporation Providing archiving of individual mail content while maintaining a single copy mail store
JP2006065999A (en) * 2004-08-30 2006-03-09 Fujitsu Ltd Magnetic tape control apparatus, and method and program therefor
US7546335B2 (en) 2004-09-02 2009-06-09 Broadway Technology, Llc System and method for a data protocol layer and the transfer of data objects using the data protocol layer
US7865470B2 (en) 2004-09-09 2011-01-04 Microsoft Corporation Method, system, and apparatus for translating logical information representative of physical data in a data protection system
US7631194B2 (en) 2004-09-09 2009-12-08 Microsoft Corporation Method, system, and apparatus for creating saved searches and auto discovery groups for a data protection system
US7786881B2 (en) 2004-09-17 2010-08-31 Koninklijke Philips Electronics N.V. Content status provision related to volatile memories
US20060095470A1 (en) 2004-11-04 2006-05-04 Cochran Robert A Managing a file in a network environment
WO2006052830A2 (en) * 2004-11-05 2006-05-18 Trusted Data Corporation Storage system condition indicator and method
WO2006053084A2 (en) 2004-11-05 2006-05-18 Commvault Systems, Inc. Method and system of pooling storage devices
US7490207B2 (en) * 2004-11-08 2009-02-10 Commvault Systems, Inc. System and method for performing auxillary storage operations
JP4349301B2 (en) 2004-11-12 2009-10-21 日本電気株式会社 Storage management system, method and program
ATE405051T1 (en) * 2004-11-22 2008-08-15 Research In Motion Ltd METHOD FOR SECURE REDUNDANCY ADDITION OF AN ELECTRONIC MESSAGE
US7512135B2 (en) 2004-12-10 2009-03-31 International Business Machines Corporation Method for transferring data among a logical layer, physical layer, and storage device
US20060129771A1 (en) 2004-12-14 2006-06-15 International Business Machines Corporation Managing data migration
US7721292B2 (en) * 2004-12-16 2010-05-18 International Business Machines Corporation System for adjusting resource allocation to a logical partition based on rate of page swaps and utilization by changing a boot configuration file
US7493314B2 (en) * 2005-01-10 2009-02-17 Cyberlink Corp. System and method for providing access to computer files across computer operating systems
US7451166B2 (en) 2005-01-13 2008-11-11 International Business Machines Corporation System and method for maintaining checkpoints of a keyed data structure using a sequential log
US7506010B2 (en) * 2005-02-08 2009-03-17 Pro Softnet Corporation Storing and retrieving computer data files using an encrypted network drive file system
US7457835B2 (en) 2005-03-08 2008-11-25 Cisco Technology, Inc. Movement of data in a distributed database system to a storage location closest to a center of activity for the data
US7966495B2 (en) 2005-03-21 2011-06-21 Revinetix, Inc. Conserving file system with backup and validation
EP1708095A1 (en) 2005-03-31 2006-10-04 Ubs Ag Computer network system for constructing, synchronizing and/or managing a second database from/with a first database, and methods therefore
US7320059B1 (en) * 2005-08-26 2008-01-15 Emc Corporation Methods and apparatus for deleting content from a storage system
US20070067399A1 (en) 2005-09-22 2007-03-22 Raghavendra Kulkarni Electronic mail archiving system and method
US20070079170A1 (en) 2005-09-30 2007-04-05 Zimmer Vincent J Data migration in response to predicted disk failure
US7551572B2 (en) 2005-10-21 2009-06-23 Isilon Systems, Inc. Systems and methods for providing variable protection
US7409522B1 (en) * 2005-10-26 2008-08-05 Network Appliance, Inc. Method and system for reallocating data in a file system
US7716445B2 (en) * 2005-11-04 2010-05-11 Oracle America, Inc. Method and system for storing a sparse file using fill counts
US7532979B2 (en) 2005-11-10 2009-05-12 Tele Atlas North America, Inc. Method and system for creating universal location referencing objects
US7899990B2 (en) * 2005-11-15 2011-03-01 Oracle America, Inc. Power conservation via DRAM access
US20070118573A1 (en) 2005-11-23 2007-05-24 Solix, Inc. System and method to create a subset of a database
US7747579B2 (en) 2005-11-28 2010-06-29 Commvault Systems, Inc. Metabase for facilitating data classification
US20070136200A1 (en) 2005-12-09 2007-06-14 Microsoft Corporation Backup broker for private, integral and affordable distributed storage
ES2582364T3 (en) * 2005-12-19 2016-09-12 Commvault Systems, Inc. Systems and methods to perform data replication
US20070156998A1 (en) * 2005-12-21 2007-07-05 Gorobets Sergey A Methods for memory allocation in non-volatile memories with a directly mapped file storage system
US7831793B2 (en) 2006-03-01 2010-11-09 Quantum Corporation Data storage system including unique block pool manager and applications in tiered storage
US7603529B1 (en) * 2006-03-22 2009-10-13 Emc Corporation Methods, systems, and computer program products for mapped logical unit (MLU) replications, storage, and retrieval in a redundant array of inexpensive disks (RAID) environment
US7478113B1 (en) * 2006-04-13 2009-01-13 Symantec Operating Corporation Boundaries
US7685459B1 (en) * 2006-04-13 2010-03-23 Symantec Operating Corporation Parallel backup
US7376805B2 (en) * 2006-04-21 2008-05-20 Hewlett-Packard Development Company, L.P. Distributed storage array
US8165221B2 (en) 2006-04-28 2012-04-24 Netapp, Inc. System and method for sampling based elimination of duplicate data
US7512745B2 (en) * 2006-04-28 2009-03-31 International Business Machines Corporation Method for garbage collection in heterogeneous multiprocessor systems
US20080005194A1 (en) 2006-05-05 2008-01-03 Lockheed Martin Corporation System and method for immutably cataloging and storing electronic assets in a large scale computer system
US20070271316A1 (en) 2006-05-22 2007-11-22 I3Archives, Inc. System and method for backing up medical records
US8112396B2 (en) * 2006-06-07 2012-02-07 Emc Corporation Backup and recovery of integrated linked databases
US7480782B2 (en) * 2006-06-14 2009-01-20 Sun Microsystems, Inc. Reference-updating using per-chunk referenced-address ranges in a compacting garbage collector
US7636824B1 (en) 2006-06-28 2009-12-22 Acronis Inc. System and method for efficient backup using hashes
US7921077B2 (en) 2006-06-29 2011-04-05 Netapp, Inc. System and method for managing data deduplication of storage systems utilizing persistent consistency point images
US8412682B2 (en) 2006-06-29 2013-04-02 Netapp, Inc. System and method for retrieving and using block fingerprints for data deduplication
JP4749266B2 (en) * 2006-07-27 2011-08-17 株式会社日立製作所 Backup control apparatus and method without duplication of information resources
US7831707B2 (en) * 2006-08-02 2010-11-09 Scenera Technologies, Llc Methods, systems, and computer program products for managing electronic subscriptions
US7747584B1 (en) * 2006-08-22 2010-06-29 Netapp, Inc. System and method for enabling de-duplication in a storage system architecture
US20080082714A1 (en) * 2006-09-29 2008-04-03 Nasa Hq's. Systems, methods and apparatus for flash drive
US7685177B1 (en) * 2006-10-03 2010-03-23 Emc Corporation Detecting and managing orphan files between primary and secondary data stores
US8078603B1 (en) 2006-10-05 2011-12-13 Blinkx Uk Ltd Various methods and apparatuses for moving thumbnails
US7493456B2 (en) * 2006-10-13 2009-02-17 International Business Machines Corporation Memory queue with supplemental locations for consecutive addresses
US7882077B2 (en) 2006-10-17 2011-02-01 Commvault Systems, Inc. Method and system for offline indexing of content and classifying stored data
US8635194B2 (en) * 2006-10-19 2014-01-21 Oracle International Corporation System and method for data compression
US9465823B2 (en) * 2006-10-19 2016-10-11 Oracle International Corporation System and method for data de-duplication
US7920700B2 (en) * 2006-10-19 2011-04-05 Oracle International Corporation System and method for data encryption
US20080162321A1 (en) 2006-11-07 2008-07-03 Breeden Benjamin T System and method for processing duplicative electronic check return files
US8875199B2 (en) * 2006-11-13 2014-10-28 Cisco Technology, Inc. Indicating picture usefulness for playback optimization
JP4751814B2 (en) * 2006-11-28 2011-08-17 富士通東芝モバイルコミュニケーションズ株式会社 Mobile device
JP2008134966A (en) 2006-11-29 2008-06-12 Sony Corp Data management server, data management system, data management method and program
CA2705379C (en) 2006-12-04 2016-08-30 Commvault Systems, Inc. Systems and methods for creating copies of data, such as archive copies
US7840537B2 (en) 2006-12-22 2010-11-23 Commvault Systems, Inc. System and method for storing redundant information
EP2102751B1 (en) 2006-12-22 2013-11-20 Commvault Systems, Inc. System and method for storing redundant information
US8099386B2 (en) 2006-12-27 2012-01-17 Research In Motion Limited Method and apparatus for synchronizing databases connected by wireless interface
US20080162518A1 (en) 2007-01-03 2008-07-03 International Business Machines Corporation Data aggregation and grooming in multiple geo-locations
US7788230B2 (en) 2007-01-23 2010-08-31 International Business Machines Corporation Backing-up and restoring files including files referenced with multiple file names
US7870486B2 (en) * 2007-01-26 2011-01-11 Kabushiki Kaisha Toshiba System and method for simultaneously commencing output of disparately encoded electronic documents
US7853750B2 (en) 2007-01-30 2010-12-14 Netapp, Inc. Method and an apparatus to store data patterns
US7672981B1 (en) 2007-02-28 2010-03-02 Emc Corporation Object classification and indexing of very large name spaces using grid technology
US7769971B2 (en) 2007-03-29 2010-08-03 Data Center Technologies Replication and restoration of single-instance storage pools
US7873809B2 (en) * 2007-03-29 2011-01-18 Hitachi, Ltd. Method and apparatus for de-duplication after mirror operation
US20080243769A1 (en) * 2007-03-30 2008-10-02 Symantec Corporation System and method for exporting data directly from deduplication storage to non-deduplication storage
US8019016B1 (en) 2007-04-27 2011-09-13 Marvell International Ltd. System and method of transmit beam selection
US8315984B2 (en) 2007-05-22 2012-11-20 Netapp, Inc. System and method for on-the-fly elimination of redundant data
US8099392B2 (en) * 2007-06-08 2012-01-17 Apple Inc. Electronic backup of applications
US7631155B1 (en) * 2007-06-30 2009-12-08 Emc Corporation Thin provisioning of a file system and an iSCSI LUN through a common mechanism
US20090012984A1 (en) * 2007-07-02 2009-01-08 Equivio Ltd. Method for Organizing Large Numbers of Documents
US8028106B2 (en) 2007-07-06 2011-09-27 Proster Systems, Inc. Hardware acceleration of commonality factoring with removable media
US20090049260A1 (en) 2007-08-13 2009-02-19 Upadhyayula Shivarama Narasimh High performance data deduplication in a virtual tape system
US8315999B2 (en) * 2007-08-29 2012-11-20 Nirvanix, Inc. Policy-based file management for a storage delivery network
US8880797B2 (en) * 2007-09-05 2014-11-04 Emc Corporation De-duplication in a virtualized server environment
US8788457B2 (en) 2007-09-21 2014-07-22 International Business Machines Corporation Ensuring that the archival data deleted in relational source table is already stored in relational target table
JP5331323B2 (en) 2007-09-26 2013-10-30 株式会社日立製作所 Storage subsystem and control method thereof
JP2009080671A (en) 2007-09-26 2009-04-16 Hitachi Ltd Computer system, management computer and file management method
US8170994B2 (en) 2007-09-28 2012-05-01 Symantec Corporation Techniques for virtual archiving
JP5026213B2 (en) 2007-09-28 2012-09-12 株式会社日立製作所 Storage apparatus and data deduplication method
US20090106369A1 (en) * 2007-10-18 2009-04-23 Yen-Fu Chen Duplicate email address detection for a contact
US8432797B2 (en) * 2007-10-24 2013-04-30 Motorola Solutions, Inc. Method and system for distributed admission control in mobile ad hoc networks (MANETs)
US8782368B2 (en) 2007-10-25 2014-07-15 Hewlett-Packard Development Company, L.P. Storing chunks in containers
US20090112870A1 (en) * 2007-10-31 2009-04-30 Microsoft Corporation Management of distributed storage
US20090119678A1 (en) 2007-11-02 2009-05-07 Jimmy Shih Systems and methods for supporting downloadable applications on a portable client device
US8548953B2 (en) 2007-11-12 2013-10-01 F5 Networks, Inc. File deduplication using storage tiers
US20090204650A1 (en) * 2007-11-15 2009-08-13 Attune Systems, Inc. File Deduplication using Copy-on-Write Storage Tiers
US7870105B2 (en) * 2007-11-20 2011-01-11 Hitachi, Ltd. Methods and apparatus for deduplication in storage system
US20090150498A1 (en) * 2007-12-07 2009-06-11 Steven Joseph Branda Identifying a Plurality of Related Electronic Messages and Combining the Plurality of Related Messages Into a Composite View
US7962452B2 (en) 2007-12-28 2011-06-14 International Business Machines Corporation Data deduplication by separating data from meta data
US8190835B1 (en) * 2007-12-31 2012-05-29 Emc Corporation Global de-duplication in shared architectures
US20090193210A1 (en) 2008-01-29 2009-07-30 Hewett Jeffrey R System for Automatic Legal Discovery Management and Data Collection
US20090204636A1 (en) 2008-02-11 2009-08-13 Microsoft Corporation Multimodal object de-duplication
JP4815459B2 (en) 2008-03-06 2011-11-16 株式会社日立製作所 Load balancing control server, load balancing control method, and computer program
US8234444B2 (en) 2008-03-11 2012-07-31 International Business Machines Corporation Apparatus and method to select a deduplication protocol for a data storage library
US7937371B2 (en) * 2008-03-14 2011-05-03 International Business Machines Corporation Ordering compression and deduplication of data
US7519635B1 (en) * 2008-03-31 2009-04-14 International Business Machines Corporation Method of and system for adaptive selection of a deduplication chunking technique
US9395929B2 (en) 2008-04-25 2016-07-19 Netapp, Inc. Network storage server with integrated encryption, compression and deduplication capability
US8346730B2 (en) 2008-04-25 2013-01-01 Netapp. Inc. Deduplication of data on disk devices based on a threshold number of sequential blocks
US8515909B2 (en) 2008-04-29 2013-08-20 International Business Machines Corporation Enhanced method and system for assuring integrity of deduplicated data
US8266114B2 (en) * 2008-09-22 2012-09-11 Riverbed Technology, Inc. Log structured content addressable deduplicating storage
US7890454B2 (en) 2008-05-08 2011-02-15 International Business Machines Corporation Method and system for data disaggregation
US8484162B2 (en) 2008-06-24 2013-07-09 Commvault Systems, Inc. De-duplication systems and methods for application-specific data
US9098495B2 (en) 2008-06-24 2015-08-04 Commvault Systems, Inc. Application-aware and remote single instance data management
US8219524B2 (en) 2008-06-24 2012-07-10 Commvault Systems, Inc. Application-aware and remote single instance data management
US8041907B1 (en) 2008-06-30 2011-10-18 Symantec Operating Corporation Method and system for efficient space management for single-instance-storage volumes
US8176269B2 (en) 2008-06-30 2012-05-08 International Business Machines Corporation Managing metadata for data blocks used in a deduplication system
US8166263B2 (en) * 2008-07-03 2012-04-24 Commvault Systems, Inc. Continuous data protection over intermittent connections, such as continuous data backup for laptops or wireless devices
US8788466B2 (en) * 2008-08-05 2014-07-22 International Business Machines Corporation Efficient transfer of deduplicated data
US8086799B2 (en) 2008-08-12 2011-12-27 Netapp, Inc. Scalable deduplication of stored data
US8239348B1 (en) 2008-08-14 2012-08-07 Symantec Corporation Method and apparatus for automatically archiving data items from backup storage
US10642794B2 (en) * 2008-09-11 2020-05-05 Vmware, Inc. Computer storage deduplication
US8290915B2 (en) * 2008-09-15 2012-10-16 International Business Machines Corporation Retrieval and recovery of data chunks from alternate data stores in a deduplicating system
US8190823B2 (en) 2008-09-18 2012-05-29 Lenovo (Singapore) Pte. Ltd. Apparatus, system and method for storage cache deduplication
US8620845B2 (en) 2008-09-24 2013-12-31 Timothy John Stoakes Identifying application metadata in a backup stream
US9015181B2 (en) * 2008-09-26 2015-04-21 Commvault Systems, Inc. Systems and methods for managing single instancing data
JP2012504266A (en) 2008-09-30 2012-02-16 レインスター リミテッド System and method for data storage
US8495032B2 (en) * 2008-10-01 2013-07-23 International Business Machines Corporation Policy based sharing of redundant data across storage pools in a deduplicating system
US20100088296A1 (en) * 2008-10-03 2010-04-08 Netapp, Inc. System and method for organizing data to facilitate data deduplication
US8626723B2 (en) * 2008-10-14 2014-01-07 Vmware, Inc. Storage-network de-duplication
US8131687B2 (en) 2008-11-13 2012-03-06 International Business Machines Corporation File system with internal deduplication and management of data blocks
US8751462B2 (en) * 2008-11-14 2014-06-10 Emc Corporation Delta compression after identity deduplication
US8412677B2 (en) 2008-11-26 2013-04-02 Commvault Systems, Inc. Systems and methods for byte-level or quasi byte-level single instancing
US8131809B2 (en) 2008-12-03 2012-03-06 Microsoft Corporation Online archiving of message objects
CN104166673B (en) * 2008-12-22 2017-09-19 谷歌公司 Asynchronous distributed duplicate removal for reproducting content addressable storage cluster
US8161255B2 (en) * 2009-01-06 2012-04-17 International Business Machines Corporation Optimized simultaneous storing of data into deduplicated and non-deduplicated storage pools
US8401996B2 (en) 2009-03-30 2013-03-19 Commvault Systems, Inc. Storing a variable number of instances of data objects
US8095756B1 (en) 2009-04-28 2012-01-10 Netapp, Inc. System and method for coordinating deduplication operations and backup operations of a storage volume
US8195636B2 (en) * 2009-04-29 2012-06-05 Netapp, Inc. Predicting space reclamation in deduplicated datasets
US8578120B2 (en) 2009-05-22 2013-11-05 Commvault Systems, Inc. Block-level single instancing
US20100333116A1 (en) 2009-06-30 2010-12-30 Anand Prahlad Cloud gateway system for managing data storage to cloud storage sites
US8930306B1 (en) 2009-07-08 2015-01-06 Commvault Systems, Inc. Synchronized data deduplication
US8244914B1 (en) 2009-07-31 2012-08-14 Symantec Corporation Systems and methods for restoring email databases
EP2336910B1 (en) 2009-11-23 2018-02-21 BlackBerry Limited Generating device specific thumbnails
US8965852B2 (en) 2009-11-24 2015-02-24 Dell Products L.P. Methods and apparatus for network efficient deduplication
WO2012025915A1 (en) 2010-07-21 2012-03-01 Sqream Technologies Ltd A system and method for the parallel execution of database queries over cpus and multi core processors
US8364652B2 (en) 2010-09-30 2013-01-29 Commvault Systems, Inc. Content aligned block-based deduplication
US8572340B2 (en) 2010-09-30 2013-10-29 Commvault Systems, Inc. Systems and methods for retaining and using data block signatures in data protection operations
WO2012045023A2 (en) 2010-09-30 2012-04-05 Commvault Systems, Inc. Archiving data objects using secondary copies
US9720606B2 (en) 2010-10-26 2017-08-01 Avago Technologies General Ip (Singapore) Pte. Ltd. Methods and structure for online migration of data in storage systems comprising a plurality of storage devices
US8396841B1 (en) * 2010-11-30 2013-03-12 Symantec Corporation Method and system of multi-level and multi-mode cloud-based deduplication
US8954446B2 (en) 2010-12-14 2015-02-10 Comm Vault Systems, Inc. Client-side repository in a networked deduplicated storage system
US9020900B2 (en) 2010-12-14 2015-04-28 Commvault Systems, Inc. Distributed deduplicated storage system
US20120159098A1 (en) 2010-12-17 2012-06-21 Microsoft Corporation Garbage collection and hotspots relief for a data deduplication chunk store
US9823981B2 (en) 2011-03-11 2017-11-21 Microsoft Technology Licensing, Llc Backup and restore strategies for data deduplication
CN102799598A (en) 2011-05-25 2012-11-28 英业达股份有限公司 Data recovery method for deleting repeated data
US8954967B2 (en) 2011-05-31 2015-02-10 International Business Machines Corporation Adaptive parallel data processing
US8533231B2 (en) 2011-08-12 2013-09-10 Nexenta Systems, Inc. Cloud storage system with distributed metadata
US8521704B2 (en) 2011-09-30 2013-08-27 Symantec Corporation System and method for filesystem deduplication using variable length sharing
US9116633B2 (en) 2011-09-30 2015-08-25 Commvault Systems, Inc. Information management of virtual machines having mapped storage devices
US9715434B1 (en) * 2011-09-30 2017-07-25 EMC IP Holding Company LLC System and method for estimating storage space needed to store data migrated from a source storage to a target storage
US9372479B1 (en) 2012-02-21 2016-06-21 Omniboard, Inc. System and method for a database layer for managing a set of energy consuming devices
US9367057B2 (en) 2012-02-21 2016-06-14 Omniboard, Inc. System and method for real-time controls of energy consuming devices including tiered architecture
US8732403B1 (en) * 2012-03-14 2014-05-20 Netapp, Inc. Deduplication of data blocks on storage devices
US9020890B2 (en) 2012-03-30 2015-04-28 Commvault Systems, Inc. Smart archiving and data previewing for mobile devices
US9703500B2 (en) 2012-04-25 2017-07-11 International Business Machines Corporation Reducing power consumption by migration of data within a tiered storage system
US20130339310A1 (en) 2012-06-13 2013-12-19 Commvault Systems, Inc. Restore using a client side signature repository in a networked storage system
US8862571B2 (en) 2012-06-29 2014-10-14 International Business Machines Corporation Predicate pushdown with late materialization in database query processing
US9792320B2 (en) 2012-07-06 2017-10-17 Box, Inc. System and method for performing shard migration to support functions of a cloud-based service
US9348538B2 (en) * 2012-10-18 2016-05-24 Netapp, Inc. Selective deduplication
US20140129961A1 (en) 2012-11-07 2014-05-08 Sergey Mikhailovich Zubarev Tool for managing user task information
US20140188532A1 (en) 2012-11-13 2014-07-03 Nec Laboratories America, Inc. Multitenant Database Placement with a Cost Based Query Scheduler
US9146960B2 (en) 2012-12-20 2015-09-29 Teradata Us, Inc. Adaptive optimization of iterative or recursive query execution by database systems
US9633022B2 (en) 2012-12-28 2017-04-25 Commvault Systems, Inc. Backup and restoration for a deduplicated file system
US9633033B2 (en) 2013-01-11 2017-04-25 Commvault Systems, Inc. High availability distributed deduplicated storage system
US9740741B2 (en) 2013-04-11 2017-08-22 Hasso-Plattner-Institut Fur Softwaresystemtechnik Gmbh Aggregate query-caching in databases architectures with a differential buffer and a main store
US9646166B2 (en) 2013-08-05 2017-05-09 International Business Machines Corporation Masking query data access pattern in encrypted data
IN2013MU04016A (en) 2013-12-23 2015-08-07 Tata Consultancy Services Ltd
US10324897B2 (en) 2014-01-27 2019-06-18 Commvault Systems, Inc. Techniques for serving archived electronic mail
US9633056B2 (en) 2014-03-17 2017-04-25 Commvault Systems, Inc. Maintaining a deduplication database
US10380072B2 (en) * 2014-03-17 2019-08-13 Commvault Systems, Inc. Managing deletions from a deduplication database
US9276871B1 (en) 2014-03-20 2016-03-01 Cisco Technology, Inc. LISP stretched subnet mode for data center migrations
US9575673B2 (en) 2014-10-29 2017-02-21 Commvault Systems, Inc. Accessing a file system using tiered deduplication
US9612766B2 (en) 2014-12-19 2017-04-04 Oracle International Corporation Systems and methods for shadow migration progress estimation
US9781000B1 (en) 2014-12-22 2017-10-03 EMC IP Holding Company LLC Storage mobility using locator-identifier separation protocol
US10108687B2 (en) 2015-01-21 2018-10-23 Commvault Systems, Inc. Database protection using block-level mapping
US10339106B2 (en) 2015-04-09 2019-07-02 Commvault Systems, Inc. Highly reusable deduplication database after disaster recovery
US9213540B1 (en) 2015-05-05 2015-12-15 Archive Solutions Providers Automated workflow management system for application and data retirement
US10324914B2 (en) 2015-05-20 2019-06-18 Commvalut Systems, Inc. Handling user queries against production and archive storage systems, such as for enterprise customers having large and/or numerous files
US20160350391A1 (en) 2015-05-26 2016-12-01 Commvault Systems, Inc. Replication using deduplicated secondary copy data
US9940337B2 (en) * 2015-05-31 2018-04-10 Vmware, Inc. Predictive probabilistic deduplication of storage
US20170038978A1 (en) * 2015-08-05 2017-02-09 HGST Netherlands B.V. Delta Compression Engine for Similarity Based Data Deduplication
US10592357B2 (en) 2015-12-30 2020-03-17 Commvault Systems, Inc. Distributed file system in a distributed deduplication data storage system
US10146697B1 (en) * 2016-09-26 2018-12-04 EMC IP Holding Company LLC NUMA-aware perfect hash algorithm
US10558646B2 (en) * 2017-04-30 2020-02-11 International Business Machines Corporation Cognitive deduplication-aware data placement in large scale storage systems
US10795812B1 (en) * 2017-06-30 2020-10-06 EMC IP Holding Company LLC Virtual copy forward method and system for garbage collection in cloud computing networks
US20190192978A1 (en) 2017-12-27 2019-06-27 Josiah Eatedali Video game group dynamic building
US11010258B2 (en) 2018-11-27 2021-05-18 Commvault Systems, Inc. Generating backup copies through interoperability between components of a data storage management system and appliances for data storage and deduplication
US11507305B2 (en) * 2019-03-29 2022-11-22 EMC IP Holding Company LLC Concurrently performing normal system operations and garbage collection
US20200327017A1 (en) 2019-04-10 2020-10-15 Commvault Systems, Inc. Restore using deduplicated secondary copy data
US11463264B2 (en) 2019-05-08 2022-10-04 Commvault Systems, Inc. Use of data block signatures for monitoring in an information management system
US10664165B1 (en) * 2019-05-10 2020-05-26 EMC IP Holding Company LLC Managing inline data compression and deduplication in storage systems
US11533063B2 (en) * 2019-08-01 2022-12-20 EMC IP Holding Company LLC Techniques for determining compression tiers and using collected compression hints
US11687424B2 (en) * 2020-05-28 2023-06-27 Commvault Systems, Inc. Automated media agent state management

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6928459B1 (en) * 2000-07-18 2005-08-09 International Business Machines Corporation Plurality of file systems using weighted allocation to allocate space on one or more storage devices
US20020112022A1 (en) * 2000-12-18 2002-08-15 Spinnaker Networks, Inc. Mechanism for handling file level and block level remote file accesses using the same server
US20050210460A1 (en) * 2004-03-22 2005-09-22 Microsoft Corporation Computing device with relatively limited storage space and operating/file system thereof
US20090106480A1 (en) * 2007-10-23 2009-04-23 Keicy Chung Computer storage device having separate read-only space and read-write space, removable media component, system management interface, and network interface

Also Published As

Publication number Publication date
WO2010135136A3 (en) 2011-02-17
US11455212B2 (en) 2022-09-27
US20210263803A1 (en) 2021-08-26
US11709739B2 (en) 2023-07-25
US20140040582A1 (en) 2014-02-06
US20190179713A1 (en) 2019-06-13
US8578120B2 (en) 2013-11-05
US20230367678A1 (en) 2023-11-16
US20220382643A1 (en) 2022-12-01
US20100299490A1 (en) 2010-11-25
US10956274B2 (en) 2021-03-23
US9058117B2 (en) 2015-06-16
US20150199242A1 (en) 2015-07-16

Similar Documents

Publication Publication Date Title
US11709739B2 (en) Block-level single instancing
US11640338B2 (en) Data recovery operations, such as recovery from modified network data management protocol data
US11768800B2 (en) Archiving data objects using secondary copies
US11016858B2 (en) Systems and methods for managing single instancing data
US11593217B2 (en) Systems and methods for managing single instancing data
US10970304B2 (en) Storing a variable number of instances of data objects
AU2013206404B2 (en) Systems and methods for managing single instancing data

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10778136

Country of ref document: EP

Kind code of ref document: A2

122 Ep: pct application non-entry in european phase

Ref document number: 10778136

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE