US20120054264A1 - Techniques for Migrating Active I/O Connections with Migrating Servers and Clients - Google Patents

Techniques for Migrating Active I/O Connections with Migrating Servers and Clients Download PDF

Info

Publication number
US20120054264A1
US20120054264A1 US12/872,077 US87207710A US2012054264A1 US 20120054264 A1 US20120054264 A1 US 20120054264A1 US 87207710 A US87207710 A US 87207710A US 2012054264 A1 US2012054264 A1 US 2012054264A1
Authority
US
United States
Prior art keywords
client
virtualized
server
connection
storage access
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/872,077
Inventor
Julianne F. Haugh
Gerald F. McBrearty
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Airbnb Inc
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US12/872,077 priority Critical patent/US20120054264A1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HAUGH, JULIANNE F., MCBREARTY, GERALD F.
Publication of US20120054264A1 publication Critical patent/US20120054264A1/en
Priority to US13/450,633 priority patent/US8898224B2/en
Priority to US14/510,983 priority patent/US9645764B2/en
Assigned to Airbnb, Inc. reassignment Airbnb, Inc. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERNATIONAL BUSINESS MACHINES CORPORATION
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0628Interfaces specially adapted for storage systems making use of a particular technique
    • G06F3/0655Vertical data movement, i.e. input-output transfer; data movement between one or more hosts and one or more storage devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5077Logical partitioning of resources; Management or configuration of virtualized resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0602Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
    • G06F3/0604Improving or facilitating administration, e.g. storage management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/06Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
    • G06F3/0601Interfaces specially adapted for storage systems
    • G06F3/0668Interfaces specially adapted for storage systems adopting a particular infrastructure
    • G06F3/067Distributed or networked storage systems, e.g. storage area networks [SAN], network attached storage [NAS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]

Definitions

  • This disclosure relates generally to migrating active input/output (I/O) connections and, more particularly to techniques for migrating active I/O connections with migrating virtualized I/O servers and/or I/O clients.
  • I/O active input/output
  • IB InfiniBand
  • FC Fibre channel
  • PCI peripheral component interconnect
  • ATA serial advanced technology attachment
  • IB offers point-to-point bidirectional serial links for connecting processors with high-speed peripherals, e.g., disks.
  • FC is widely used as the standard connection type for storage area networks (SANs).
  • FC protocol (FCP) is a transport protocol (similar to transmission control protocol (TCP) used in Internet protocol (IP) networks) that predominantly transports small computer system interface (SCSI) commands over FC networks.
  • TCP transmission control protocol
  • IP Internet protocol
  • SCSI small computer system interface
  • FC supports a number of upper layer protocols, including SCSI, asynchronous transfer mode (ATM), and IP.
  • ATM asynchronous transfer mode
  • IP asynchronous transfer mode
  • the SCSI standards define commands, protocols, and electrical and optical interfaces. SCSI is most commonly used for hard disks and tape drives, but can connect a wide range of other devices, e.g., scanners and compact disc (CD) drives.
  • iSCSI Internet SCSI
  • iSCSI Internet SCSI
  • LANs local area networks
  • WANs wide area networks
  • iSCSI Internet SCSI
  • iSCSI is a popular SAN protocol that allows organizations to consolidate storage into data center storage arrays, while providing hosts (such as database and web servers) with the illusion of locally-attached disks.
  • a SAN architecture is typically used to attach remote computer storage devices (e.g., disk arrays, tape libraries, and optical jukeboxes) to servers in such a way that the devices appear locally attached, with respect to an operating system (OS).
  • OS operating system
  • a SAN only provides block-level operations, as contrasted with file abstraction.
  • file systems may be built on top of SANs to provide file abstraction.
  • network attached storage uses file-based protocols, where it is clear that the storage is remote and computers request a portion of an abstract file rather than a disk block.
  • NAS provides both storage and a file system, in contrast to SANs, which typically provide only block-based storage and leave file system concerns to the client-side.
  • Protocols employed in SANs include SCSI, FC, iSCSI, advanced technology attachment (ATA), and ATA over Ethernet (AoE), among other protocols.
  • SCSI SCSI
  • FC Fibre Channel
  • iSCSI advanced technology attachment
  • ATA ATA over Ethernet
  • AoE ATA over Ethernet
  • I/O virtualization is a methodology that is typically implemented to simplify management, lower costs, and improve performance of servers.
  • I/O virtualization environments are created by abstracting upper layer protocols from physical connections.
  • I/O virtualization enables one physical adapter card to appear as multiple virtual network interface cards (vNICs) and virtual host bus adapters (vHBAs).
  • vNICs and vHBAs function as conventional NICs and HBAs and are designed to be compatible with existing OSs, hypervisors, and applications.
  • To networking resources e.g., LANs and SANs
  • the vNICs and vHBAs appear as normal cards and adapters.
  • Virtual I/O provides a shared transport for all network and storage connections.
  • Virtual I/O addresses performance bottlenecks by consolidating I/O to a single connection, whose bandwidth ideally exceeds the I/O capacity of an associated server (to ensure that an associated I/O link is not a bottleneck). Bandwidth is then dynamically allocated in real-time across multiple virtual connections to both storage and network resources. In I/O intensive applications, I/O virtualization can help increase both virtual machine (VM) performance and the number of VMs that may execute on a given server.
  • VM virtual machine
  • Storage virtualization refers to the process of abstracting logical storage from physical storage. In general, physical storage resources are aggregated into storage pools, from which logical storage is created. Storage virtualization presents a logical space for data storage and transparently handles the process of mapping the logical space to the actual physical location. Storage virtualization is typically implemented in modern disk arrays using vendor proprietary solutions. However, the goal of storage virtualization is usually to virtualize multiple disk arrays from different vendors, scattered over a network, into a single monolithic storage device, which can be managed uniformly.
  • File systems which are the most common means of accessing disk storage, are abstractions of a physical storage object (e.g., rotating magnetic media, solid state electronic devices, directory structured magnetic tapes) into a more human friendly format, where data can be organized into files, folders and other similar objects.
  • OSs including those derived from the UNIX or Windows OSs, abstract the final connection between the file system into, for example, a device switch, a device array, or a device control block.
  • Each physical device, as well as each type of physical device (which may also include network connected storage, virtualized disk storage, etc.) has a different handle, identifier, or other type of object.
  • These data structures or objects include the information needed by the OS or disk I/O subsystem to convert requests for disk I/O into actual data transfer (i.e., either reading or writing).
  • a technique for modifying an input/output (I/O) connection type between a client and a server includes determining, by the client, when at least partial storage access control for the client is to be migrated from a first virtualized I/O server to a second virtualized I/O server.
  • An I/O connection type from among multiple available I/O connection types available between the client and the second virtualized I/O server, is then selected by the client, when the at least partial storage access control for the client is to be migrated.
  • the selected I/O connection type is then communicated, by the client, to the second virtualized I/O server, which is configured to facilitate storage access for the client using the selected I/O connection type.
  • FIG. 1 is a block diagram of an example data processing environment that may be configured to migrate active input/output (I/O) connections according to various aspects of the present disclosure.
  • FIG. 2 is a block diagram of another example data processing environment that may be configured to migrate active I/O connections according to various embodiments of the present disclosure.
  • FIG. 3 is a flowchart of an example process for migrating an active I/O connection for a client according to various aspects of the present disclosure.
  • FIG. 4 is a flowchart of an example process for migrating active I/O connections for clients in a client cluster according to various aspects of the present disclosure.
  • aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • the term “coupled” includes both a direct electrical connection between blocks or components and an indirect electrical connection between blocks or components achieved using one or more intervening blocks or components.
  • I/O input/output
  • logical partitions, network attached storage (NAS), and/or storage area network (SAN) devices are migrated opportunities for improving I/O connections may exist when, for example, a new physical location of an I/O client is closer to a new physical location of a migrated virtualized I/O server.
  • NAS network attached storage
  • SAN storage area network
  • each time an I/O client e.g., a disk I/O client
  • a virtualized I/O server e.g., a virtualized disk I/O server
  • related endpoints i.e., a server in the case of a client migrating or a client in the case of a server migrating
  • a client may query an I/O server as to which I/O connection types are possible given an Internet protocol (IP) address.
  • IP Internet protocol
  • client and server data structures may be modified to, for example, select the most efficient (e.g., in terms of data transfer speed) I/O connection type available.
  • disk I/O requests are suspended pending the I/O connection type modification (which may be an upgrade or a downgrade in the I/O connection type), with the I/O connection type modification being considered both prior to migration and once migration is complete.
  • I/O connection changes e.g., local to network
  • I/O connection type changes e.g., network to local
  • disk I/O may resume on a connection-by-connection basis.
  • a new I/O connection may be established between a new virtualized I/O server and an I/O client, while an old I/O connection between the I/O client and an old virtualized I/O server is maintained.
  • new I/Os are sent to the new virtualized I/O server, while the old virtualized I/O server completes service of the old I/Os.
  • the old virtualized I/O server finishes servicing the old I/Os, the old I/O connection is closed. In this manner, I/O access downtime is reduced (and possibly eliminated) and rebooting of the I/O client is avoided.
  • an I/O connection may be upgraded or downgraded when an endpoint (server or client) is moved from its origin.
  • migration of an endpoint results in either a network connection (e.g., an IP address of an iSCSI device changes) or a physical I/O connection (e.g., an available adapter changes from: FC to SCSI; FC to iSCSI; FC to IB; IB to FC; IB to SCSI; IB to iSCSI; iSCSI to SCSI) being changed.
  • a network connection e.g., an IP address of an iSCSI device changes
  • a physical I/O connection e.g., an available adapter changes from: FC to SCSI; FC to iSCSI; FC to IB; IB to FC; IB to SCSI; IB to iSCSI; iSCSI to SCSI
  • a network connection might be changed by migrating a virtual I/O server to a different virtual host (which is not local to a previous connection, or which becomes local to the previous connection).
  • Each type of change may result in a physical type I/O connection being changed (e.g., I/O connections where physical blocks are specified, rather than logical connections where the physical I/O structure is concealed, such as with network file system (NFS), common Internet file system (CIFS), and similar mechanisms).
  • NFS network file system
  • CIFS common Internet file system
  • an I/O connection type may change from iSCSI to SCSI or from one network file system type (e.g., network attached storage (NAS), SAN, or other block-structured I/O mechanism) to another, as a result of an I/O server being migrated.
  • NAS network attached storage
  • SAN or other block-structured I/O mechanism
  • U.S. Pat. No. 7,219,096 discloses a method for migrating files on a busy file system, with the migrated files being recreated on a different file system.
  • the '096 patent does not address the case where there is no desire to break a client/server connection, but rather to change characteristics of the connection to one which is more efficient (if possible), or less efficient (if required to maintain the client/server connection).
  • U.S. Pat. No. 6,976,103 (hereinafter “the '103 patent”) discloses a method for migrating storage.
  • the '103 patent does not address modifying a connection type when a virtualized I/O server (or network disk I/O service) is migrated.
  • U.S. Pat. No. 6,931,441 discloses a method for migrating an actual network connection by means of an encapsulating mechanism.
  • network communication is made between capsules, with the capsules handling changes in network addressing.
  • the '441 patent does not address migrating the physical connection between a disk I/O client and a virtualized disk I/O server.
  • migration as disclosed in the '441 patent may include changing a network address, the network address change is handled by changing the underlying abstracted block device interface, or other such mechanism.
  • IP address 192.168.1.12 if a client moved from IP address 192.168.1.12 to IP address 192.168.0.7 and the client is now located within a virtual host that also contains an associated disk I/O server (e.g., at IP address 192.168.0.2), following the disclosure of the '441 patent, packet routing would be changed without upgrading the I/O connection. While some performance benefit may result from changing the network transport (e.g., a virtual network where traffic on the 192.168.0/24 subnet was not transmitted over physical connections, as with products such as VMware and others), greater benefits are usually provided by removing the software layer which abstracts the physical connection into a network protocol (according to one or more embodiments of the present disclosure).
  • an example data processing environment 100 includes a client 110 and a client 130 that are configured to select (or at least participate in the selection of) an I/O connection type when an associated virtualized I/O server is migrated (e.g., created in a different virtual host) and/or when clients 110 and 130 are re-located (when, for example, clients 110 and 130 are mobile devices).
  • Clients 110 and 130 may take various forms, such as workstations, laptop computer systems, notebook computer systems, smart phones, web-enabled portable devices, or desktop computer systems.
  • client 110 may correspond to a desktop computer system of a computer system user and client 130 may correspond to a web-enabled device of another computer system user.
  • Client 110 includes a processor 102 (which may include one or more processor cores for executing program code) coupled to a data storage subsystem 104 , a display 106 , one or more input devices 108 , and an input/output adapter (IOA) 109 .
  • IOA 109 may be virtualized as: a Fibre Channel (FC) adapter, a small computer system interface (SCSI) adapter, an Internet SCSI (iSCSI) adapter, or an InfiniBand (IB) adapter, among other adapters.
  • FC Fibre Channel
  • SCSI small computer system interface
  • iSCSI Internet SCSI
  • IB InfiniBand
  • Data storage subsystem 104 may include, for example, an application appropriate amount of volatile memory (e.g., dynamic random access memory (DRAM)), non-volatile memory (e.g., read-only memory (ROM) or static RAM), and/or non-volatile mass storage device, such as a magnetic or optical disk drive.
  • Data storage subsystem 104 includes an operating system (OS) 114 for client 110 , as well as application programs, such as a browser 112 (which may optionally include customized plug-ins to support various client applications), application 118 (which may include an email application, a word processing application, etc.), and an I/O connection type application 120 .
  • OS operating system
  • application programs such as a browser 112 (which may optionally include customized plug-ins to support various client applications), application 118 (which may include an email application, a word processing application, etc.), and an I/O connection type application 120 .
  • I/O connection type application 120 when executed by clients 110 and 130 , is configured to interact with a virtualized I/O server (e.g., servers 124 or 126 ) to determine what I/O connection types are available when clients 110 and 130 are migrated and/or an associated virtualized I/O server (e.g., servers 126 or 124 ) is migrated.
  • a virtualized I/O server e.g., servers 124 or 126
  • an associated virtualized I/O server e.g., servers 126 or 124
  • Display 106 may be, for example, a cathode ray tube (CRT) or a liquid crystal display (LCD).
  • Input device(s) 108 of client 110 may include, for example, a mouse, a keyboard, haptic devices, and/or a touch screen.
  • IOA 109 supports communication of client 110 with one or more wired and/or wireless networks utilizing one or more communication protocols, such as 802.x, HTTP, simple mail transfer protocol (SMTP), etc.
  • IOA 109 also facilitates communication between clients 110 and 130 and may physically correspond to, for example, an FC adapter, an IB adapter, an SCSI adapter, or an iSCSI adapter, among other adapters.
  • Clients 110 and 130 are coupled via one or more wired or wireless networks, such as Internet/intranet 122 , to optional storage manager 128 (which may be a SAN manager) and hosts 134 and/or 136 and respective virtualized I/O servers 124 and 126 .
  • Virtualized I/O servers 124 and 126 provide respective access to data stored on disk arrays 115 and 125 .
  • Hosts 124 and 126 may be physically located in, for example, a same room of a building, a different room of a building, a different building in a limited geographical area, or a different geographical area (e.g., different cites, counties, states, or countries).
  • virtualized I/O servers 124 and 126 may each represent six virtualized I/O servers.
  • hosts 134 and 136 may each implement more or less than six virtualized I/O servers and a virtual machine manager (VMM) or hypervisor for controlling operation of the virtualized I/O servers (which may each utilize the same or a different OS), as well as an I/O connection type application that, when executed by hosts 134 and 136 , facilitates communication with client 110 and 130 to determine what I/O connection types are available following migration of clients 110 and 130 and/or an associated virtualized I/O server (e.g., servers 126 or 124 ).
  • VMM virtual machine manager
  • hypervisor for controlling operation of the virtualized I/O servers
  • I/O connection type application that, when executed by hosts 134 and 136 , facilitates communication with client 110 and 130 to determine what I/O connection types are available following migration of clients 110 and 130 and/or an associated virtualized I/O server (e.g., servers 126 or 124 ).
  • Host 134 and array 115 may be included within a NAS or SAN (NAS/SAN) 131 .
  • host 136 and array 125 may be included within a different NAS or SAN (NAS/SAN) 133 .
  • Arrays 115 and 125 may each correspond to, for example, redundant array of inexpensive disks (RAID) arrays.
  • an example data processing environment 200 includes a client cluster 210 that includes multiple clients 110 and a client cluster 230 that includes multiple clients 130 .
  • Clients 110 and 130 are configured to select an I/O connection type when an associated virtualized I/O server is migrated (e.g., to a different virtual host) and/or when clients 110 and 130 are re-located (when, for example, clients 110 and 130 are mobile devices).
  • Clients 110 and 130 are coupled via one or more wired or wireless networks, such as Internet/intranet 122 , to optional storage manager 128 (which may be, for example, a SAN manager) and hosts 134 and/or 136 and respective virtualized I/O servers 124 and 126 .
  • virtualized I/O servers 124 and 126 provide access to data stored on respective disk arrays 115 and 125 .
  • virtualized I/O servers 124 and 126 may each represent one or more virtualized I/O servers.
  • hosts 134 and 136 may each implement more than one virtualized I/O server and a virtual machine manager (VMM) or hypervisor for managing operation of implemented virtualized I/O servers. While only two client clusters are shown associated with data processing environment 200 , it should be appreciated that more or less than two client clusters may be associated with a data processing environment configured according to the present disclosure. In any case, when a virtualized I/O server that is servicing a client cluster is migrated, clients within a client cluster may coordinate between themselves and/or between the clients and the migrated (new) virtualized I/O server to determine what I/O connection type each of the clients may implement. For example, clients in a client cluster may implement different I/O connection types to balance a load on an associated virtualized I/O server.
  • VMM virtual machine manager
  • process 300 migrates active I/O connection types of a client according to various aspects of the present disclosure. It should be appreciated that portions of process 300 may execute, at any given point in time, on clients 110 or 130 , hosts 134 or 136 , optional storage manager 128 , or a combination thereof. Process 300 may be periodically initiated at block 302 by, for example, storage manager 128 , which may be implemented as a stand-alone application on a dedicated server or on hosts 134 and 136 . In block 303 , client 110 accesses storage via an old (first) virtualized I/O server.
  • Process 300 then proceeds to decision block 304 , where client 110 (alone or in conjunction with, for example, storage manager 128 ) determines whether at least partial storage access control requires migration (e.g., due to an addition of an upgraded host (which may be included in a SAN or NAS) or maintenance on a current host and/or a disk array associated therewith).
  • migration e.g., due to an addition of an upgraded host (which may be included in a SAN or NAS) or maintenance on a current host and/or a disk array associated therewith).
  • control transfers to block 310 , where process 300 terminates and control returns to a calling routine.
  • the selected I/O connection type may be the same or different than the I/O connection type employed between the client and the old virtualized I/O server.
  • the new virtualized I/O server may correspond to, for example, a replica of the old virtualized I/O server. In this case, both the new and old virtualized servers may facilitate storage access for the client.
  • the first and second virtualized I/O servers may be, for example, located in a same storage area network (SAN).
  • SAN storage area network
  • each client selects its own I/O connection type.
  • the new virtualized I/O server selects (alone or in combination with, for example, storage manager 128 ) an I/O connection type for each client.
  • the new virtualized I/O server may be, for example, a replica of the old virtualized I/O server that is migrated to a new virtual host.
  • the I/O connection type may be selected based on testing (by a client) each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types.
  • the I/O connection type may be selected (by a client), for example, based on which of the multiple available I/O connection types has a highest performance level and allows a connection between the client and the new virtualized I/O server to be maintained.
  • client 110 communicates the selected I/O connection type to the new virtualized I/O server.
  • client 110 accesses storage using the new virtualized I/O server and optionally the old virtualized I/O server.
  • control transfers to block 310 where process 300 terminates and control returns to a calling routine.
  • process 400 migrates active I/O connection types of clients in a client cluster according to various aspects of the present disclosure. It should be appreciated that portions of process 400 may execute, at any given point in time, on clients 110 or 130 , hosts 134 or 136 , optional storage manager 128 , or a combination thereof. Process 400 may be periodically initiated at block 402 by, for example, storage manager 128 , which may be implemented as a stand-alone application on a dedicated server or on hosts 134 and 136 .
  • Process 400 then proceeds to decision block 404 , where client 110 (alone or in conjunction with, for example, storage manager 128 ) determines whether at least partial storage access control requires migration (e.g., due to an addition of an upgraded host (which may be included in a SAN or NAS) or maintenance on a current host and/or a disk array associated therewith).
  • migration e.g., due to an addition of an upgraded host (which may be included in a SAN or NAS) or maintenance on a current host and/or a disk array associated therewith).
  • control transfers to block 410 , where process 400 terminates and control returns to a calling routine.
  • each of the clients, in a client cluster selects in conjunction with other clients in the client cluster and/or in conjunction with a new (second) virtualized I/O server) an I/O connection type.
  • the clients in the client cluster may need to select a less efficient I/O connection type than is available in order to ensure that the new virtualized I/O server can meet I/O demands of all of the clients in the client cluster.
  • the new virtualized I/O server may correspond to, for example, a replica of the old virtualized I/O server. In this case, both the new and old virtualized servers may control storage access for the clients.
  • the first and second virtualized I/O servers may be, for example, located in a same storage area network (SAN).
  • the new virtualized I/O server may, for example, correspond to a replica of the old virtualized I/O server migrated to a new virtual host.
  • an I/O link may be limited such that only two I/O connection types can be IB, two I/O connections types are required to be SCSI, and two connection types are required to be iSCSI.
  • the I/O connection type may be selected based on testing (by a client) each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types. In this case, the I/O connection type may be selected (by a client in conjunction with other clients), for example, based on which of the multiple available I/O connection types provides a highest performance level and allows a connection between the client and the new virtualized I/O server to be maintained.
  • the virtualized I/O server controls storage access for each client using selected I/O connection types.
  • storage access control for at least some clients in a client cluster may remain under control of the first virtualized I/O server to substantially balance a load for the client cluster between the first and second virtualized I/O servers.
  • control transfers to block 410 where process 400 terminates and control returns to a calling routine.
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.

Abstract

A technique for modifying an input/output (I/O) connection type between a client and a server includes determining, by the client, when at least partial storage access control for the client is to be migrated from a first virtualized I/O server to a second virtualized I/O server. An I/O connection type, from among multiple available I/O connection types available between the client and the second virtualized I/O server, is then selected by the client when the at least partial storage access control for the client is to be migrated. The selected I/O connection type is then communicated, by the client, to the second virtualized I/O server, which is configured to facilitate storage access for the client using the selected I/O connection type.

Description

    BACKGROUND
  • This disclosure relates generally to migrating active input/output (I/O) connections and, more particularly to techniques for migrating active I/O connections with migrating virtualized I/O servers and/or I/O clients.
  • InfiniBand (IB) is a switched fabric communications link that has been employed in high performance computing (HPC) applications. Similar to Fibre channel (FC), peripheral component interconnect (PCI) express, and serial advanced technology attachment (ATA), as well as other modern interconnects, IB offers point-to-point bidirectional serial links for connecting processors with high-speed peripherals, e.g., disks. Today, FC is widely used as the standard connection type for storage area networks (SANs). FC protocol (FCP) is a transport protocol (similar to transmission control protocol (TCP) used in Internet protocol (IP) networks) that predominantly transports small computer system interface (SCSI) commands over FC networks. FC supports a number of upper layer protocols, including SCSI, asynchronous transfer mode (ATM), and IP. The SCSI standards define commands, protocols, and electrical and optical interfaces. SCSI is most commonly used for hard disks and tape drives, but can connect a wide range of other devices, e.g., scanners and compact disc (CD) drives.
  • Unlike FC, which has traditionally required special-purpose cabling, Internet SCSI (iSCSI) can be run over relatively long distances using many existing network infrastructures. iSCSI is an IP-based storage networking standard for linking data storage facilities. iSCSI can be used to transmit data over local area networks (LANs), wide area networks (WANs), or the Internet and facilitates location-independent data storage and retrieval. iSCSI is a popular SAN protocol that allows organizations to consolidate storage into data center storage arrays, while providing hosts (such as database and web servers) with the illusion of locally-attached disks. A SAN architecture is typically used to attach remote computer storage devices (e.g., disk arrays, tape libraries, and optical jukeboxes) to servers in such a way that the devices appear locally attached, with respect to an operating system (OS). A SAN only provides block-level operations, as contrasted with file abstraction. However, file systems may be built on top of SANs to provide file abstraction. In contrast to SANs, network attached storage (NAS) uses file-based protocols, where it is clear that the storage is remote and computers request a portion of an abstract file rather than a disk block. NAS provides both storage and a file system, in contrast to SANs, which typically provide only block-based storage and leave file system concerns to the client-side. Protocols employed in SANs include SCSI, FC, iSCSI, advanced technology attachment (ATA), and ATA over Ethernet (AoE), among other protocols. Despite their differences, SANs and NAS are not mutually exclusive, and may be combined, offering both file-level protocols and block-level protocols from the same system.
  • I/O virtualization is a methodology that is typically implemented to simplify management, lower costs, and improve performance of servers. I/O virtualization environments are created by abstracting upper layer protocols from physical connections. I/O virtualization enables one physical adapter card to appear as multiple virtual network interface cards (vNICs) and virtual host bus adapters (vHBAs). vNICs and vHBAs function as conventional NICs and HBAs and are designed to be compatible with existing OSs, hypervisors, and applications. To networking resources (e.g., LANs and SANs), the vNICs and vHBAs appear as normal cards and adapters. Virtual I/O provides a shared transport for all network and storage connections. Virtual I/O addresses performance bottlenecks by consolidating I/O to a single connection, whose bandwidth ideally exceeds the I/O capacity of an associated server (to ensure that an associated I/O link is not a bottleneck). Bandwidth is then dynamically allocated in real-time across multiple virtual connections to both storage and network resources. In I/O intensive applications, I/O virtualization can help increase both virtual machine (VM) performance and the number of VMs that may execute on a given server.
  • Storage virtualization refers to the process of abstracting logical storage from physical storage. In general, physical storage resources are aggregated into storage pools, from which logical storage is created. Storage virtualization presents a logical space for data storage and transparently handles the process of mapping the logical space to the actual physical location. Storage virtualization is typically implemented in modern disk arrays using vendor proprietary solutions. However, the goal of storage virtualization is usually to virtualize multiple disk arrays from different vendors, scattered over a network, into a single monolithic storage device, which can be managed uniformly.
  • File systems, which are the most common means of accessing disk storage, are abstractions of a physical storage object (e.g., rotating magnetic media, solid state electronic devices, directory structured magnetic tapes) into a more human friendly format, where data can be organized into files, folders and other similar objects. Many modern OSs, including those derived from the UNIX or Windows OSs, abstract the final connection between the file system into, for example, a device switch, a device array, or a device control block. Each physical device, as well as each type of physical device (which may also include network connected storage, virtualized disk storage, etc.) has a different handle, identifier, or other type of object. These data structures or objects include the information needed by the OS or disk I/O subsystem to convert requests for disk I/O into actual data transfer (i.e., either reading or writing).
  • SUMMARY
  • According to one aspect of the present disclosure, a technique for modifying an input/output (I/O) connection type between a client and a server includes determining, by the client, when at least partial storage access control for the client is to be migrated from a first virtualized I/O server to a second virtualized I/O server. An I/O connection type, from among multiple available I/O connection types available between the client and the second virtualized I/O server, is then selected by the client, when the at least partial storage access control for the client is to be migrated. The selected I/O connection type is then communicated, by the client, to the second virtualized I/O server, which is configured to facilitate storage access for the client using the selected I/O connection type.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention is illustrated by way of example and is not intended to be limited by the accompanying figures, in which like references indicate similar elements. Elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale.
  • FIG. 1 is a block diagram of an example data processing environment that may be configured to migrate active input/output (I/O) connections according to various aspects of the present disclosure.
  • FIG. 2 is a block diagram of another example data processing environment that may be configured to migrate active I/O connections according to various embodiments of the present disclosure.
  • FIG. 3 is a flowchart of an example process for migrating an active I/O connection for a client according to various aspects of the present disclosure.
  • FIG. 4 is a flowchart of an example process for migrating active I/O connections for clients in a client cluster according to various aspects of the present disclosure.
  • DETAILED DESCRIPTION
  • As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. As may be used herein, the term “coupled” includes both a direct electrical connection between blocks or components and an indirect electrical connection between blocks or components achieved using one or more intervening blocks or components.
  • When virtualized input/output (I/O) servers, logical partitions, network attached storage (NAS), and/or storage area network (SAN) devices are migrated opportunities for improving I/O connections may exist when, for example, a new physical location of an I/O client is closer to a new physical location of a migrated virtualized I/O server. According to various embodiments of the present disclosure, each time an I/O client (e.g., a disk I/O client) or a virtualized I/O server (e.g., a virtualized disk I/O server) is migrated, related endpoints (i.e., a server in the case of a client migrating or a client in the case of a server migrating) may determine the network topology of each connection, as well as available I/O connection types.
  • For example, a client may query an I/O server as to which I/O connection types are possible given an Internet protocol (IP) address. Based on the connection types available, client and server data structures may be modified to, for example, select the most efficient (e.g., in terms of data transfer speed) I/O connection type available. In one embodiment, disk I/O requests are suspended pending the I/O connection type modification (which may be an upgrade or a downgrade in the I/O connection type), with the I/O connection type modification being considered both prior to migration and once migration is complete. For some I/O connection changes (e.g., local to network), it may be desirable to complete the I/O connection type change prior to migration. For other I/O connection type changes (e.g., network to local), it may be desirable to complete the I/O connection type change following migration. In any case, following disk I/O connection migration, disk I/O may resume on a connection-by-connection basis.
  • According to another aspect of the present disclosure, a new I/O connection may be established between a new virtualized I/O server and an I/O client, while an old I/O connection between the I/O client and an old virtualized I/O server is maintained. In this case, following migration initiation, new I/Os are sent to the new virtualized I/O server, while the old virtualized I/O server completes service of the old I/Os. When the old virtualized I/O server finishes servicing the old I/Os, the old I/O connection is closed. In this manner, I/O access downtime is reduced (and possibly eliminated) and rebooting of the I/O client is avoided.
  • According to various aspects of the present disclosure, an I/O connection may be upgraded or downgraded when an endpoint (server or client) is moved from its origin. In a typical situation, migration of an endpoint, (either a client or a server) results in either a network connection (e.g., an IP address of an iSCSI device changes) or a physical I/O connection (e.g., an available adapter changes from: FC to SCSI; FC to iSCSI; FC to IB; IB to FC; IB to SCSI; IB to iSCSI; iSCSI to SCSI) being changed. For example, a network connection might be changed by migrating a virtual I/O server to a different virtual host (which is not local to a previous connection, or which becomes local to the previous connection). Each type of change may result in a physical type I/O connection being changed (e.g., I/O connections where physical blocks are specified, rather than logical connections where the physical I/O structure is concealed, such as with network file system (NFS), common Internet file system (CIFS), and similar mechanisms). For example, an I/O connection type may change from iSCSI to SCSI or from one network file system type (e.g., network attached storage (NAS), SAN, or other block-structured I/O mechanism) to another, as a result of an I/O server being migrated.
  • A number of patents disclose migrating files or storage. For example, U.S. Pat. No. 7,219,096 (hereinafter “the '096 patent”) discloses a method for migrating files on a busy file system, with the migrated files being recreated on a different file system. However, the '096 patent does not address the case where there is no desire to break a client/server connection, but rather to change characteristics of the connection to one which is more efficient (if possible), or less efficient (if required to maintain the client/server connection). As another example, U.S. Pat. No. 6,976,103 (hereinafter “the '103 patent”) discloses a method for migrating storage. However, the '103 patent does not address modifying a connection type when a virtualized I/O server (or network disk I/O service) is migrated.
  • As yet another example, U.S. Pat. No. 6,931,441 (hereinafter “the '441 patent”) discloses a method for migrating an actual network connection by means of an encapsulating mechanism. In the '441 patent, network communication is made between capsules, with the capsules handling changes in network addressing. However, the '441 patent does not address migrating the physical connection between a disk I/O client and a virtualized disk I/O server. While migration as disclosed in the '441 patent may include changing a network address, the network address change is handled by changing the underlying abstracted block device interface, or other such mechanism. As an example, if a client moved from IP address 192.168.1.12 to IP address 192.168.0.7 and the client is now located within a virtual host that also contains an associated disk I/O server (e.g., at IP address 192.168.0.2), following the disclosure of the '441 patent, packet routing would be changed without upgrading the I/O connection. While some performance benefit may result from changing the network transport (e.g., a virtual network where traffic on the 192.168.0/24 subnet was not transmitted over physical connections, as with products such as VMware and others), greater benefits are usually provided by removing the software layer which abstracts the physical connection into a network protocol (according to one or more embodiments of the present disclosure).
  • With reference to FIG. 1, an example data processing environment 100 is illustrated that includes a client 110 and a client 130 that are configured to select (or at least participate in the selection of) an I/O connection type when an associated virtualized I/O server is migrated (e.g., created in a different virtual host) and/or when clients 110 and 130 are re-located (when, for example, clients 110 and 130 are mobile devices). Clients 110 and 130 may take various forms, such as workstations, laptop computer systems, notebook computer systems, smart phones, web-enabled portable devices, or desktop computer systems. For example, client 110 may correspond to a desktop computer system of a computer system user and client 130 may correspond to a web-enabled device of another computer system user.
  • Client 110 includes a processor 102 (which may include one or more processor cores for executing program code) coupled to a data storage subsystem 104, a display 106, one or more input devices 108, and an input/output adapter (IOA) 109. IOA 109 may be virtualized as: a Fibre Channel (FC) adapter, a small computer system interface (SCSI) adapter, an Internet SCSI (iSCSI) adapter, or an InfiniBand (IB) adapter, among other adapters. Data storage subsystem 104 may include, for example, an application appropriate amount of volatile memory (e.g., dynamic random access memory (DRAM)), non-volatile memory (e.g., read-only memory (ROM) or static RAM), and/or non-volatile mass storage device, such as a magnetic or optical disk drive. Data storage subsystem 104 includes an operating system (OS) 114 for client 110, as well as application programs, such as a browser 112 (which may optionally include customized plug-ins to support various client applications), application 118 (which may include an email application, a word processing application, etc.), and an I/O connection type application 120. I/O connection type application 120, when executed by clients 110 and 130, is configured to interact with a virtualized I/O server (e.g., servers 124 or 126) to determine what I/O connection types are available when clients 110 and 130 are migrated and/or an associated virtualized I/O server (e.g., servers 126 or 124) is migrated.
  • Display 106 may be, for example, a cathode ray tube (CRT) or a liquid crystal display (LCD). Input device(s) 108 of client 110 may include, for example, a mouse, a keyboard, haptic devices, and/or a touch screen. IOA 109 supports communication of client 110 with one or more wired and/or wireless networks utilizing one or more communication protocols, such as 802.x, HTTP, simple mail transfer protocol (SMTP), etc. IOA 109 also facilitates communication between clients 110 and 130 and may physically correspond to, for example, an FC adapter, an IB adapter, an SCSI adapter, or an iSCSI adapter, among other adapters.
  • Clients 110 and 130 are coupled via one or more wired or wireless networks, such as Internet/intranet 122, to optional storage manager 128 (which may be a SAN manager) and hosts 134 and/or 136 and respective virtualized I/ O servers 124 and 126. Virtualized I/ O servers 124 and 126 provide respective access to data stored on disk arrays 115 and 125. Hosts 124 and 126 may be physically located in, for example, a same room of a building, a different room of a building, a different building in a limited geographical area, or a different geographical area (e.g., different cites, counties, states, or countries). As one example, virtualized I/ O servers 124 and 126 may each represent six virtualized I/O servers.
  • It should be appreciated that hosts 134 and 136 may each implement more or less than six virtualized I/O servers and a virtual machine manager (VMM) or hypervisor for controlling operation of the virtualized I/O servers (which may each utilize the same or a different OS), as well as an I/O connection type application that, when executed by hosts 134 and 136, facilitates communication with client 110 and 130 to determine what I/O connection types are available following migration of clients 110 and 130 and/or an associated virtualized I/O server (e.g., servers 126 or 124). While only two clients and two hosts (i.e., physical platforms) are shown associated with data processing environment 100, it should be appreciated that more or less than two clients and more or less than two hosts may be implemented in a data processing environment configured according to the present disclosure. Host 134 and array 115 may be included within a NAS or SAN (NAS/SAN) 131. Similarly, host 136 and array 125 may be included within a different NAS or SAN (NAS/SAN) 133. Arrays 115 and 125 may each correspond to, for example, redundant array of inexpensive disks (RAID) arrays.
  • With reference to FIG. 2, an example data processing environment 200 is illustrated that includes a client cluster 210 that includes multiple clients 110 and a client cluster 230 that includes multiple clients 130. Clients 110 and 130 are configured to select an I/O connection type when an associated virtualized I/O server is migrated (e.g., to a different virtual host) and/or when clients 110 and 130 are re-located (when, for example, clients 110 and 130 are mobile devices). Clients 110 and 130 are coupled via one or more wired or wireless networks, such as Internet/intranet 122, to optional storage manager 128 (which may be, for example, a SAN manager) and hosts 134 and/or 136 and respective virtualized I/ O servers 124 and 126. As with the data processing environment 100 of FIG. 1, virtualized I/ O servers 124 and 126 provide access to data stored on respective disk arrays 115 and 125. For example, virtualized I/ O servers 124 and 126 may each represent one or more virtualized I/O servers.
  • It should be appreciated that hosts 134 and 136 may each implement more than one virtualized I/O server and a virtual machine manager (VMM) or hypervisor for managing operation of implemented virtualized I/O servers. While only two client clusters are shown associated with data processing environment 200, it should be appreciated that more or less than two client clusters may be associated with a data processing environment configured according to the present disclosure. In any case, when a virtualized I/O server that is servicing a client cluster is migrated, clients within a client cluster may coordinate between themselves and/or between the clients and the migrated (new) virtualized I/O server to determine what I/O connection type each of the clients may implement. For example, clients in a client cluster may implement different I/O connection types to balance a load on an associated virtualized I/O server.
  • With reference to FIG. 3, an example process 300 is illustrated that migrates active I/O connection types of a client according to various aspects of the present disclosure. It should be appreciated that portions of process 300 may execute, at any given point in time, on clients 110 or 130, hosts 134 or 136, optional storage manager 128, or a combination thereof. Process 300 may be periodically initiated at block 302 by, for example, storage manager 128, which may be implemented as a stand-alone application on a dedicated server or on hosts 134 and 136. In block 303, client 110 accesses storage via an old (first) virtualized I/O server. Process 300 then proceeds to decision block 304, where client 110 (alone or in conjunction with, for example, storage manager 128) determines whether at least partial storage access control requires migration (e.g., due to an addition of an upgraded host (which may be included in a SAN or NAS) or maintenance on a current host and/or a disk array associated therewith). When migration is not indicated in block 304, control transfers to block 310, where process 300 terminates and control returns to a calling routine.
  • When migration is indicated in block 304, control transfers to block 306, where client 110 (alone or in conjunction with, for example, a new (second) virtualized I/O server) selects an I/O connection type for the connection between client 110 and the new virtualized I/O server. The selected I/O connection type may be the same or different than the I/O connection type employed between the client and the old virtualized I/O server. The new virtualized I/O server may correspond to, for example, a replica of the old virtualized I/O server. In this case, both the new and old virtualized servers may facilitate storage access for the client. The first and second virtualized I/O servers may be, for example, located in a same storage area network (SAN). In one or more embodiments, each client selects its own I/O connection type. In another embodiment, the new virtualized I/O server selects (alone or in combination with, for example, storage manager 128) an I/O connection type for each client. The new virtualized I/O server may be, for example, a replica of the old virtualized I/O server that is migrated to a new virtual host.
  • As one example, the I/O connection type may be selected based on testing (by a client) each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types. In this case, the I/O connection type may be selected (by a client), for example, based on which of the multiple available I/O connection types has a highest performance level and allows a connection between the client and the new virtualized I/O server to be maintained. Next, in block 308, client 110 communicates the selected I/O connection type to the new virtualized I/O server. Then, in block 309, client 110 accesses storage using the new virtualized I/O server and optionally the old virtualized I/O server. Following block 309, control transfers to block 310, where process 300 terminates and control returns to a calling routine.
  • With reference to FIG. 4, an example process 400 is illustrated that migrates active I/O connection types of clients in a client cluster according to various aspects of the present disclosure. It should be appreciated that portions of process 400 may execute, at any given point in time, on clients 110 or 130, hosts 134 or 136, optional storage manager 128, or a combination thereof. Process 400 may be periodically initiated at block 402 by, for example, storage manager 128, which may be implemented as a stand-alone application on a dedicated server or on hosts 134 and 136. Process 400 then proceeds to decision block 404, where client 110 (alone or in conjunction with, for example, storage manager 128) determines whether at least partial storage access control requires migration (e.g., due to an addition of an upgraded host (which may be included in a SAN or NAS) or maintenance on a current host and/or a disk array associated therewith). When migration is not indicated in block 404, control transfers to block 410, where process 400 terminates and control returns to a calling routine.
  • When migration is indicated in block 404, control transfers to block 406, where each of the clients, in a client cluster, selects in conjunction with other clients in the client cluster and/or in conjunction with a new (second) virtualized I/O server) an I/O connection type. In this case, at least some of the clients in the client cluster may need to select a less efficient I/O connection type than is available in order to ensure that the new virtualized I/O server can meet I/O demands of all of the clients in the client cluster. The new virtualized I/O server may correspond to, for example, a replica of the old virtualized I/O server. In this case, both the new and old virtualized servers may control storage access for the clients. The first and second virtualized I/O servers may be, for example, located in a same storage area network (SAN). The new virtualized I/O server may, for example, correspond to a replica of the old virtualized I/O server migrated to a new virtual host.
  • As one example, when a client cluster includes six clients, an I/O link may be limited such that only two I/O connection types can be IB, two I/O connections types are required to be SCSI, and two connection types are required to be iSCSI. As another example, the I/O connection type may be selected based on testing (by a client) each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types. In this case, the I/O connection type may be selected (by a client in conjunction with other clients), for example, based on which of the multiple available I/O connection types provides a highest performance level and allows a connection between the client and the new virtualized I/O server to be maintained. Next, in block 408, the virtualized I/O server controls storage access for each client using selected I/O connection types. As mentioned above, storage access control for at least some clients in a client cluster may remain under control of the first virtualized I/O server to substantially balance a load for the client cluster between the first and second virtualized I/O servers. Following block 408, control transfers to block 410, where process 400 terminates and control returns to a calling routine.
  • Accordingly, techniques have been disclosed herein that select an I/O connection type to maximize data transfer efficiency (between a disk storage client and a disk storage I/O server) when either the client or the I/O server are migrated.
  • The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
  • The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
  • The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below, if any, are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
  • Having thus described the invention of the present application in detail and by reference to preferred embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of the invention defined in the appended claims.

Claims (20)

What is claimed is:
1. A method for modifying an input/output (I/O) connection type between a client and a server, comprising:
determining, by a client, when at least partial storage access control for the client is to be migrated from a first virtualized I/O server to a second virtualized I/O server;
selecting, by the client, an I/O connection type from among multiple available I/O connection types available between the client and the second virtualized I/O server when the at least partial storage access control for the client is to be migrated; and
communicating, by the client, the selected I/O connection type to the second virtualized I/O server, wherein the second virtualized I/O server is configured to facilitate storage access for the client using the selected I/O connection type.
2. The method of claim 1, wherein the second virtualized I/O server is a replica of the first virtualized I/O server and the method further comprises:
accessing, by the client, storage via both the first and second virtualized I/O servers.
3. The method of claim 1, wherein the second virtualized I/O server resides on a different virtual host than the first virtualized I/O server and the second virtualized I/O server is further configured to facilitate storage access for the client during and following migration of the at least partial storage access control from the first virtualized I/O server to the second virtualized I/O server.
4. The method of claim 1, wherein the client includes a cluster of clients and each of the clients in the cluster is configured to select respective I/O connection types in conjunction with the other clients to balance a load on the second virtualized I/O server.
5. The method of claim 1, wherein the I/O connection type is a Fibre Channel connection, a small computer system interface (SCSI) connection, an Internet SCSI connection, or an InfiniBand connection.
6. The method of claim 1, wherein the first and second virtualized I/O servers are located in a same storage area network and the second virtualized I/O server is further configured to facilitate storage access for the client following migration of the at least partial storage access control from the first virtualized I/O server to the second virtualized I/O server.
7. The method of claim 1, wherein the selecting, by the client, an I/O connection type from among multiple available I/O connection types available between the client and the second virtualized I/O server further comprises:
testing, by the client, each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types; and
selecting, by the client, the I/O connection type based on which of the multiple available I/O connection types yields a highest performance level, as indicated by the testing, and allows a connection between the client and the second virtualized I/O server to be maintained.
8. The method of claim 1, wherein the client is included in a client cluster that includes at least one other client, and wherein storage access control for the at least one other client remains under control of the first virtualized I/O server to substantially balance a load for the client cluster between the first and second virtualized I/O servers, where I/O connection types for clients in the client cluster may be the same or different.
9. The method of claim 1, wherein the client queries the second virtualized I/O server to determine the multiple available I/O connection types.
10. The method of claim 1, wherein the second virtualized I/O server is a replica of the first virtualized I/O server that is migrated to a new virtual host.
11. A computer-readable storage medium including computer-readable code for modifying an input/output (I/O) connection type between a client and a server, the computer-readable code comprising:
code for determining, by a client, when at least partial storage access control for the client is to be migrated from a first virtualized I/O server to a second virtualized I/O server;
code for selecting, by the client, an I/O connection type from among multiple available I/O connection types available between the client and the second virtualized I/O server when the at least partial storage access control for the client is to be migrated; and
code for communicating, by the client, the selected I/O connection type to the second virtualized I/O server, wherein the second virtualized I/O server is configured to facilitate storage access for the client using the selected I/O connection type.
12. The computer-readable storage medium of claim 11, wherein the second virtualized I/O server is a replica of the first virtualized I/O server and the code further comprises:
code for accessing, by the client, storage via both the first and second virtualized I/O servers.
13. The computer-readable storage medium of claim 11, wherein the second virtualized I/O server resides on a different virtual host than the first virtualized I/O server and the second virtualized I/O server is further configured to facilitate storage access for the client during and following migration of the at least partial storage access control from the first virtualized I/O server to the second virtualized I/O server.
14. The computer-readable storage medium of claim 11, wherein the client includes a cluster of clients and each of the clients in the cluster is configured to select respective I/O connection types in conjunction with the other clients to balance a load on the second virtualized I/O server, and wherein the I/O connection types include a Fibre Channel connection, a small computer system interface (SCSI) connection, an Internet SCSI connection, and an InfiniBand connection.
15. The computer-readable storage medium of claim 11, wherein the first and second virtualized I/O servers are located in a same storage area network and the second virtualized I/O server is further configured to facilitate storage access for the client following migration of the at least partial storage access control from the first virtualized I/O server to the second virtualized I/O server.
16. The computer-readable storage medium of claim 11, wherein the code for selecting an I/O connection type, from among multiple available I/O connection types, available between the client and the second virtualized I/O server further comprises:
code for testing, by the client, each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types; and
code for selecting, by the client, the I/O connection type based on which of the multiple available I/O connection types yields a highest performance level, as indicated by the testing, and allows a connection between the client and the second virtualized I/O server to be maintained.
17. The computer-readable storage medium of claim 11, wherein the client is included in a client cluster that includes at least one other client and storage access control for the at least one other client remains under control of the first virtualized I/O server to substantially balance a load for the client cluster between the first and second virtualized I/O servers, and wherein the client queries the second virtualized I/O server to determine the multiple available I/O connection types, where I/O connection types for clients in the client cluster may be the same or different.
18. A data processing system, comprising:
a first virtualized server configured to facilitate storage access control;
a client configured to select an I/O connection type, from among multiple available I/O connection types, when at least partial storage access control for the client is migrated from the first virtualized server; and
a second virtualized I/O server configured to transmit the multiple available I/O connection types to the client in response to a query from the client, wherein the second virtualized I/O server is further configured to facilitate storage access control for the client during and following migration of the at least partial storage access control from the first virtualized I/O server to the second virtualized I/O server.
19. The data processing system of claim 18, wherein the client is further configured to test each of the multiple available I/O connection types to determine a performance level of each of the multiple available I/O connection types and select the I/O connection type based on which of the multiple available I/O connection types yields a highest performance level, as indicated by the testing, and allows a connection between the client and the second virtualized I/O server to be maintained.
20. The data processing system of claim 18, wherein the client is included in a client cluster that includes at least one other client and storage access control for the at least one other client remains under control of the first virtualized I/O server to substantially balance a load for the client cluster between the first and second virtualized I/O servers, and wherein the client queries the second virtualized I/O server to determine the multiple available I/O connection types, where I/O connection types for clients in the client cluster may be the same or different.
US12/872,077 2010-08-31 2010-08-31 Techniques for Migrating Active I/O Connections with Migrating Servers and Clients Abandoned US20120054264A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US12/872,077 US20120054264A1 (en) 2010-08-31 2010-08-31 Techniques for Migrating Active I/O Connections with Migrating Servers and Clients
US13/450,633 US8898224B2 (en) 2010-08-31 2012-04-19 Migrating active I/O connections with migrating servers and clients
US14/510,983 US9645764B2 (en) 2010-08-31 2014-10-09 Techniques for migrating active I/O connections with migrating servers and clients

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/872,077 US20120054264A1 (en) 2010-08-31 2010-08-31 Techniques for Migrating Active I/O Connections with Migrating Servers and Clients

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/450,633 Continuation US8898224B2 (en) 2010-08-31 2012-04-19 Migrating active I/O connections with migrating servers and clients

Publications (1)

Publication Number Publication Date
US20120054264A1 true US20120054264A1 (en) 2012-03-01

Family

ID=45698561

Family Applications (3)

Application Number Title Priority Date Filing Date
US12/872,077 Abandoned US20120054264A1 (en) 2010-08-31 2010-08-31 Techniques for Migrating Active I/O Connections with Migrating Servers and Clients
US13/450,633 Expired - Fee Related US8898224B2 (en) 2010-08-31 2012-04-19 Migrating active I/O connections with migrating servers and clients
US14/510,983 Active 2031-03-09 US9645764B2 (en) 2010-08-31 2014-10-09 Techniques for migrating active I/O connections with migrating servers and clients

Family Applications After (2)

Application Number Title Priority Date Filing Date
US13/450,633 Expired - Fee Related US8898224B2 (en) 2010-08-31 2012-04-19 Migrating active I/O connections with migrating servers and clients
US14/510,983 Active 2031-03-09 US9645764B2 (en) 2010-08-31 2014-10-09 Techniques for migrating active I/O connections with migrating servers and clients

Country Status (1)

Country Link
US (3) US20120054264A1 (en)

Cited By (200)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015179523A1 (en) 2014-05-20 2015-11-26 Fresenius Medical Care Holdings, Inc. Absolute blood volume estimation using hemodilution
US9392058B2 (en) * 2014-12-15 2016-07-12 International Business Machines Corporation Migration of executing applications and associated stored data
US9521200B1 (en) * 2015-05-26 2016-12-13 Pure Storage, Inc. Locally providing cloud storage array services
US9594678B1 (en) 2015-05-27 2017-03-14 Pure Storage, Inc. Preventing duplicate entries of identical data in a storage device
US9594512B1 (en) 2015-06-19 2017-03-14 Pure Storage, Inc. Attributing consumed storage capacity among entities storing data in a storage array
US9716755B2 (en) 2015-05-26 2017-07-25 Pure Storage, Inc. Providing cloud storage array services by a local storage array in a data center
US9740414B2 (en) 2015-10-29 2017-08-22 Pure Storage, Inc. Optimizing copy operations
US9760297B2 (en) 2016-02-12 2017-09-12 Pure Storage, Inc. Managing input/output (‘I/O’) queues in a data storage system
US9760479B2 (en) 2015-12-02 2017-09-12 Pure Storage, Inc. Writing data in a storage system that includes a first type of storage device and a second type of storage device
US9811264B1 (en) 2016-04-28 2017-11-07 Pure Storage, Inc. Deploying client-specific applications in a storage system utilizing redundant system resources
US9817603B1 (en) 2016-05-20 2017-11-14 Pure Storage, Inc. Data migration in a storage array that includes a plurality of storage devices
US9841921B2 (en) 2016-04-27 2017-12-12 Pure Storage, Inc. Migrating data in a storage array that includes a plurality of storage devices
US9851762B1 (en) 2015-08-06 2017-12-26 Pure Storage, Inc. Compliant printed circuit board (‘PCB’) within an enclosure
US9882913B1 (en) 2015-05-29 2018-01-30 Pure Storage, Inc. Delivering authorization and authentication for a user of a storage array from a cloud
US9886314B2 (en) 2016-01-28 2018-02-06 Pure Storage, Inc. Placing workloads in a multi-array system
US9887914B2 (en) 2014-02-04 2018-02-06 Fastly, Inc. Communication path selection for content delivery
US9892071B2 (en) 2015-08-03 2018-02-13 Pure Storage, Inc. Emulating a remote direct memory access (‘RDMA’) link between controllers in a storage array
US9910618B1 (en) 2017-04-10 2018-03-06 Pure Storage, Inc. Migrating applications executing on a storage system
US9928093B2 (en) 2015-02-24 2018-03-27 Red Hat Israel, Ltd. Methods and systems for establishing connections associated with virtual machine migrations
US9959043B2 (en) 2016-03-16 2018-05-01 Pure Storage, Inc. Performing a non-disruptive upgrade of data in a storage system
US9971995B2 (en) 2015-06-18 2018-05-15 International Business Machines Corporation Prioritization of e-mail files for migration
US10007459B2 (en) 2016-10-20 2018-06-26 Pure Storage, Inc. Performance tuning in a storage system that includes one or more storage devices
US10021170B2 (en) 2015-05-29 2018-07-10 Pure Storage, Inc. Managing a storage array using client-side services
US10146585B2 (en) 2016-09-07 2018-12-04 Pure Storage, Inc. Ensuring the fair utilization of system resources using workload based, time-independent scheduling
US10162566B2 (en) 2016-11-22 2018-12-25 Pure Storage, Inc. Accumulating application-level statistics in a storage system
US10162835B2 (en) 2015-12-15 2018-12-25 Pure Storage, Inc. Proactive management of a plurality of storage arrays in a multi-array system
US10198194B2 (en) 2015-08-24 2019-02-05 Pure Storage, Inc. Placing data within a storage device of a flash array
US10198205B1 (en) 2016-12-19 2019-02-05 Pure Storage, Inc. Dynamically adjusting a number of storage devices utilized to simultaneously service write operations
US10235229B1 (en) 2016-09-07 2019-03-19 Pure Storage, Inc. Rehabilitating storage devices in a storage array that includes a plurality of storage devices
US10275176B1 (en) 2017-10-19 2019-04-30 Pure Storage, Inc. Data transformation offloading in an artificial intelligence infrastructure
US10284232B2 (en) 2015-10-28 2019-05-07 Pure Storage, Inc. Dynamic error processing in a storage device
US10296258B1 (en) 2018-03-09 2019-05-21 Pure Storage, Inc. Offloading data storage to a decentralized storage network
US10296236B2 (en) 2015-07-01 2019-05-21 Pure Storage, Inc. Offloading device management responsibilities from a storage device in an array of storage devices
US10303390B1 (en) 2016-05-02 2019-05-28 Pure Storage, Inc. Resolving fingerprint collisions in flash storage system
US10310740B2 (en) 2015-06-23 2019-06-04 Pure Storage, Inc. Aligning memory access operations to a geometry of a storage device
US10318196B1 (en) 2015-06-10 2019-06-11 Pure Storage, Inc. Stateless storage system controller in a direct flash storage system
US10326836B2 (en) 2015-12-08 2019-06-18 Pure Storage, Inc. Partially replicating a snapshot between storage systems
US10331588B2 (en) 2016-09-07 2019-06-25 Pure Storage, Inc. Ensuring the appropriate utilization of system resources using weighted workload based, time-independent scheduling
US10346043B2 (en) 2015-12-28 2019-07-09 Pure Storage, Inc. Adaptive computing for data compression
US10353777B2 (en) 2015-10-30 2019-07-16 Pure Storage, Inc. Ensuring crash-safe forward progress of a system configuration update
US10360214B2 (en) 2017-10-19 2019-07-23 Pure Storage, Inc. Ensuring reproducibility in an artificial intelligence infrastructure
US10365982B1 (en) 2017-03-10 2019-07-30 Pure Storage, Inc. Establishing a synchronous replication relationship between two or more storage systems
US10374868B2 (en) 2015-10-29 2019-08-06 Pure Storage, Inc. Distributed command processing in a flash storage system
US10417092B2 (en) 2017-09-07 2019-09-17 Pure Storage, Inc. Incremental RAID stripe update parity calculation
US10452310B1 (en) 2016-07-13 2019-10-22 Pure Storage, Inc. Validating cabling for storage component admission to a storage array
US10452444B1 (en) 2017-10-19 2019-10-22 Pure Storage, Inc. Storage system with compute resources and shared storage resources
US10454810B1 (en) 2017-03-10 2019-10-22 Pure Storage, Inc. Managing host definitions across a plurality of storage systems
US10459652B2 (en) 2016-07-27 2019-10-29 Pure Storage, Inc. Evacuating blades in a storage array that includes a plurality of blades
US10459664B1 (en) 2017-04-10 2019-10-29 Pure Storage, Inc. Virtualized copy-by-reference
US10467107B1 (en) 2017-11-01 2019-11-05 Pure Storage, Inc. Maintaining metadata resiliency among storage device failures
US10474363B1 (en) 2016-07-29 2019-11-12 Pure Storage, Inc. Space reporting in a storage system
US10484174B1 (en) 2017-11-01 2019-11-19 Pure Storage, Inc. Protecting an encryption key for data stored in a storage system that includes a plurality of storage devices
US10489307B2 (en) 2017-01-05 2019-11-26 Pure Storage, Inc. Periodically re-encrypting user data stored on a storage device
US10503427B2 (en) 2017-03-10 2019-12-10 Pure Storage, Inc. Synchronously replicating datasets and other managed objects to cloud-based storage systems
US10503700B1 (en) 2017-01-19 2019-12-10 Pure Storage, Inc. On-demand content filtering of snapshots within a storage system
US10509581B1 (en) 2017-11-01 2019-12-17 Pure Storage, Inc. Maintaining write consistency in a multi-threaded storage system
US10514978B1 (en) 2015-10-23 2019-12-24 Pure Storage, Inc. Automatic deployment of corrective measures for storage arrays
US10521151B1 (en) 2018-03-05 2019-12-31 Pure Storage, Inc. Determining effective space utilization in a storage system
US10552090B2 (en) 2017-09-07 2020-02-04 Pure Storage, Inc. Solid state drives with multiple types of addressable memory
US10572460B2 (en) 2016-02-11 2020-02-25 Pure Storage, Inc. Compressing data in dependence upon characteristics of a storage system
US10599536B1 (en) 2015-10-23 2020-03-24 Pure Storage, Inc. Preventing storage errors using problem signatures
US10613791B2 (en) 2017-06-12 2020-04-07 Pure Storage, Inc. Portable snapshot replication between storage systems
US10671302B1 (en) 2018-10-26 2020-06-02 Pure Storage, Inc. Applying a rate limit across a plurality of storage systems
US10671494B1 (en) 2017-11-01 2020-06-02 Pure Storage, Inc. Consistent selection of replicated datasets during storage system recovery
US10671439B1 (en) 2016-09-07 2020-06-02 Pure Storage, Inc. Workload planning with quality-of-service (‘QOS’) integration
US10691567B2 (en) 2016-06-03 2020-06-23 Pure Storage, Inc. Dynamically forming a failure domain in a storage system that includes a plurality of blades
US10789020B2 (en) 2017-06-12 2020-09-29 Pure Storage, Inc. Recovering data within a unified storage element
US10795598B1 (en) 2017-12-07 2020-10-06 Pure Storage, Inc. Volume migration for storage systems synchronously replicating a dataset
US10803978B2 (en) * 2011-01-14 2020-10-13 Dispersive Networks, Inc. Selective access to medical symptom tracking data using dispersive storage area network (SAN)
US10817392B1 (en) 2017-11-01 2020-10-27 Pure Storage, Inc. Ensuring resiliency to storage device failures in a storage system that includes a plurality of storage devices
US10834086B1 (en) 2015-05-29 2020-11-10 Pure Storage, Inc. Hybrid cloud-based authentication for flash storage array access
US10838833B1 (en) 2018-03-26 2020-11-17 Pure Storage, Inc. Providing for high availability in a data analytics pipeline without replicas
US10853148B1 (en) 2017-06-12 2020-12-01 Pure Storage, Inc. Migrating workloads between a plurality of execution environments
US10871922B2 (en) 2018-05-22 2020-12-22 Pure Storage, Inc. Integrated storage management between storage systems and container orchestrators
US10884636B1 (en) 2017-06-12 2021-01-05 Pure Storage, Inc. Presenting workload performance in a storage system
US10908966B1 (en) 2016-09-07 2021-02-02 Pure Storage, Inc. Adapting target service times in a storage system
US10917470B1 (en) 2018-11-18 2021-02-09 Pure Storage, Inc. Cloning storage systems in a cloud computing environment
US10917471B1 (en) 2018-03-15 2021-02-09 Pure Storage, Inc. Active membership in a cloud-based storage system
US10924548B1 (en) 2018-03-15 2021-02-16 Pure Storage, Inc. Symmetric storage using a cloud-based storage system
US10929226B1 (en) 2017-11-21 2021-02-23 Pure Storage, Inc. Providing for increased flexibility for large scale parity
US10936238B2 (en) 2017-11-28 2021-03-02 Pure Storage, Inc. Hybrid data tiering
US10942650B1 (en) 2018-03-05 2021-03-09 Pure Storage, Inc. Reporting capacity utilization in a storage system
US10963189B1 (en) 2018-11-18 2021-03-30 Pure Storage, Inc. Coalescing write operations in a cloud-based storage system
US10976962B2 (en) 2018-03-15 2021-04-13 Pure Storage, Inc. Servicing I/O operations in a cloud-based storage system
US10992533B1 (en) 2018-01-30 2021-04-27 Pure Storage, Inc. Policy based path management
US10992598B2 (en) 2018-05-21 2021-04-27 Pure Storage, Inc. Synchronously replicating when a mediation service becomes unavailable
US10990282B1 (en) 2017-11-28 2021-04-27 Pure Storage, Inc. Hybrid data tiering with cloud storage
US11003369B1 (en) 2019-01-14 2021-05-11 Pure Storage, Inc. Performing a tune-up procedure on a storage device during a boot process
US11016824B1 (en) 2017-06-12 2021-05-25 Pure Storage, Inc. Event identification with out-of-order reporting in a cloud-based environment
US11036677B1 (en) 2017-12-14 2021-06-15 Pure Storage, Inc. Replicated data integrity
US11042452B1 (en) 2019-03-20 2021-06-22 Pure Storage, Inc. Storage system data recovery using data recovery as a service
US11048590B1 (en) 2018-03-15 2021-06-29 Pure Storage, Inc. Data consistency during recovery in a cloud-based storage system
US11068162B1 (en) 2019-04-09 2021-07-20 Pure Storage, Inc. Storage management in a cloud data store
US11086553B1 (en) 2019-08-28 2021-08-10 Pure Storage, Inc. Tiering duplicated objects in a cloud-based object store
US11089105B1 (en) 2017-12-14 2021-08-10 Pure Storage, Inc. Synchronously replicating datasets in cloud-based storage systems
US11093139B1 (en) 2019-07-18 2021-08-17 Pure Storage, Inc. Durably storing data within a virtual storage system
US11095706B1 (en) 2018-03-21 2021-08-17 Pure Storage, Inc. Secure cloud-based storage system management
CN113282246A (en) * 2021-06-15 2021-08-20 杭州海康威视数字技术股份有限公司 Data processing method and device
US11102298B1 (en) 2015-05-26 2021-08-24 Pure Storage, Inc. Locally providing cloud storage services for fleet management
US11112990B1 (en) 2016-04-27 2021-09-07 Pure Storage, Inc. Managing storage device evacuation
US11126364B2 (en) 2019-07-18 2021-09-21 Pure Storage, Inc. Virtual storage system architecture
US11146564B1 (en) 2018-07-24 2021-10-12 Pure Storage, Inc. Login authentication in a cloud storage platform
US11150834B1 (en) 2018-03-05 2021-10-19 Pure Storage, Inc. Determining storage consumption in a storage system
US11163624B2 (en) 2017-01-27 2021-11-02 Pure Storage, Inc. Dynamically adjusting an amount of log data generated for a storage system
US11169727B1 (en) 2017-03-10 2021-11-09 Pure Storage, Inc. Synchronous replication between storage systems with virtualized storage
US11171950B1 (en) 2018-03-21 2021-11-09 Pure Storage, Inc. Secure cloud-based storage system management
US11210133B1 (en) 2017-06-12 2021-12-28 Pure Storage, Inc. Workload mobility between disparate execution environments
US11210009B1 (en) 2018-03-15 2021-12-28 Pure Storage, Inc. Staging data in a cloud-based storage system
US11221778B1 (en) 2019-04-02 2022-01-11 Pure Storage, Inc. Preparing data for deduplication
US11231858B2 (en) 2016-05-19 2022-01-25 Pure Storage, Inc. Dynamically configuring a storage system to facilitate independent scaling of resources
US11288138B1 (en) 2018-03-15 2022-03-29 Pure Storage, Inc. Recovery from a system fault in a cloud-based storage system
US11294588B1 (en) 2015-08-24 2022-04-05 Pure Storage, Inc. Placing data within a storage device
US11301152B1 (en) 2020-04-06 2022-04-12 Pure Storage, Inc. Intelligently moving data between storage systems
US11321006B1 (en) 2020-03-25 2022-05-03 Pure Storage, Inc. Data loss prevention during transitions from a replication source
US11327676B1 (en) 2019-07-18 2022-05-10 Pure Storage, Inc. Predictive data streaming in a virtual storage system
US11340837B1 (en) 2018-11-18 2022-05-24 Pure Storage, Inc. Storage system management via a remote console
US11340800B1 (en) 2017-01-19 2022-05-24 Pure Storage, Inc. Content masking in a storage system
US11340939B1 (en) 2017-06-12 2022-05-24 Pure Storage, Inc. Application-aware analytics for storage systems
US11347697B1 (en) 2015-12-15 2022-05-31 Pure Storage, Inc. Proactively optimizing a storage system
US11349917B2 (en) 2020-07-23 2022-05-31 Pure Storage, Inc. Replication handling among distinct networks
US11360689B1 (en) 2019-09-13 2022-06-14 Pure Storage, Inc. Cloning a tracking copy of replica data
US11360844B1 (en) 2015-10-23 2022-06-14 Pure Storage, Inc. Recovery of a container storage provider
US11379132B1 (en) 2016-10-20 2022-07-05 Pure Storage, Inc. Correlating medical sensor data
US11392555B2 (en) 2019-05-15 2022-07-19 Pure Storage, Inc. Cloud-based file services
US11392553B1 (en) 2018-04-24 2022-07-19 Pure Storage, Inc. Remote data management
US11397545B1 (en) 2021-01-20 2022-07-26 Pure Storage, Inc. Emulating persistent reservations in a cloud-based storage system
US11403000B1 (en) 2018-07-20 2022-08-02 Pure Storage, Inc. Resiliency in a cloud-based storage system
US11416298B1 (en) 2018-07-20 2022-08-16 Pure Storage, Inc. Providing application-specific storage by a storage system
US11422731B1 (en) 2017-06-12 2022-08-23 Pure Storage, Inc. Metadata-based replication of a dataset
US11431488B1 (en) 2020-06-08 2022-08-30 Pure Storage, Inc. Protecting local key generation using a remote key management service
US11436344B1 (en) 2018-04-24 2022-09-06 Pure Storage, Inc. Secure encryption in deduplication cluster
US11442652B1 (en) 2020-07-23 2022-09-13 Pure Storage, Inc. Replication handling during storage system transportation
US11442669B1 (en) 2018-03-15 2022-09-13 Pure Storage, Inc. Orchestrating a virtual storage system
US11442825B2 (en) 2017-03-10 2022-09-13 Pure Storage, Inc. Establishing a synchronous replication relationship between two or more storage systems
US11455409B2 (en) 2018-05-21 2022-09-27 Pure Storage, Inc. Storage layer data obfuscation
US11455168B1 (en) 2017-10-19 2022-09-27 Pure Storage, Inc. Batch building for deep learning training workloads
US11461273B1 (en) 2016-12-20 2022-10-04 Pure Storage, Inc. Modifying storage distribution in a storage system that includes one or more storage devices
US11477280B1 (en) 2017-07-26 2022-10-18 Pure Storage, Inc. Integrating cloud storage services
US11481261B1 (en) 2016-09-07 2022-10-25 Pure Storage, Inc. Preventing extended latency in a storage system
US11487715B1 (en) 2019-07-18 2022-11-01 Pure Storage, Inc. Resiliency in a cloud-based storage system
US11494692B1 (en) 2018-03-26 2022-11-08 Pure Storage, Inc. Hyperscale artificial intelligence and machine learning infrastructure
US11494267B2 (en) 2020-04-14 2022-11-08 Pure Storage, Inc. Continuous value data redundancy
US11503031B1 (en) 2015-05-29 2022-11-15 Pure Storage, Inc. Storage array access control from cloud-based user authorization and authentication
US11526405B1 (en) 2018-11-18 2022-12-13 Pure Storage, Inc. Cloud-based disaster recovery
US11526408B2 (en) 2019-07-18 2022-12-13 Pure Storage, Inc. Data recovery in a virtual storage system
US11531487B1 (en) 2019-12-06 2022-12-20 Pure Storage, Inc. Creating a replica of a storage system
US11531577B1 (en) 2016-09-07 2022-12-20 Pure Storage, Inc. Temporarily limiting access to a storage device
US11550514B2 (en) 2019-07-18 2023-01-10 Pure Storage, Inc. Efficient transfers between tiers of a virtual storage system
US11561714B1 (en) 2017-07-05 2023-01-24 Pure Storage, Inc. Storage efficiency driven migration
US11573864B1 (en) 2019-09-16 2023-02-07 Pure Storage, Inc. Automating database management in a storage system
US11588716B2 (en) 2021-05-12 2023-02-21 Pure Storage, Inc. Adaptive storage processing for storage-as-a-service
US11592991B2 (en) 2017-09-07 2023-02-28 Pure Storage, Inc. Converting raid data between persistent storage types
US11609718B1 (en) 2017-06-12 2023-03-21 Pure Storage, Inc. Identifying valid data after a storage system recovery
US11616834B2 (en) 2015-12-08 2023-03-28 Pure Storage, Inc. Efficient replication of a dataset to the cloud
US11620075B2 (en) 2016-11-22 2023-04-04 Pure Storage, Inc. Providing application aware storage
US11625181B1 (en) 2015-08-24 2023-04-11 Pure Storage, Inc. Data tiering using snapshots
US11630585B1 (en) 2016-08-25 2023-04-18 Pure Storage, Inc. Processing evacuation events in a storage array that includes a plurality of storage devices
US11630598B1 (en) 2020-04-06 2023-04-18 Pure Storage, Inc. Scheduling data replication operations
US11632360B1 (en) 2018-07-24 2023-04-18 Pure Storage, Inc. Remote access to a storage device
US11637896B1 (en) 2020-02-25 2023-04-25 Pure Storage, Inc. Migrating applications to a cloud-computing environment
US11650749B1 (en) 2018-12-17 2023-05-16 Pure Storage, Inc. Controlling access to sensitive data in a shared dataset
US11669386B1 (en) 2019-10-08 2023-06-06 Pure Storage, Inc. Managing an application's resource stack
US11675520B2 (en) 2017-03-10 2023-06-13 Pure Storage, Inc. Application replication among storage systems synchronously replicating a dataset
US11675503B1 (en) 2018-05-21 2023-06-13 Pure Storage, Inc. Role-based data access
US11693713B1 (en) 2019-09-04 2023-07-04 Pure Storage, Inc. Self-tuning clusters for resilient microservices
US11706895B2 (en) 2016-07-19 2023-07-18 Pure Storage, Inc. Independent scaling of compute resources and storage resources in a storage system
US11709636B1 (en) 2020-01-13 2023-07-25 Pure Storage, Inc. Non-sequential readahead for deep learning training
US11714723B2 (en) 2021-10-29 2023-08-01 Pure Storage, Inc. Coordinated snapshots for data stored across distinct storage environments
US11720497B1 (en) 2020-01-13 2023-08-08 Pure Storage, Inc. Inferred nonsequential prefetch based on data access patterns
US11733901B1 (en) 2020-01-13 2023-08-22 Pure Storage, Inc. Providing persistent storage to transient cloud computing services
US11762764B1 (en) 2015-12-02 2023-09-19 Pure Storage, Inc. Writing data in a storage system that includes a first type of storage device and a second type of storage device
US11762781B2 (en) 2017-01-09 2023-09-19 Pure Storage, Inc. Providing end-to-end encryption for data stored in a storage system
US11782614B1 (en) 2017-12-21 2023-10-10 Pure Storage, Inc. Encrypting data to optimize data reduction
US11797569B2 (en) 2019-09-13 2023-10-24 Pure Storage, Inc. Configurable data replication
US11803453B1 (en) 2017-03-10 2023-10-31 Pure Storage, Inc. Using host connectivity states to avoid queuing I/O requests
US11809727B1 (en) 2016-04-27 2023-11-07 Pure Storage, Inc. Predicting failures in a storage system that includes a plurality of storage devices
US11816129B2 (en) 2021-06-22 2023-11-14 Pure Storage, Inc. Generating datasets using approximate baselines
US11847071B2 (en) 2021-12-30 2023-12-19 Pure Storage, Inc. Enabling communication between a single-port device and multiple storage system controllers
US11853266B2 (en) 2019-05-15 2023-12-26 Pure Storage, Inc. Providing a file system in a cloud environment
US11853285B1 (en) 2021-01-22 2023-12-26 Pure Storage, Inc. Blockchain logging of volume-level events in a storage system
US11861170B2 (en) 2018-03-05 2024-01-02 Pure Storage, Inc. Sizing resources for a replication target
US11860820B1 (en) 2018-09-11 2024-01-02 Pure Storage, Inc. Processing data through a storage system in a data pipeline
US11861221B1 (en) 2019-07-18 2024-01-02 Pure Storage, Inc. Providing scalable and reliable container-based storage services
US11860780B2 (en) 2022-01-28 2024-01-02 Pure Storage, Inc. Storage cache management
US11861423B1 (en) 2017-10-19 2024-01-02 Pure Storage, Inc. Accelerating artificial intelligence (‘AI’) workflows
US11868629B1 (en) 2017-05-05 2024-01-09 Pure Storage, Inc. Storage system sizing service
US11868622B2 (en) 2020-02-25 2024-01-09 Pure Storage, Inc. Application recovery across storage systems
US11886922B2 (en) 2016-09-07 2024-01-30 Pure Storage, Inc. Scheduling input/output operations for a storage system
US11886295B2 (en) 2022-01-31 2024-01-30 Pure Storage, Inc. Intra-block error correction
US11893263B2 (en) 2021-10-29 2024-02-06 Pure Storage, Inc. Coordinated checkpoints among storage systems implementing checkpoint-based replication
US11914867B2 (en) 2021-10-29 2024-02-27 Pure Storage, Inc. Coordinated snapshots among storage systems implementing a promotion/demotion model
US11921908B2 (en) 2017-08-31 2024-03-05 Pure Storage, Inc. Writing data to compressed and encrypted volumes
US11922052B2 (en) 2021-12-15 2024-03-05 Pure Storage, Inc. Managing links between storage objects
US11921670B1 (en) 2020-04-20 2024-03-05 Pure Storage, Inc. Multivariate data backup retention policies
US11941279B2 (en) 2017-03-10 2024-03-26 Pure Storage, Inc. Data path virtualization
US11954238B1 (en) 2018-07-24 2024-04-09 Pure Storage, Inc. Role-based access control for a storage system
US11954220B2 (en) 2018-05-21 2024-04-09 Pure Storage, Inc. Data protection for container storage
US11960777B2 (en) 2017-06-12 2024-04-16 Pure Storage, Inc. Utilizing multiple redundancy schemes within a unified storage element
US11960348B2 (en) 2016-09-07 2024-04-16 Pure Storage, Inc. Cloud-based monitoring of hardware components in a fleet of storage systems
US11972134B2 (en) 2022-01-12 2024-04-30 Pure Storage, Inc. Resource utilization using normalized input/output (‘I/O’) operations

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9264384B1 (en) 2004-07-22 2016-02-16 Oracle International Corporation Resource virtualization mechanism including virtual host bus adapters
US9813283B2 (en) 2005-08-09 2017-11-07 Oracle International Corporation Efficient data transfer between servers and remote peripherals
US9973446B2 (en) 2009-08-20 2018-05-15 Oracle International Corporation Remote shared server peripherals over an Ethernet network for resource virtualization
US20120054264A1 (en) 2010-08-31 2012-03-01 International Business Machines Corporation Techniques for Migrating Active I/O Connections with Migrating Servers and Clients
US9331963B2 (en) 2010-09-24 2016-05-03 Oracle International Corporation Wireless host I/O using virtualized I/O controllers
US9083550B2 (en) 2012-10-29 2015-07-14 Oracle International Corporation Network virtualization over infiniband
US10747475B2 (en) 2013-08-26 2020-08-18 Vmware, Inc. Virtual disk blueprints for a virtualized storage area network, wherein virtual disk objects are created from local physical storage of host computers that are running multiple virtual machines
US9887924B2 (en) * 2013-08-26 2018-02-06 Vmware, Inc. Distributed policy-based provisioning and enforcement for quality of service
US9811531B2 (en) 2013-08-26 2017-11-07 Vmware, Inc. Scalable distributed storage architecture
US11016820B2 (en) 2013-08-26 2021-05-25 Vmware, Inc. Load balancing of resources
US10666513B2 (en) * 2017-05-03 2020-05-26 International Business Machines Corporation Filesystem share auto-detect
CN110677492B (en) * 2019-10-11 2022-08-02 北京字节跳动网络技术有限公司 Access request processing method and device, electronic equipment and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5950011A (en) * 1996-03-01 1999-09-07 Bull S.A. System using designer editor and knowledge base for configuring preconfigured software in an open system in a distributed environment
US6931441B1 (en) * 2001-06-29 2005-08-16 Cisco Technology, Inc. Method and apparatus for managing a network using link state information
US20050228835A1 (en) * 2004-04-12 2005-10-13 Guillermo Roa System and method for supporting block-based protocols on a virtual storage appliance executing within a physical storage appliance
US6976103B1 (en) * 2000-03-03 2005-12-13 Hitachi, Ltd. Data migration method using storage area network
US7219096B2 (en) * 2002-06-06 2007-05-15 International Business Machines Corporation Method for migrating open files on a busy file system
US20070220121A1 (en) * 2006-03-18 2007-09-20 Ignatia Suwarna Virtual machine migration between servers
US20100125712A1 (en) * 2008-11-17 2010-05-20 Hitachi, Ltd. Data backup system and method for virtual infrastructure

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090216880A1 (en) 2008-02-26 2009-08-27 Viasat, Inc. Methods and Systems for Dynamic Transport Selection Based on Last Mile Network Detection
JP2009237763A (en) * 2008-03-26 2009-10-15 Hitachi Ltd Server system and control method therefor
US20120054264A1 (en) 2010-08-31 2012-03-01 International Business Machines Corporation Techniques for Migrating Active I/O Connections with Migrating Servers and Clients

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5950011A (en) * 1996-03-01 1999-09-07 Bull S.A. System using designer editor and knowledge base for configuring preconfigured software in an open system in a distributed environment
US6976103B1 (en) * 2000-03-03 2005-12-13 Hitachi, Ltd. Data migration method using storage area network
US6931441B1 (en) * 2001-06-29 2005-08-16 Cisco Technology, Inc. Method and apparatus for managing a network using link state information
US7219096B2 (en) * 2002-06-06 2007-05-15 International Business Machines Corporation Method for migrating open files on a busy file system
US20050228835A1 (en) * 2004-04-12 2005-10-13 Guillermo Roa System and method for supporting block-based protocols on a virtual storage appliance executing within a physical storage appliance
US20070220121A1 (en) * 2006-03-18 2007-09-20 Ignatia Suwarna Virtual machine migration between servers
US20100125712A1 (en) * 2008-11-17 2010-05-20 Hitachi, Ltd. Data backup system and method for virtual infrastructure

Cited By (380)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10803978B2 (en) * 2011-01-14 2020-10-13 Dispersive Networks, Inc. Selective access to medical symptom tracking data using dispersive storage area network (SAN)
US10536372B2 (en) 2014-02-04 2020-01-14 Fastly Inc. Communication path selection for content delivery
US9887914B2 (en) 2014-02-04 2018-02-06 Fastly, Inc. Communication path selection for content delivery
CN106456865A (en) * 2014-05-20 2017-02-22 弗雷泽纽斯医疗保健控股有限公司 Absolute blood volume estimation using hemodilution
WO2015179523A1 (en) 2014-05-20 2015-11-26 Fresenius Medical Care Holdings, Inc. Absolute blood volume estimation using hemodilution
US9392058B2 (en) * 2014-12-15 2016-07-12 International Business Machines Corporation Migration of executing applications and associated stored data
US9389789B2 (en) * 2014-12-15 2016-07-12 International Business Machines Corporation Migration of executing applications and associated stored data
US9600195B2 (en) * 2014-12-15 2017-03-21 International Business Machines Corporation Migration of executing applications and associated stored data
US9600196B2 (en) * 2014-12-15 2017-03-21 International Business Machines Corporation Migration of executing applications and associated stored data
US9928093B2 (en) 2015-02-24 2018-03-27 Red Hat Israel, Ltd. Methods and systems for establishing connections associated with virtual machine migrations
US10664301B2 (en) 2015-02-24 2020-05-26 Red Hat Israel, Ltd. Methods and systems for establishing connections associated with virtual machine migrations
US10027757B1 (en) 2015-05-26 2018-07-17 Pure Storage, Inc. Locally providing cloud storage array services
US11102298B1 (en) 2015-05-26 2021-08-24 Pure Storage, Inc. Locally providing cloud storage services for fleet management
US9521200B1 (en) * 2015-05-26 2016-12-13 Pure Storage, Inc. Locally providing cloud storage array services
US11711426B2 (en) 2015-05-26 2023-07-25 Pure Storage, Inc. Providing storage resources from a storage pool
US10652331B1 (en) 2015-05-26 2020-05-12 Pure Storage, Inc. Locally providing highly available cloud-based storage system services
US9716755B2 (en) 2015-05-26 2017-07-25 Pure Storage, Inc. Providing cloud storage array services by a local storage array in a data center
US10761759B1 (en) 2015-05-27 2020-09-01 Pure Storage, Inc. Deduplication of data in a storage device
US11360682B1 (en) 2015-05-27 2022-06-14 Pure Storage, Inc. Identifying duplicative write data in a storage system
US9594678B1 (en) 2015-05-27 2017-03-14 Pure Storage, Inc. Preventing duplicate entries of identical data in a storage device
US11921633B2 (en) 2015-05-27 2024-03-05 Pure Storage, Inc. Deduplicating data based on recently reading the data
US11936719B2 (en) 2015-05-29 2024-03-19 Pure Storage, Inc. Using cloud services to provide secure access to a storage system
US10560517B1 (en) 2015-05-29 2020-02-11 Pure Storage, Inc. Remote management of a storage array
US10834086B1 (en) 2015-05-29 2020-11-10 Pure Storage, Inc. Hybrid cloud-based authentication for flash storage array access
US11503031B1 (en) 2015-05-29 2022-11-15 Pure Storage, Inc. Storage array access control from cloud-based user authorization and authentication
US11201913B1 (en) 2015-05-29 2021-12-14 Pure Storage, Inc. Cloud-based authentication of a storage system user
US11936654B2 (en) 2015-05-29 2024-03-19 Pure Storage, Inc. Cloud-based user authorization control for storage system access
US10021170B2 (en) 2015-05-29 2018-07-10 Pure Storage, Inc. Managing a storage array using client-side services
US9882913B1 (en) 2015-05-29 2018-01-30 Pure Storage, Inc. Delivering authorization and authentication for a user of a storage array from a cloud
US11868625B2 (en) 2015-06-10 2024-01-09 Pure Storage, Inc. Alert tracking in storage
US11137918B1 (en) 2015-06-10 2021-10-05 Pure Storage, Inc. Administration of control information in a storage system
US10318196B1 (en) 2015-06-10 2019-06-11 Pure Storage, Inc. Stateless storage system controller in a direct flash storage system
US10600032B2 (en) 2015-06-18 2020-03-24 International Business Machines Corporation Prioritization of e-mail files for migration
US9971995B2 (en) 2015-06-18 2018-05-15 International Business Machines Corporation Prioritization of e-mail files for migration
US10082971B1 (en) 2015-06-19 2018-09-25 Pure Storage, Inc. Calculating capacity utilization in a storage system
US9594512B1 (en) 2015-06-19 2017-03-14 Pure Storage, Inc. Attributing consumed storage capacity among entities storing data in a storage array
US11586359B1 (en) 2015-06-19 2023-02-21 Pure Storage, Inc. Tracking storage consumption in a storage array
US10866744B1 (en) 2015-06-19 2020-12-15 Pure Storage, Inc. Determining capacity utilization in a deduplicating storage system
US9804779B1 (en) 2015-06-19 2017-10-31 Pure Storage, Inc. Determining storage capacity to be made available upon deletion of a shared data object
US10310753B1 (en) 2015-06-19 2019-06-04 Pure Storage, Inc. Capacity attribution in a storage system
US10310740B2 (en) 2015-06-23 2019-06-04 Pure Storage, Inc. Aligning memory access operations to a geometry of a storage device
US10296236B2 (en) 2015-07-01 2019-05-21 Pure Storage, Inc. Offloading device management responsibilities from a storage device in an array of storage devices
US11385801B1 (en) 2015-07-01 2022-07-12 Pure Storage, Inc. Offloading device management responsibilities of a storage device to a storage controller
US10540307B1 (en) 2015-08-03 2020-01-21 Pure Storage, Inc. Providing an active/active front end by coupled controllers in a storage system
US11681640B2 (en) 2015-08-03 2023-06-20 Pure Storage, Inc. Multi-channel communications between controllers in a storage system
US9892071B2 (en) 2015-08-03 2018-02-13 Pure Storage, Inc. Emulating a remote direct memory access (‘RDMA’) link between controllers in a storage array
US9910800B1 (en) 2015-08-03 2018-03-06 Pure Storage, Inc. Utilizing remote direct memory access (‘RDMA’) for communication between controllers in a storage array
US9851762B1 (en) 2015-08-06 2017-12-26 Pure Storage, Inc. Compliant printed circuit board (‘PCB’) within an enclosure
US11294588B1 (en) 2015-08-24 2022-04-05 Pure Storage, Inc. Placing data within a storage device
US11625181B1 (en) 2015-08-24 2023-04-11 Pure Storage, Inc. Data tiering using snapshots
US11868636B2 (en) 2015-08-24 2024-01-09 Pure Storage, Inc. Prioritizing garbage collection based on the extent to which data is deduplicated
US10198194B2 (en) 2015-08-24 2019-02-05 Pure Storage, Inc. Placing data within a storage device of a flash array
US11360844B1 (en) 2015-10-23 2022-06-14 Pure Storage, Inc. Recovery of a container storage provider
US10514978B1 (en) 2015-10-23 2019-12-24 Pure Storage, Inc. Automatic deployment of corrective measures for storage arrays
US11874733B2 (en) 2015-10-23 2024-01-16 Pure Storage, Inc. Recovering a container storage system
US11934260B2 (en) 2015-10-23 2024-03-19 Pure Storage, Inc. Problem signature-based corrective measure deployment
US11061758B1 (en) 2015-10-23 2021-07-13 Pure Storage, Inc. Proactively providing corrective measures for storage arrays
US10599536B1 (en) 2015-10-23 2020-03-24 Pure Storage, Inc. Preventing storage errors using problem signatures
US11593194B2 (en) 2015-10-23 2023-02-28 Pure Storage, Inc. Cloud-based providing of one or more corrective measures for a storage system
US11784667B2 (en) 2015-10-28 2023-10-10 Pure Storage, Inc. Selecting optimal responses to errors in a storage system
US10432233B1 (en) 2015-10-28 2019-10-01 Pure Storage Inc. Error correction processing in a storage device
US10284232B2 (en) 2015-10-28 2019-05-07 Pure Storage, Inc. Dynamic error processing in a storage device
US10374868B2 (en) 2015-10-29 2019-08-06 Pure Storage, Inc. Distributed command processing in a flash storage system
US11836357B2 (en) 2015-10-29 2023-12-05 Pure Storage, Inc. Memory aligned copy operation execution
US11032123B1 (en) 2015-10-29 2021-06-08 Pure Storage, Inc. Hierarchical storage system management
US10268403B1 (en) 2015-10-29 2019-04-23 Pure Storage, Inc. Combining multiple copy operations into a single copy operation
US10956054B1 (en) 2015-10-29 2021-03-23 Pure Storage, Inc. Efficient performance of copy operations in a storage system
US9740414B2 (en) 2015-10-29 2017-08-22 Pure Storage, Inc. Optimizing copy operations
US11422714B1 (en) 2015-10-29 2022-08-23 Pure Storage, Inc. Efficient copying of data in a storage system
US10929231B1 (en) 2015-10-30 2021-02-23 Pure Storage, Inc. System configuration selection in a storage system
US10353777B2 (en) 2015-10-30 2019-07-16 Pure Storage, Inc. Ensuring crash-safe forward progress of a system configuration update
US9760479B2 (en) 2015-12-02 2017-09-12 Pure Storage, Inc. Writing data in a storage system that includes a first type of storage device and a second type of storage device
US10970202B1 (en) 2015-12-02 2021-04-06 Pure Storage, Inc. Managing input/output (‘I/O’) requests in a storage system that includes multiple types of storage devices
US10255176B1 (en) 2015-12-02 2019-04-09 Pure Storage, Inc. Input/output (‘I/O’) in a storage system that includes multiple types of storage devices
US11762764B1 (en) 2015-12-02 2023-09-19 Pure Storage, Inc. Writing data in a storage system that includes a first type of storage device and a second type of storage device
US10326836B2 (en) 2015-12-08 2019-06-18 Pure Storage, Inc. Partially replicating a snapshot between storage systems
US11616834B2 (en) 2015-12-08 2023-03-28 Pure Storage, Inc. Efficient replication of a dataset to the cloud
US10986179B1 (en) 2015-12-08 2021-04-20 Pure Storage, Inc. Cloud-based snapshot replication
US11030160B1 (en) 2015-12-15 2021-06-08 Pure Storage, Inc. Projecting the effects of implementing various actions on a storage system
US11347697B1 (en) 2015-12-15 2022-05-31 Pure Storage, Inc. Proactively optimizing a storage system
US11836118B2 (en) 2015-12-15 2023-12-05 Pure Storage, Inc. Performance metric-based improvement of one or more conditions of a storage array
US10162835B2 (en) 2015-12-15 2018-12-25 Pure Storage, Inc. Proactive management of a plurality of storage arrays in a multi-array system
US11281375B1 (en) 2015-12-28 2022-03-22 Pure Storage, Inc. Optimizing for data reduction in a storage system
US10346043B2 (en) 2015-12-28 2019-07-09 Pure Storage, Inc. Adaptive computing for data compression
US10929185B1 (en) 2016-01-28 2021-02-23 Pure Storage, Inc. Predictive workload placement
US9886314B2 (en) 2016-01-28 2018-02-06 Pure Storage, Inc. Placing workloads in a multi-array system
US10572460B2 (en) 2016-02-11 2020-02-25 Pure Storage, Inc. Compressing data in dependence upon characteristics of a storage system
US11748322B2 (en) 2016-02-11 2023-09-05 Pure Storage, Inc. Utilizing different data compression algorithms based on characteristics of a storage system
US11392565B1 (en) 2016-02-11 2022-07-19 Pure Storage, Inc. Optimizing data compression in a storage system
US10884666B1 (en) 2016-02-12 2021-01-05 Pure Storage, Inc. Dynamic path selection in a storage network
US10001951B1 (en) 2016-02-12 2018-06-19 Pure Storage, Inc. Path selection in a data storage system
US9760297B2 (en) 2016-02-12 2017-09-12 Pure Storage, Inc. Managing input/output (‘I/O’) queues in a data storage system
US11561730B1 (en) 2016-02-12 2023-01-24 Pure Storage, Inc. Selecting paths between a host and a storage system
US10289344B1 (en) 2016-02-12 2019-05-14 Pure Storage, Inc. Bandwidth-based path selection in a storage network
US11340785B1 (en) 2016-03-16 2022-05-24 Pure Storage, Inc. Upgrading data in a storage system using background processes
US10768815B1 (en) 2016-03-16 2020-09-08 Pure Storage, Inc. Upgrading a storage system
US9959043B2 (en) 2016-03-16 2018-05-01 Pure Storage, Inc. Performing a non-disruptive upgrade of data in a storage system
US11112990B1 (en) 2016-04-27 2021-09-07 Pure Storage, Inc. Managing storage device evacuation
US11809727B1 (en) 2016-04-27 2023-11-07 Pure Storage, Inc. Predicting failures in a storage system that includes a plurality of storage devices
US9841921B2 (en) 2016-04-27 2017-12-12 Pure Storage, Inc. Migrating data in a storage array that includes a plurality of storage devices
US11934681B2 (en) 2016-04-27 2024-03-19 Pure Storage, Inc. Data migration for write groups
US10564884B1 (en) 2016-04-27 2020-02-18 Pure Storage, Inc. Intelligent data migration within a flash storage array
US10996859B1 (en) 2016-04-28 2021-05-04 Pure Storage, Inc. Utilizing redundant resources in a storage system
US10545676B1 (en) 2016-04-28 2020-01-28 Pure Storage, Inc. Providing high availability to client-specific applications executing in a storage system
US9811264B1 (en) 2016-04-28 2017-11-07 Pure Storage, Inc. Deploying client-specific applications in a storage system utilizing redundant system resources
US11461009B2 (en) 2016-04-28 2022-10-04 Pure Storage, Inc. Supporting applications across a fleet of storage systems
US10303390B1 (en) 2016-05-02 2019-05-28 Pure Storage, Inc. Resolving fingerprint collisions in flash storage system
US10620864B1 (en) 2016-05-02 2020-04-14 Pure Storage, Inc. Improving the accuracy of in-line data deduplication
US11231858B2 (en) 2016-05-19 2022-01-25 Pure Storage, Inc. Dynamically configuring a storage system to facilitate independent scaling of resources
US10642524B1 (en) 2016-05-20 2020-05-05 Pure Storage, Inc. Upgrading a write buffer in a storage system that includes a plurality of storage devices and a plurality of write buffer devices
US10078469B1 (en) 2016-05-20 2018-09-18 Pure Storage, Inc. Preparing for cache upgrade in a storage array that includes a plurality of storage devices and a plurality of write buffer devices
US9817603B1 (en) 2016-05-20 2017-11-14 Pure Storage, Inc. Data migration in a storage array that includes a plurality of storage devices
US10691567B2 (en) 2016-06-03 2020-06-23 Pure Storage, Inc. Dynamically forming a failure domain in a storage system that includes a plurality of blades
US11126516B2 (en) 2016-06-03 2021-09-21 Pure Storage, Inc. Dynamic formation of a failure domain
US10452310B1 (en) 2016-07-13 2019-10-22 Pure Storage, Inc. Validating cabling for storage component admission to a storage array
US11706895B2 (en) 2016-07-19 2023-07-18 Pure Storage, Inc. Independent scaling of compute resources and storage resources in a storage system
US10459652B2 (en) 2016-07-27 2019-10-29 Pure Storage, Inc. Evacuating blades in a storage array that includes a plurality of blades
US10474363B1 (en) 2016-07-29 2019-11-12 Pure Storage, Inc. Space reporting in a storage system
US11630585B1 (en) 2016-08-25 2023-04-18 Pure Storage, Inc. Processing evacuation events in a storage array that includes a plurality of storage devices
US11481261B1 (en) 2016-09-07 2022-10-25 Pure Storage, Inc. Preventing extended latency in a storage system
US11921567B2 (en) 2016-09-07 2024-03-05 Pure Storage, Inc. Temporarily preventing access to a storage device
US11886922B2 (en) 2016-09-07 2024-01-30 Pure Storage, Inc. Scheduling input/output operations for a storage system
US10671439B1 (en) 2016-09-07 2020-06-02 Pure Storage, Inc. Workload planning with quality-of-service (‘QOS’) integration
US10146585B2 (en) 2016-09-07 2018-12-04 Pure Storage, Inc. Ensuring the fair utilization of system resources using workload based, time-independent scheduling
US10896068B1 (en) 2016-09-07 2021-01-19 Pure Storage, Inc. Ensuring the fair utilization of system resources using workload based, time-independent scheduling
US10908966B1 (en) 2016-09-07 2021-02-02 Pure Storage, Inc. Adapting target service times in a storage system
US11803492B2 (en) 2016-09-07 2023-10-31 Pure Storage, Inc. System resource management using time-independent scheduling
US11449375B1 (en) 2016-09-07 2022-09-20 Pure Storage, Inc. Performing rehabilitative actions on storage devices
US10353743B1 (en) 2016-09-07 2019-07-16 Pure Storage, Inc. System resource utilization balancing in a storage system
US11914455B2 (en) 2016-09-07 2024-02-27 Pure Storage, Inc. Addressing storage device performance
US10534648B2 (en) 2016-09-07 2020-01-14 Pure Storage, Inc. System resource utilization balancing
US10585711B2 (en) 2016-09-07 2020-03-10 Pure Storage, Inc. Crediting entity utilization of system resources
US11531577B1 (en) 2016-09-07 2022-12-20 Pure Storage, Inc. Temporarily limiting access to a storage device
US10853281B1 (en) 2016-09-07 2020-12-01 Pure Storage, Inc. Administration of storage system resource utilization
US11789780B1 (en) 2016-09-07 2023-10-17 Pure Storage, Inc. Preserving quality-of-service (‘QOS’) to storage system workloads
US11960348B2 (en) 2016-09-07 2024-04-16 Pure Storage, Inc. Cloud-based monitoring of hardware components in a fleet of storage systems
US10963326B1 (en) 2016-09-07 2021-03-30 Pure Storage, Inc. Self-healing storage devices
US10235229B1 (en) 2016-09-07 2019-03-19 Pure Storage, Inc. Rehabilitating storage devices in a storage array that includes a plurality of storage devices
US11520720B1 (en) 2016-09-07 2022-12-06 Pure Storage, Inc. Weighted resource allocation for workload scheduling
US10331588B2 (en) 2016-09-07 2019-06-25 Pure Storage, Inc. Ensuring the appropriate utilization of system resources using weighted workload based, time-independent scheduling
US10331370B2 (en) 2016-10-20 2019-06-25 Pure Storage, Inc. Tuning a storage system in dependence upon workload access patterns
US10007459B2 (en) 2016-10-20 2018-06-26 Pure Storage, Inc. Performance tuning in a storage system that includes one or more storage devices
US11379132B1 (en) 2016-10-20 2022-07-05 Pure Storage, Inc. Correlating medical sensor data
US10162566B2 (en) 2016-11-22 2018-12-25 Pure Storage, Inc. Accumulating application-level statistics in a storage system
US11016700B1 (en) 2016-11-22 2021-05-25 Pure Storage, Inc. Analyzing application-specific consumption of storage system resources
US10416924B1 (en) 2016-11-22 2019-09-17 Pure Storage, Inc. Identifying workload characteristics in dependence upon storage utilization
US11620075B2 (en) 2016-11-22 2023-04-04 Pure Storage, Inc. Providing application aware storage
US11061573B1 (en) 2016-12-19 2021-07-13 Pure Storage, Inc. Accelerating write operations in a storage system
US11687259B2 (en) 2016-12-19 2023-06-27 Pure Storage, Inc. Reconfiguring a storage system based on resource availability
US10198205B1 (en) 2016-12-19 2019-02-05 Pure Storage, Inc. Dynamically adjusting a number of storage devices utilized to simultaneously service write operations
US11461273B1 (en) 2016-12-20 2022-10-04 Pure Storage, Inc. Modifying storage distribution in a storage system that includes one or more storage devices
US10574454B1 (en) 2017-01-05 2020-02-25 Pure Storage, Inc. Current key data encryption
US10489307B2 (en) 2017-01-05 2019-11-26 Pure Storage, Inc. Periodically re-encrypting user data stored on a storage device
US11146396B1 (en) 2017-01-05 2021-10-12 Pure Storage, Inc. Data re-encryption in a storage system
US11762781B2 (en) 2017-01-09 2023-09-19 Pure Storage, Inc. Providing end-to-end encryption for data stored in a storage system
US10503700B1 (en) 2017-01-19 2019-12-10 Pure Storage, Inc. On-demand content filtering of snapshots within a storage system
US11340800B1 (en) 2017-01-19 2022-05-24 Pure Storage, Inc. Content masking in a storage system
US11861185B2 (en) 2017-01-19 2024-01-02 Pure Storage, Inc. Protecting sensitive data in snapshots
US11726850B2 (en) 2017-01-27 2023-08-15 Pure Storage, Inc. Increasing or decreasing the amount of log data generated based on performance characteristics of a device
US11163624B2 (en) 2017-01-27 2021-11-02 Pure Storage, Inc. Dynamically adjusting an amount of log data generated for a storage system
US10454810B1 (en) 2017-03-10 2019-10-22 Pure Storage, Inc. Managing host definitions across a plurality of storage systems
US11086555B1 (en) 2017-03-10 2021-08-10 Pure Storage, Inc. Synchronously replicating datasets
US10585733B1 (en) 2017-03-10 2020-03-10 Pure Storage, Inc. Determining active membership among storage systems synchronously replicating a dataset
US11645173B2 (en) 2017-03-10 2023-05-09 Pure Storage, Inc. Resilient mediation between storage systems replicating a dataset
US10613779B1 (en) 2017-03-10 2020-04-07 Pure Storage, Inc. Determining membership among storage systems synchronously replicating a dataset
US10558537B1 (en) 2017-03-10 2020-02-11 Pure Storage, Inc. Mediating between storage systems synchronously replicating a dataset
US11442825B2 (en) 2017-03-10 2022-09-13 Pure Storage, Inc. Establishing a synchronous replication relationship between two or more storage systems
US10990490B1 (en) 2017-03-10 2021-04-27 Pure Storage, Inc. Creating a synchronous replication lease between two or more storage systems
US11500745B1 (en) 2017-03-10 2022-11-15 Pure Storage, Inc. Issuing operations directed to synchronously replicated data
US11422730B1 (en) 2017-03-10 2022-08-23 Pure Storage, Inc. Recovery for storage systems synchronously replicating a dataset
US11716385B2 (en) 2017-03-10 2023-08-01 Pure Storage, Inc. Utilizing cloud-based storage systems to support synchronous replication of a dataset
US11789831B2 (en) 2017-03-10 2023-10-17 Pure Storage, Inc. Directing operations to synchronously replicated storage systems
US11941279B2 (en) 2017-03-10 2024-03-26 Pure Storage, Inc. Data path virtualization
US10671408B1 (en) 2017-03-10 2020-06-02 Pure Storage, Inc. Automatic storage system configuration for mediation services
US11797403B2 (en) 2017-03-10 2023-10-24 Pure Storage, Inc. Maintaining a synchronous replication relationship between two or more storage systems
US11698844B2 (en) 2017-03-10 2023-07-11 Pure Storage, Inc. Managing storage systems that are synchronously replicating a dataset
US11169727B1 (en) 2017-03-10 2021-11-09 Pure Storage, Inc. Synchronous replication between storage systems with virtualized storage
US11954002B1 (en) 2017-03-10 2024-04-09 Pure Storage, Inc. Automatically provisioning mediation services for a storage system
US10884993B1 (en) 2017-03-10 2021-01-05 Pure Storage, Inc. Synchronizing metadata among storage systems synchronously replicating a dataset
US11687423B2 (en) 2017-03-10 2023-06-27 Pure Storage, Inc. Prioritizing highly performant storage systems for servicing a synchronously replicated dataset
US11803453B1 (en) 2017-03-10 2023-10-31 Pure Storage, Inc. Using host connectivity states to avoid queuing I/O requests
US11379285B1 (en) 2017-03-10 2022-07-05 Pure Storage, Inc. Mediation for synchronous replication
US11210219B1 (en) 2017-03-10 2021-12-28 Pure Storage, Inc. Synchronously replicating a dataset across a plurality of storage systems
US11687500B1 (en) 2017-03-10 2023-06-27 Pure Storage, Inc. Updating metadata for a synchronously replicated dataset
US10680932B1 (en) 2017-03-10 2020-06-09 Pure Storage, Inc. Managing connectivity to synchronously replicated storage systems
US10365982B1 (en) 2017-03-10 2019-07-30 Pure Storage, Inc. Establishing a synchronous replication relationship between two or more storage systems
US11237927B1 (en) 2017-03-10 2022-02-01 Pure Storage, Inc. Resolving disruptions between storage systems replicating a dataset
US10503427B2 (en) 2017-03-10 2019-12-10 Pure Storage, Inc. Synchronously replicating datasets and other managed objects to cloud-based storage systems
US11347606B2 (en) 2017-03-10 2022-05-31 Pure Storage, Inc. Responding to a change in membership among storage systems synchronously replicating a dataset
US10521344B1 (en) 2017-03-10 2019-12-31 Pure Storage, Inc. Servicing input/output (‘I/O’) operations directed to a dataset that is synchronized across a plurality of storage systems
US11829629B2 (en) 2017-03-10 2023-11-28 Pure Storage, Inc. Synchronously replicating data using virtual volumes
US11675520B2 (en) 2017-03-10 2023-06-13 Pure Storage, Inc. Application replication among storage systems synchronously replicating a dataset
US10459664B1 (en) 2017-04-10 2019-10-29 Pure Storage, Inc. Virtualized copy-by-reference
US10534677B2 (en) 2017-04-10 2020-01-14 Pure Storage, Inc. Providing high availability for applications executing on a storage system
US11656804B2 (en) 2017-04-10 2023-05-23 Pure Storage, Inc. Copy using metadata representation
US11126381B1 (en) 2017-04-10 2021-09-21 Pure Storage, Inc. Lightweight copy
US9910618B1 (en) 2017-04-10 2018-03-06 Pure Storage, Inc. Migrating applications executing on a storage system
US11868629B1 (en) 2017-05-05 2024-01-09 Pure Storage, Inc. Storage system sizing service
US11567810B1 (en) 2017-06-12 2023-01-31 Pure Storage, Inc. Cost optimized workload placement
US11593036B2 (en) 2017-06-12 2023-02-28 Pure Storage, Inc. Staging data within a unified storage element
US11340939B1 (en) 2017-06-12 2022-05-24 Pure Storage, Inc. Application-aware analytics for storage systems
US11609718B1 (en) 2017-06-12 2023-03-21 Pure Storage, Inc. Identifying valid data after a storage system recovery
US10884636B1 (en) 2017-06-12 2021-01-05 Pure Storage, Inc. Presenting workload performance in a storage system
US10613791B2 (en) 2017-06-12 2020-04-07 Pure Storage, Inc. Portable snapshot replication between storage systems
US11016824B1 (en) 2017-06-12 2021-05-25 Pure Storage, Inc. Event identification with out-of-order reporting in a cloud-based environment
US11960777B2 (en) 2017-06-12 2024-04-16 Pure Storage, Inc. Utilizing multiple redundancy schemes within a unified storage element
US11422731B1 (en) 2017-06-12 2022-08-23 Pure Storage, Inc. Metadata-based replication of a dataset
US10853148B1 (en) 2017-06-12 2020-12-01 Pure Storage, Inc. Migrating workloads between a plurality of execution environments
US11210133B1 (en) 2017-06-12 2021-12-28 Pure Storage, Inc. Workload mobility between disparate execution environments
US10789020B2 (en) 2017-06-12 2020-09-29 Pure Storage, Inc. Recovering data within a unified storage element
US11561714B1 (en) 2017-07-05 2023-01-24 Pure Storage, Inc. Storage efficiency driven migration
US11477280B1 (en) 2017-07-26 2022-10-18 Pure Storage, Inc. Integrating cloud storage services
US11921908B2 (en) 2017-08-31 2024-03-05 Pure Storage, Inc. Writing data to compressed and encrypted volumes
US10417092B2 (en) 2017-09-07 2019-09-17 Pure Storage, Inc. Incremental RAID stripe update parity calculation
US11392456B1 (en) 2017-09-07 2022-07-19 Pure Storage, Inc. Calculating parity as a data stripe is modified
US11592991B2 (en) 2017-09-07 2023-02-28 Pure Storage, Inc. Converting raid data between persistent storage types
US10891192B1 (en) 2017-09-07 2021-01-12 Pure Storage, Inc. Updating raid stripe parity calculations
US11714718B2 (en) 2017-09-07 2023-08-01 Pure Storage, Inc. Performing partial redundant array of independent disks (RAID) stripe parity calculations
US10552090B2 (en) 2017-09-07 2020-02-04 Pure Storage, Inc. Solid state drives with multiple types of addressable memory
US11403290B1 (en) 2017-10-19 2022-08-02 Pure Storage, Inc. Managing an artificial intelligence infrastructure
US10452444B1 (en) 2017-10-19 2019-10-22 Pure Storage, Inc. Storage system with compute resources and shared storage resources
US11210140B1 (en) 2017-10-19 2021-12-28 Pure Storage, Inc. Data transformation delegation for a graphical processing unit (‘GPU’) server
US10275285B1 (en) 2017-10-19 2019-04-30 Pure Storage, Inc. Data transformation caching in an artificial intelligence infrastructure
US11861423B1 (en) 2017-10-19 2024-01-02 Pure Storage, Inc. Accelerating artificial intelligence (‘AI’) workflows
US11768636B2 (en) 2017-10-19 2023-09-26 Pure Storage, Inc. Generating a transformed dataset for use by a machine learning model in an artificial intelligence infrastructure
US10275176B1 (en) 2017-10-19 2019-04-30 Pure Storage, Inc. Data transformation offloading in an artificial intelligence infrastructure
US10671434B1 (en) 2017-10-19 2020-06-02 Pure Storage, Inc. Storage based artificial intelligence infrastructure
US11307894B1 (en) 2017-10-19 2022-04-19 Pure Storage, Inc. Executing a big data analytics pipeline using shared storage resources
US11556280B2 (en) 2017-10-19 2023-01-17 Pure Storage, Inc. Data transformation for a machine learning model
US11803338B2 (en) 2017-10-19 2023-10-31 Pure Storage, Inc. Executing a machine learning model in an artificial intelligence infrastructure
US11455168B1 (en) 2017-10-19 2022-09-27 Pure Storage, Inc. Batch building for deep learning training workloads
US10360214B2 (en) 2017-10-19 2019-07-23 Pure Storage, Inc. Ensuring reproducibility in an artificial intelligence infrastructure
US10671435B1 (en) 2017-10-19 2020-06-02 Pure Storage, Inc. Data transformation caching in an artificial intelligence infrastructure
US10649988B1 (en) 2017-10-19 2020-05-12 Pure Storage, Inc. Artificial intelligence and machine learning infrastructure
US11451391B1 (en) 2017-11-01 2022-09-20 Pure Storage, Inc. Encryption key management in a storage system
US11263096B1 (en) 2017-11-01 2022-03-01 Pure Storage, Inc. Preserving tolerance to storage device failures in a storage system
US10484174B1 (en) 2017-11-01 2019-11-19 Pure Storage, Inc. Protecting an encryption key for data stored in a storage system that includes a plurality of storage devices
US10671494B1 (en) 2017-11-01 2020-06-02 Pure Storage, Inc. Consistent selection of replicated datasets during storage system recovery
US10509581B1 (en) 2017-11-01 2019-12-17 Pure Storage, Inc. Maintaining write consistency in a multi-threaded storage system
US11663097B2 (en) 2017-11-01 2023-05-30 Pure Storage, Inc. Mirroring data to survive storage device failures
US10467107B1 (en) 2017-11-01 2019-11-05 Pure Storage, Inc. Maintaining metadata resiliency among storage device failures
US10817392B1 (en) 2017-11-01 2020-10-27 Pure Storage, Inc. Ensuring resiliency to storage device failures in a storage system that includes a plurality of storage devices
US11500724B1 (en) 2017-11-21 2022-11-15 Pure Storage, Inc. Flexible parity information for storage systems
US10929226B1 (en) 2017-11-21 2021-02-23 Pure Storage, Inc. Providing for increased flexibility for large scale parity
US11847025B2 (en) 2017-11-21 2023-12-19 Pure Storage, Inc. Storage system parity based on system characteristics
US10936238B2 (en) 2017-11-28 2021-03-02 Pure Storage, Inc. Hybrid data tiering
US11604583B2 (en) 2017-11-28 2023-03-14 Pure Storage, Inc. Policy based data tiering
US10990282B1 (en) 2017-11-28 2021-04-27 Pure Storage, Inc. Hybrid data tiering with cloud storage
US10795598B1 (en) 2017-12-07 2020-10-06 Pure Storage, Inc. Volume migration for storage systems synchronously replicating a dataset
US11579790B1 (en) 2017-12-07 2023-02-14 Pure Storage, Inc. Servicing input/output (‘I/O’) operations during data migration
US11036677B1 (en) 2017-12-14 2021-06-15 Pure Storage, Inc. Replicated data integrity
US11089105B1 (en) 2017-12-14 2021-08-10 Pure Storage, Inc. Synchronously replicating datasets in cloud-based storage systems
US11782614B1 (en) 2017-12-21 2023-10-10 Pure Storage, Inc. Encrypting data to optimize data reduction
US10992533B1 (en) 2018-01-30 2021-04-27 Pure Storage, Inc. Policy based path management
US11296944B2 (en) 2018-01-30 2022-04-05 Pure Storage, Inc. Updating path selection as paths between a computing device and a storage system change
US11861170B2 (en) 2018-03-05 2024-01-02 Pure Storage, Inc. Sizing resources for a replication target
US11150834B1 (en) 2018-03-05 2021-10-19 Pure Storage, Inc. Determining storage consumption in a storage system
US10521151B1 (en) 2018-03-05 2019-12-31 Pure Storage, Inc. Determining effective space utilization in a storage system
US10942650B1 (en) 2018-03-05 2021-03-09 Pure Storage, Inc. Reporting capacity utilization in a storage system
US11614881B2 (en) 2018-03-05 2023-03-28 Pure Storage, Inc. Calculating storage consumption for distinct client entities
US11474701B1 (en) 2018-03-05 2022-10-18 Pure Storage, Inc. Determining capacity consumption in a deduplicating storage system
US11836349B2 (en) 2018-03-05 2023-12-05 Pure Storage, Inc. Determining storage capacity utilization based on deduplicated data
US11112989B2 (en) 2018-03-09 2021-09-07 Pure Storage, Inc. Utilizing a decentralized storage network for data storage
US10296258B1 (en) 2018-03-09 2019-05-21 Pure Storage, Inc. Offloading data storage to a decentralized storage network
US11533364B1 (en) 2018-03-15 2022-12-20 Pure Storage, Inc. Maintaining metadata associated with a replicated dataset
US10917471B1 (en) 2018-03-15 2021-02-09 Pure Storage, Inc. Active membership in a cloud-based storage system
US10924548B1 (en) 2018-03-15 2021-02-16 Pure Storage, Inc. Symmetric storage using a cloud-based storage system
US11539793B1 (en) 2018-03-15 2022-12-27 Pure Storage, Inc. Responding to membership changes to a set of storage systems that are synchronously replicating a dataset
US11442669B1 (en) 2018-03-15 2022-09-13 Pure Storage, Inc. Orchestrating a virtual storage system
US11288138B1 (en) 2018-03-15 2022-03-29 Pure Storage, Inc. Recovery from a system fault in a cloud-based storage system
US11210009B1 (en) 2018-03-15 2021-12-28 Pure Storage, Inc. Staging data in a cloud-based storage system
US11698837B2 (en) 2018-03-15 2023-07-11 Pure Storage, Inc. Consistent recovery of a dataset
US11048590B1 (en) 2018-03-15 2021-06-29 Pure Storage, Inc. Data consistency during recovery in a cloud-based storage system
US10976962B2 (en) 2018-03-15 2021-04-13 Pure Storage, Inc. Servicing I/O operations in a cloud-based storage system
US11838359B2 (en) 2018-03-15 2023-12-05 Pure Storage, Inc. Synchronizing metadata in a cloud-based storage system
US11704202B2 (en) 2018-03-15 2023-07-18 Pure Storage, Inc. Recovering from system faults for replicated datasets
US11171950B1 (en) 2018-03-21 2021-11-09 Pure Storage, Inc. Secure cloud-based storage system management
US11888846B2 (en) 2018-03-21 2024-01-30 Pure Storage, Inc. Configuring storage systems in a fleet of storage systems
US11729251B2 (en) 2018-03-21 2023-08-15 Pure Storage, Inc. Remote and secure management of a storage system
US11095706B1 (en) 2018-03-21 2021-08-17 Pure Storage, Inc. Secure cloud-based storage system management
US10838833B1 (en) 2018-03-26 2020-11-17 Pure Storage, Inc. Providing for high availability in a data analytics pipeline without replicas
US11494692B1 (en) 2018-03-26 2022-11-08 Pure Storage, Inc. Hyperscale artificial intelligence and machine learning infrastructure
US11714728B2 (en) 2018-03-26 2023-08-01 Pure Storage, Inc. Creating a highly available data analytics pipeline without replicas
US11263095B1 (en) 2018-03-26 2022-03-01 Pure Storage, Inc. Managing a data analytics pipeline
US11392553B1 (en) 2018-04-24 2022-07-19 Pure Storage, Inc. Remote data management
US11436344B1 (en) 2018-04-24 2022-09-06 Pure Storage, Inc. Secure encryption in deduplication cluster
US10992598B2 (en) 2018-05-21 2021-04-27 Pure Storage, Inc. Synchronously replicating when a mediation service becomes unavailable
US11954220B2 (en) 2018-05-21 2024-04-09 Pure Storage, Inc. Data protection for container storage
US11677687B2 (en) 2018-05-21 2023-06-13 Pure Storage, Inc. Switching between fault response models in a storage system
US11757795B2 (en) 2018-05-21 2023-09-12 Pure Storage, Inc. Resolving mediator unavailability
US11128578B2 (en) 2018-05-21 2021-09-21 Pure Storage, Inc. Switching between mediator services for a storage system
US11455409B2 (en) 2018-05-21 2022-09-27 Pure Storage, Inc. Storage layer data obfuscation
US11675503B1 (en) 2018-05-21 2023-06-13 Pure Storage, Inc. Role-based data access
US11748030B1 (en) 2018-05-22 2023-09-05 Pure Storage, Inc. Storage system metric optimization for container orchestrators
US10871922B2 (en) 2018-05-22 2020-12-22 Pure Storage, Inc. Integrated storage management between storage systems and container orchestrators
US11416298B1 (en) 2018-07-20 2022-08-16 Pure Storage, Inc. Providing application-specific storage by a storage system
US11403000B1 (en) 2018-07-20 2022-08-02 Pure Storage, Inc. Resiliency in a cloud-based storage system
US11146564B1 (en) 2018-07-24 2021-10-12 Pure Storage, Inc. Login authentication in a cloud storage platform
US11632360B1 (en) 2018-07-24 2023-04-18 Pure Storage, Inc. Remote access to a storage device
US11954238B1 (en) 2018-07-24 2024-04-09 Pure Storage, Inc. Role-based access control for a storage system
US11860820B1 (en) 2018-09-11 2024-01-02 Pure Storage, Inc. Processing data through a storage system in a data pipeline
US10671302B1 (en) 2018-10-26 2020-06-02 Pure Storage, Inc. Applying a rate limit across a plurality of storage systems
US11586365B2 (en) 2018-10-26 2023-02-21 Pure Storage, Inc. Applying a rate limit across a plurality of storage systems
US10990306B1 (en) 2018-10-26 2021-04-27 Pure Storage, Inc. Bandwidth sharing for paired storage systems
US11340837B1 (en) 2018-11-18 2022-05-24 Pure Storage, Inc. Storage system management via a remote console
US11184233B1 (en) 2018-11-18 2021-11-23 Pure Storage, Inc. Non-disruptive upgrades to a cloud-based storage system
US11822825B2 (en) 2018-11-18 2023-11-21 Pure Storage, Inc. Distributed cloud-based storage system
US10917470B1 (en) 2018-11-18 2021-02-09 Pure Storage, Inc. Cloning storage systems in a cloud computing environment
US11455126B1 (en) 2018-11-18 2022-09-27 Pure Storage, Inc. Copying a cloud-based storage system
US11379254B1 (en) 2018-11-18 2022-07-05 Pure Storage, Inc. Dynamic configuration of a cloud-based storage system
US11768635B2 (en) 2018-11-18 2023-09-26 Pure Storage, Inc. Scaling storage resources in a storage volume
US11941288B1 (en) 2018-11-18 2024-03-26 Pure Storage, Inc. Servicing write operations in a cloud-based storage system
US10963189B1 (en) 2018-11-18 2021-03-30 Pure Storage, Inc. Coalescing write operations in a cloud-based storage system
US11861235B2 (en) 2018-11-18 2024-01-02 Pure Storage, Inc. Maximizing data throughput in a cloud-based storage system
US11526405B1 (en) 2018-11-18 2022-12-13 Pure Storage, Inc. Cloud-based disaster recovery
US11907590B2 (en) 2018-11-18 2024-02-20 Pure Storage, Inc. Using infrastructure-as-code (‘IaC’) to update a cloud-based storage system
US11928366B2 (en) 2018-11-18 2024-03-12 Pure Storage, Inc. Scaling a cloud-based storage system in response to a change in workload
US11023179B2 (en) 2018-11-18 2021-06-01 Pure Storage, Inc. Cloud-based storage system storage management
US11650749B1 (en) 2018-12-17 2023-05-16 Pure Storage, Inc. Controlling access to sensitive data in a shared dataset
US11003369B1 (en) 2019-01-14 2021-05-11 Pure Storage, Inc. Performing a tune-up procedure on a storage device during a boot process
US11947815B2 (en) 2019-01-14 2024-04-02 Pure Storage, Inc. Configuring a flash-based storage device
US11042452B1 (en) 2019-03-20 2021-06-22 Pure Storage, Inc. Storage system data recovery using data recovery as a service
US11221778B1 (en) 2019-04-02 2022-01-11 Pure Storage, Inc. Preparing data for deduplication
US11068162B1 (en) 2019-04-09 2021-07-20 Pure Storage, Inc. Storage management in a cloud data store
US11640239B2 (en) 2019-04-09 2023-05-02 Pure Storage, Inc. Cost conscious garbage collection
US11392555B2 (en) 2019-05-15 2022-07-19 Pure Storage, Inc. Cloud-based file services
US11853266B2 (en) 2019-05-15 2023-12-26 Pure Storage, Inc. Providing a file system in a cloud environment
US11487715B1 (en) 2019-07-18 2022-11-01 Pure Storage, Inc. Resiliency in a cloud-based storage system
US11327676B1 (en) 2019-07-18 2022-05-10 Pure Storage, Inc. Predictive data streaming in a virtual storage system
US11093139B1 (en) 2019-07-18 2021-08-17 Pure Storage, Inc. Durably storing data within a virtual storage system
US11126364B2 (en) 2019-07-18 2021-09-21 Pure Storage, Inc. Virtual storage system architecture
US11797197B1 (en) 2019-07-18 2023-10-24 Pure Storage, Inc. Dynamic scaling of a virtual storage system
US11526408B2 (en) 2019-07-18 2022-12-13 Pure Storage, Inc. Data recovery in a virtual storage system
US11861221B1 (en) 2019-07-18 2024-01-02 Pure Storage, Inc. Providing scalable and reliable container-based storage services
US11550514B2 (en) 2019-07-18 2023-01-10 Pure Storage, Inc. Efficient transfers between tiers of a virtual storage system
US11086553B1 (en) 2019-08-28 2021-08-10 Pure Storage, Inc. Tiering duplicated objects in a cloud-based object store
US11693713B1 (en) 2019-09-04 2023-07-04 Pure Storage, Inc. Self-tuning clusters for resilient microservices
US11360689B1 (en) 2019-09-13 2022-06-14 Pure Storage, Inc. Cloning a tracking copy of replica data
US11797569B2 (en) 2019-09-13 2023-10-24 Pure Storage, Inc. Configurable data replication
US11625416B1 (en) 2019-09-13 2023-04-11 Pure Storage, Inc. Uniform model for distinct types of data replication
US11704044B2 (en) 2019-09-13 2023-07-18 Pure Storage, Inc. Modifying a cloned image of replica data
US11573864B1 (en) 2019-09-16 2023-02-07 Pure Storage, Inc. Automating database management in a storage system
US11669386B1 (en) 2019-10-08 2023-06-06 Pure Storage, Inc. Managing an application's resource stack
US11943293B1 (en) 2019-12-06 2024-03-26 Pure Storage, Inc. Restoring a storage system from a replication target
US11868318B1 (en) 2019-12-06 2024-01-09 Pure Storage, Inc. End-to-end encryption in a storage system with multi-tenancy
US11947683B2 (en) 2019-12-06 2024-04-02 Pure Storage, Inc. Replicating a storage system
US11531487B1 (en) 2019-12-06 2022-12-20 Pure Storage, Inc. Creating a replica of a storage system
US11930112B1 (en) 2019-12-06 2024-03-12 Pure Storage, Inc. Multi-path end-to-end encryption in a storage system
US11720497B1 (en) 2020-01-13 2023-08-08 Pure Storage, Inc. Inferred nonsequential prefetch based on data access patterns
US11709636B1 (en) 2020-01-13 2023-07-25 Pure Storage, Inc. Non-sequential readahead for deep learning training
US11733901B1 (en) 2020-01-13 2023-08-22 Pure Storage, Inc. Providing persistent storage to transient cloud computing services
US11868622B2 (en) 2020-02-25 2024-01-09 Pure Storage, Inc. Application recovery across storage systems
US11637896B1 (en) 2020-02-25 2023-04-25 Pure Storage, Inc. Migrating applications to a cloud-computing environment
US11321006B1 (en) 2020-03-25 2022-05-03 Pure Storage, Inc. Data loss prevention during transitions from a replication source
US11625185B2 (en) 2020-03-25 2023-04-11 Pure Storage, Inc. Transitioning between replication sources for data replication operations
US11630598B1 (en) 2020-04-06 2023-04-18 Pure Storage, Inc. Scheduling data replication operations
US11301152B1 (en) 2020-04-06 2022-04-12 Pure Storage, Inc. Intelligently moving data between storage systems
US11494267B2 (en) 2020-04-14 2022-11-08 Pure Storage, Inc. Continuous value data redundancy
US11853164B2 (en) 2020-04-14 2023-12-26 Pure Storage, Inc. Generating recovery information using data redundancy
US11921670B1 (en) 2020-04-20 2024-03-05 Pure Storage, Inc. Multivariate data backup retention policies
US11431488B1 (en) 2020-06-08 2022-08-30 Pure Storage, Inc. Protecting local key generation using a remote key management service
US11882179B2 (en) 2020-07-23 2024-01-23 Pure Storage, Inc. Supporting multiple replication schemes across distinct network layers
US11789638B2 (en) 2020-07-23 2023-10-17 Pure Storage, Inc. Continuing replication during storage system transportation
US11349917B2 (en) 2020-07-23 2022-05-31 Pure Storage, Inc. Replication handling among distinct networks
US11442652B1 (en) 2020-07-23 2022-09-13 Pure Storage, Inc. Replication handling during storage system transportation
US11397545B1 (en) 2021-01-20 2022-07-26 Pure Storage, Inc. Emulating persistent reservations in a cloud-based storage system
US11693604B2 (en) 2021-01-20 2023-07-04 Pure Storage, Inc. Administering storage access in a cloud-based storage system
US11853285B1 (en) 2021-01-22 2023-12-26 Pure Storage, Inc. Blockchain logging of volume-level events in a storage system
US11588716B2 (en) 2021-05-12 2023-02-21 Pure Storage, Inc. Adaptive storage processing for storage-as-a-service
US11822809B2 (en) 2021-05-12 2023-11-21 Pure Storage, Inc. Role enforcement for storage-as-a-service
CN113282246A (en) * 2021-06-15 2021-08-20 杭州海康威视数字技术股份有限公司 Data processing method and device
US11816129B2 (en) 2021-06-22 2023-11-14 Pure Storage, Inc. Generating datasets using approximate baselines
US11914867B2 (en) 2021-10-29 2024-02-27 Pure Storage, Inc. Coordinated snapshots among storage systems implementing a promotion/demotion model
US11714723B2 (en) 2021-10-29 2023-08-01 Pure Storage, Inc. Coordinated snapshots for data stored across distinct storage environments
US11893263B2 (en) 2021-10-29 2024-02-06 Pure Storage, Inc. Coordinated checkpoints among storage systems implementing checkpoint-based replication
US11922052B2 (en) 2021-12-15 2024-03-05 Pure Storage, Inc. Managing links between storage objects
US11847071B2 (en) 2021-12-30 2023-12-19 Pure Storage, Inc. Enabling communication between a single-port device and multiple storage system controllers
US11972134B2 (en) 2022-01-12 2024-04-30 Pure Storage, Inc. Resource utilization using normalized input/output (‘I/O’) operations
US11860780B2 (en) 2022-01-28 2024-01-02 Pure Storage, Inc. Storage cache management
US11886295B2 (en) 2022-01-31 2024-01-30 Pure Storage, Inc. Intra-block error correction

Also Published As

Publication number Publication date
US20120209905A1 (en) 2012-08-16
US8898224B2 (en) 2014-11-25
US9645764B2 (en) 2017-05-09
US20150026371A1 (en) 2015-01-22

Similar Documents

Publication Publication Date Title
US9645764B2 (en) Techniques for migrating active I/O connections with migrating servers and clients
US11188254B2 (en) Using a data mover and a clone blocklist primitive to clone files on a virtual file system
US10389852B2 (en) Method and system for providing a roaming remote desktop
US9928091B2 (en) Techniques for streaming virtual machines from a server to a host
US11829793B2 (en) Unified management of virtual machines and bare metal computers
US8086808B2 (en) Method and system for migration between physical and virtual systems
US9116737B2 (en) Conversion of virtual disk snapshots between redo and copy-on-write technologies
TWI439867B (en) Dynamic physical and virtual multipath i/o
US8261268B1 (en) System and method for dynamic allocation of virtual machines in a virtual server environment
US8312471B2 (en) File system independent content aware cache
US8880687B1 (en) Detecting and managing idle virtual storage servers
US9575786B2 (en) System and method for raw device mapping in traditional NAS subsystems
US10120618B2 (en) Systems and methods for adaptive offloading of mass storage data movement
US20120331242A1 (en) Consistent unmapping of application data in presence of concurrent, unquiesced writers and readers
US20130138764A1 (en) Method and system for virtual machine data migration
US9304875B2 (en) Dynamically tracking logical units moving between input/output ports of a storage area network target
US11163463B2 (en) Non-disruptive migration of a virtual volume in a clustered data storage system
WO2009150122A1 (en) Systems and methods for virtualizing storage for wpar clients using node port id virtuali zation
US11138028B1 (en) Hot growing a cloud hosted block device
US10534566B1 (en) Cloud storage tiering using application programming interface
US8838768B2 (en) Computer system and disk sharing method used thereby
US20240012566A1 (en) Storage device write performance in remote computing environments

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HAUGH, JULIANNE F.;MCBREARTY, GERALD F.;REEL/FRAME:024916/0115

Effective date: 20100827

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: AIRBNB, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:056427/0193

Effective date: 20210106