US20130080623A1 - Dynamic route requests for multiple clouds - Google Patents

Dynamic route requests for multiple clouds Download PDF

Info

Publication number
US20130080623A1
US20130080623A1 US13/325,352 US201113325352A US2013080623A1 US 20130080623 A1 US20130080623 A1 US 20130080623A1 US 201113325352 A US201113325352 A US 201113325352A US 2013080623 A1 US2013080623 A1 US 2013080623A1
Authority
US
United States
Prior art keywords
cloud computing
requests
request
routing
computing networks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/325,352
Inventor
Jason Thireault
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Edgio Inc
Original Assignee
Limelight Networks Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Limelight Networks Inc filed Critical Limelight Networks Inc
Priority to US13/325,352 priority Critical patent/US20130080623A1/en
Publication of US20130080623A1 publication Critical patent/US20130080623A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • G06F9/505Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals considering the load
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/145Network analysis or design involving simulating, designing, planning or modelling of a network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/40Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks using virtualisation of network functions or resources, e.g. SDN or NFV entities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/50Network service management, e.g. ensuring proper service fulfilment according to agreements
    • H04L41/508Network service management, e.g. ensuring proper service fulfilment according to agreements based on type of value added network service under agreement
    • H04L41/5096Network service management, e.g. ensuring proper service fulfilment according to agreements based on type of value added network service under agreement wherein the managed service relates to distributed or central networked applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L61/00Network arrangements, protocols or services for addressing or naming
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1004Server selection for load balancing
    • H04L67/1014Server selection for load balancing based on the content of a request
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/60Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources
    • H04L67/63Routing a service request depending on the request content or context
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L45/00Routing or path finding of packets in data switching networks
    • H04L45/70Routing based on monitoring results
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network

Definitions

  • compute resources i.e., applications, etc.
  • a cloud provider's network e.g., servers which are aggregated in a centralized location, a datacenter, etc.
  • All requests are load-balanced back to that cluster.
  • a cluster e.g., servers which are aggregated in a centralized location, a datacenter, etc.
  • All requests are load-balanced back to that cluster.
  • such an implementation does not necessarily provide the best performance or experience for end users who may, for example, be located far away from the centralized cluster.
  • a data center 105 includes a compute platform 110 which is in communication with devices which produce user requests 115 .
  • user requests 115 are received by the data center 105 , which includes the cloud resources.
  • the compute platform 110 As requests increase, software and services within the data center 105 are spun-up by additional cloud resources using the compute platform 110 .
  • the distance between the compute platform 110 and the user requests 115 may be great, and therefore, responsiveness and user experience are diminished greatly.
  • aspects of the present invention distribute new resources closer to end-users which are requesting the resource.
  • additional virtualized instances are spun-up within servers that are physically near to the network equipment (i.e., web servers, switches, routers, load balancers) which are receiving the requests. Accordingly, by moving computational resources closer to the requesting users in cloud computing environments, the user experience is significantly enhanced.
  • Further aspects of the present invention include dynamically routing requests for applications to one of multiple cloud computing environments.
  • the method may dynamically route an application request to an application that is hosted in multiple clouds (deployed within a management application) based upon a specified criteria.
  • the routing of requests for the application to a specific cloud in which the application is deployed may be based upon a criteria(s) that the application owner specifies. This may provide the application owner an ability to positively affect quality of service (QoS) for application delivery, ensure uninterrupted access to the application in the event of failure by one or more clouds, and provide more efficient application performance.
  • QoS quality of service
  • FIG. 1 shows a system for implementing cloud computing.
  • FIG. 2 shows a system for implementing edge-based resource spin-up for cloud computing.
  • FIG. 3 shows a method of implementing edge-based resource spin-up for cloud computing.
  • FIGS. 4A-4D show systems for implementing edge-based resource spin-up for cloud computing.
  • FIGS. 5A-5D show methods of implementing dynamic route requests for multiple clouds.
  • FIGS. 6A and 6B show systems for implementing dynamic route requests for multiple clouds.
  • FIG. 7 shows an embodiment of a content distribution system.
  • FIG. 8 shows an embodiment of a computer system.
  • similar components and/or features may have the same reference label.
  • components of the same type are identified by following a first reference label with a dash and a second reference label that further distinguishes among the similar components. If only the first reference label is used, the description is applicable to any of the similar components designated by the first reference label.
  • FIG. 2 shows a system for implementing edge-based resource spin-up for cloud computing, in accordance with one embodiment of the present invention.
  • edge-based resource spin-up includes carrying out computational activities within a cloud computing environment closer to the end user. As such, an increase in responsiveness as well as a more efficient use of resources is realized.
  • System 200 includes a data center 205 a and 205 b .
  • data centers 205 may be a facility used to house computer systems and associated components, such as telecommunications, networking systems, storage systems, etc. Furthermore, the data centers 205 may also be designated as points of presence (PoPs).
  • the data centers 205 a and 205 b may include edge servers 210 a and 210 b , respectively. Further, edge servers 210 a and 210 b may include compute platforms 215 a and 215 b , respectively. It should be noted that one skilled in the art would conclude that any number of data centers, edge serves, and/or compute platforms may be included, and only two of each are shown for ease of explanation and illustration.
  • system 200 may include a load balancer 220 in communication with both data centers 205 a and 205 b , as well as user devices issuing user requests 225 a and 225 b .
  • a load balancer 220 in communication with both data centers 205 a and 205 b , as well as user devices issuing user requests 225 a and 225 b .
  • many user requests may be received, and proper allocation and division of cloud resources should be allocated to handle the requests.
  • many of the requests are time sensitive and latency sensitive (i.e., UI intensive applications, computation intensive applications, etc.), so ensuring fast response times to requests can be important.
  • the load balancer 220 is configured to determine the “fastest” responding edge server/compute platform to direct the request.
  • fastest response time means the edge server closest physically to the requesting user device.
  • fastest may mean the edge server with the lowest latency relative to the requesting device.
  • the closest and the lowest latency edge server may be the same server, but not always. For example, if the physically closest edge server is experiencing a heavy load of traffic and requests, the response time and/or network latency of the server may outweigh the physically close proximity to the requesting device.
  • the load balancer 220 is configured to ensure that the needed resources to respond to the user requests 225 a and 225 b are routed to the edge servers 210 a and 201 b which will provide the fastest response time for the request, which in many cases will be the edge server which is in the closest proximity to the requesting user device.
  • two groups of users make requests from two different geographical locations.
  • the load balancer 220 then receives the requests and, based on the location of the request, distributes the request to the data center 205 a or 205 b closest to the user (alternatively, the request may be routed to the data center which will provide the faster response time).
  • a “localized cloud instance” which is a de-centralized cloud computing environment with resources spun-up as physically close to the requesting device as possible.
  • such localized resources may be synchronized around the network to ensure that requests come to one localized cluster are treated in the same manner as other requests.
  • resources are spun-up in that locality based upon demand (i.e., subsequent user requests).
  • FIG. 3 shows a method 300 of implementing edge-based resource spin-up for cloud computing, in accordance with one embodiment of the present invention.
  • a request for data or a service may be received at an edge server from a user device.
  • the closest edge server to the requesting device may be determined by using an enhanced anycast methodology. Accordingly, the edge server which provides the fastest response time relative to the requesting user device is selected.
  • the request may be for an application, such as an enterprise application, a media application, etc.
  • the request may be for data, such a video file, a music file, a document, etc.
  • Each request may have associated information sent with the request which identifies the application and/or data used to service the request. The identification information may be embedded or attached to the request.
  • the edge server may extract the identification information. Then, based on the information, the edge server can identify the application/service used to process the request (process block 315 ).
  • the identification information may specifically identify the application by name or some other identifier, or alternatively the information may include an application type, etc.
  • the selected edge server may be in communication with one or more compute platforms, which may be co-located or remotely-located with the edge server.
  • each of the compute platforms may have one or more containers running which provide a virtual construct for allocating resources.
  • these containers may be a type of virtualized resource which is different from a virtualized instanced, such as elastic computing cloud (EC2).
  • EC2 elastic computing cloud
  • the containers are then configured to execute and maintain applications needed to service the user requests.
  • a determination is made whether a container maintained by a compute platform in communication with the selected edge server is running (or capable of running) the application necessary for servicing the user request.
  • all of the edge-based compute platforms may include the “DNA” for running an application (e.g., an XML dataset that specifies instructions for each application to be run in a container), and the determination for being able to run the application based on the current levels of utilization. As such, the allocation of the compute platform becomes a predictive determination.
  • a compute platform is capable of running the application if the compute platform has sufficient unused resources, if the necessary application is loaded on the compute platform, etc.
  • an available compute platform (or on other words, a compute platform which has available resources) is identified. Accordingly, it may not matter if the application is not currently running, as the application can be spun up; availability can be based on either a currently running application or the necessary capacity to support the application running, which could then be translated to actually spinning up a container, on demand, to support the requests. Then, one or more containers are spun-up by the identified compute platform to run the identified application or service (process block 335 ).
  • One example of an implementation of method 300 may be performed for the MediaTagTM application.
  • a user may click on a link/file that the user desires to purchase.
  • the file includes an associated cookie which is used to point the request to the MediaTag application.
  • the application makes a request of the cookie which has been stored on the user's machine by the website providing the music download; MediaTag then takes the cookie, explodes it, and carries out computational activity against the results.
  • the edge server then upon receiving the request interprets the tag and identifies a compute platform which is capable of spinning-up resources for the MediaTag application.
  • the determination may be based solely on geography—the closest POP with resources; there is sometimes a tradeoff between locality and capacity—as a system may chose to actually go to a more distant compute resource to carry out my request because the latency of serving the response is actually less than the latency caused in the local edge by the lack of capacity.
  • the compute platform spins-up a container running the MediaTag application.
  • the MediaTag application then creates a unique file based on the request (the file may include identification information, such as the username of the requester, the origination location, etc.). Again, alternatively, the choice may be based on both proximity and the current utilization level of that current proximal location; there is a tradeoff.
  • a response to the request is sent to the user (process block 345 ).
  • the system 400 may include a user device 405 .
  • the user device 405 may be a mobile device, a cellular device, a SmartPhone, a mobile computing platform, a user terminal, etc.
  • the user device 405 may be configured to send requests and access data, services, and applications. Further, the user device 405 may be in communication with a cloud computing network as shown in system 400 .
  • the user device 405 may be in communication with a point of presence (PoP) 410 .
  • PoP 410 may be configured as an access point to the Internet 430 , a physical location that houses servers, routers, ATM switches, digital/analog call aggregators, etc. Further, PoP 410 may be either part of the facilities of a telecommunications provider that the Internet service provider (ISP) rents or a location separate from the telecommunications provider. Generally, PoPs are also located at Internet exchange points and collocation centers.
  • An edge server 415 may be located within PoP 410 .
  • the edge server 415 may be operated by a cloud computing provider, or the like.
  • the edge server 415 may represent one of the cloud computing provider's closest connection points to the Internet 430 .
  • the edge server 415 is uniquely qualified to provide the fastest and most efficient service to the user device 405 , particularly with regard to spinning-up resources for use in a cloud computing environment.
  • the edge server 415 in communication with a compute platform 420 , may direct the compute platform 420 to handle the requests.
  • edge server 415 may direct compute platform 420 to spin-up a container 420 a to handle the incoming request.
  • spinning-up container 420 a includes allocating and assigning the necessary resources and applications to handle the request. For example, if the request is for application 420 b , then an instance of application 420 b (e.g., MediaTag, video player, gaming application, etc.) will be initiated in container 420 a .
  • application 420 b e.g., MediaTag, video player, gaming application, etc.
  • container 420 a is spun-up such that container 420 a is equipped to handle the request from user device 405 .
  • the PoP 410 may be selected as being the “closest” PoP to the user device 405 .
  • closest means the closest in physical proximity to the user device 405 , which in turn provides the fastest response time to requests, thus enhancing the cloud computing experience.
  • computations are routed away from the user device; however, in an edge-based cloud computing environment as in system 400 , the computations are performed as close as possible to the user device 405 .
  • rendering of cloud applications can be done without diminishing the experience (i.e., the application can be rendered as though the application is run “locally” on the user's device, or in a local area network or the like). This may be particularly important with regard to mobile users in that—as the user moves, so too may the “closest” available edge server “move”.
  • such a cloud computing configuration as in system 400 can provide a more efficient use of resources.
  • the de-centralized model allows for resources to be spun-up in order to handle specific user requests.
  • system 400 (additionally, systems 401 - 403 ) provides a scalable solution, such that as additional requests are received by compute platform 420 , the platform can direct additional containers to be spun-up to dynamically handle the increased load.
  • the containers allow for dynamically creating instances of the systems' operating system which has been “tuned” for a specific purpose. Therefore, the necessary resources are provided in response to the received requests, and the resources are provided at the closest location to the requesting device.
  • edge server 415 may be physically closer to user device 405 , but edge server 415 may be heavily congested.
  • edge server X (which is physically further from user device 405 ) may have a better effective distance or latency, and may ultimately be selected to process the request, despite its distance from user device 405 .
  • outages and other issues may be considered when determining which edge serve will actually (or effectively) provide the fastest response time and ultimately the most desirable user experience.
  • Such a cloud computing environment uniquely provides what a centralized cloud computing environment is unable to provide: scalability, efficiency, and faster response times.
  • configuration files may be used to determine the edge server and/or the compute platform to handle certain user requests. For example, a certain edge server may be closest to the user device 405 , but because the edge server does not have access to the requested application or other resource, a further edge server may need to be used. Therefore, the configuration file may provide such designation and mappings, such that requests are routed to edge servers and compute platforms that are actually equipped to handle the request. Additionally, the configuration files may also provide an accounting of the physical as well as the virtual resources available to each edge server, which can assist in routing decisions by not overloading edge servers above their resource capability capacity, etc.
  • An additional advantage of edge-based cloud computing may be that requests are able to be routed using URLs.
  • URLs uniquely allow for information to be appended to the URL which can provide the necessary information to the edge server for more efficient routing and resource allocation.
  • URL-based routing enables a variety of systems (e.g., anything capable of dealing with HTTP) the ability to forward the request along. This makes for a very flexible application architecture and a distributed computing environment in that individual application components, all making URL (i.e., HTTP-based) requests, can be completed from different compute resources, not necessarily all the same resource—the resources used for these requests can be fanned out.
  • FIG. 4B shows an alternative embodiment of an edge-based cloud computing configuration.
  • system 401 may have the compute platform 420 co-located with edge server 415 at PoP 410 .
  • multiple compute platforms 420 a , 420 b to 420 n may be provided.
  • the additional compute platforms may provide the edge server with access to additional resources in order to handle additional requests.
  • latency can be significantly reduced. Accordingly, computations for user requests are pushed even closer to the user device 405 originating the request. Accordingly, the edge is “super-charged” with readily available compute resources to meet specific types of requests.
  • new intelligence is added to the edge to better route request traffic (i.e., URLS) to other edge resources, different POPs, etc., depending upon performance and/or availability factors that reflect customer preferences.
  • URLS route request traffic
  • each of the compute platforms 420 a through 420 n are capable of spinning-up multiple containers 422 a through 422 n .
  • each of the containers 422 a through 422 n can also provide instances of applications 423 a through 423 n .
  • each compute platform 420 can expand or shrink to effectively and efficiently accommodate an increase or decrease in user requests.
  • the dynamic nature of resource allocation coupled with the relative closeness in proximity to the user device 405 provide for an optimal user experience in a cloud computing environment.
  • FIG. 4C shows another embodiment of an edge-based cloud computing environment.
  • System 402 shows an alternative configuration in which the compute platforms 420 a through 420 n are located at a compute server 435 , which may be remotely located from the edge server 415 and PoP 410 .
  • server space may be expensive within the PoP 410
  • the diminished response time (which is minimal) may be outweighed by the reduced cost of non-PoP space (in particular, if the same edge resources are being utilized for other purposes (i.e., streaming)).
  • the computational power at the edge server 415 can be significantly increased, and as such is able to handle an increased amount of requests; thus increasing the scalability and efficiency of the cloud computing environment.
  • FIG. 4D shows a further embodiment of an edge-based cloud computing environment.
  • a router or the like
  • the router 412 may be configured to route requests from the PoP 410 to the edge server 415 .
  • the edge server 415 can direct the requests to the appropriate compute platform 420 within the compute server 435 .
  • the compute server 435 may be co-located with edge server 415 , or alternatively the compute server 435 may be remotely-located from the edge server 415 .
  • Further aspects of this invention include dynamically routing requests for applications to one of multiple cloud computing environments.
  • the method may dynamically route an application request to an application that is hosted in multiple clouds (deployed within a management application) based upon a specified criteria.
  • the routing of requests for the application to a specific cloud in which the application is deployed may be based upon a criteria(s) that the application owner specifies. This may provide the application owner an ability to positively affect quality of service (QoS) for application delivery, ensure uninterrupted access to the application in the event of failure by one or more clouds, and provide more efficient application performance.
  • QoS quality of service
  • aspects of the present invention create a picture of that responsiveness to eventually enable a dynamic adjustment of business rules based on analysis of the data provided as part of the overall system.
  • One embodiment of the present invention may provide that a customer signs up for a “multi-cloud application deployment” which entails provisioning for a content delivery network (CDN) account (i.e., providing HTTP service or the like), as well as enabling the customer access to a portal (or other web-based UI) that allows the customer to specify the cloud-based locations of their applications, the URLs to those applications, the business rules the customer wants the applications to follow when shuttling requests to different clouds, etc.
  • CDN content delivery network
  • portal or other web-based UI
  • Each cloud that the customer specifies may require a unique hostname provided by the CDN.
  • a request may be received by the end user. This request may then be passed to the edge of the CDN via a CNAME or designated hostname provided by the CDN to the customer that configures their application URLs accordingly (process block 504 ).
  • the request may then be passed through a request analysis unit (or similar module) to gather metrics on the response provided by the application to which the request is being used. For example, a historical picture of specific cloud responses may be generated and developed.
  • a cloud control unit may then determine the cloud to which to direct the request.
  • the response from the cloud for the request is captured by a request analysis unit (or similar module) and the response is returned to the user (process block 512 ).
  • a historical picture of cloud responsiveness for the application is developed through the following two ways: 1) Actual data—the request analysis unit captures ongoing data about the responses from clouds to user requests and develops a histogram (or the like) to depict overall cloud responsiveness (which may be provided to the customer), or 2) Analytics—based upon the activities carried out in the actual data, the system develops an overall picture of cloud performance (by day, by time, by geographic request, etc.) using a systematic “pinging” of the cloud application over a period of time. Thus, the aggregation of these “pings” to the application may then be utilized to further shape the data picture of the overall responsiveness of a specific cloud.
  • aspects of the invention also provide customers with the ability to specify business rules that determine when and why a request should be routed to a specific cloud. These business rules may be dynamically adjusted within parameters (rather than, for example, hard and fast thresholds) based upon analysis provided by the data gathered from actual and analytics data.
  • customer cloud computing network routing preferences may be received.
  • the preferences may include cost, performance, applications provided, network conditions, outages, business relationships, peering relationships, proximity, etc.
  • the customer may prefer to optimize cloud usage to be most cost efficient as possible.
  • the least expensive cloud computing network may be selected for routing requests for this particular customer, even at the expense of performance.
  • cost may be secondary to providing the highest level of performance.
  • the determination of which of the cloud computing networks to route requests may be based in whole or in part on the customer's preferences.
  • cloud computing networks may be underutilized and so requests may be routed to the underutilized cloud computing networks in order to balance the load among the various clouds.
  • certain clouds may provide services and applications and, as such, requests may be directed to the clouds which align with the requested service or application.
  • the customers may also be able to provide weighting and/or ranking for each of the preferences (process block 522 ).
  • the customer may have preferences set for cost, performance, and quality of service, and each of these categories may have a weight associated with it.
  • performance may be, for example, weighted at a first value while cost may be weighted at a second (lower) value. Accordingly, when analyzing which cloud to route requests, the weights of each preferences may guide and direct the decision making process.
  • a cloud computing network routing table (or similar construction) may be generated (process block 524 ). Accordingly, in one embodiment, the routing table may be used to provide real-time routing changes and decisions for directing requests to the most optimal and favorable cloud computing networks.
  • the routing table may be dynamically updated as preferences and weighting change and in response to changes in the network conditions and performance.
  • the cloud and network conditions may be determined based in part on status requests (pinging, multicast, and the like) sent to each of the cloud computing networks (process block 526 ). As such, the status for each cloud computing network can be updated based on the response received form the status requests (process block 528 ).
  • process block 530 in addition to the performance updates, historical data for each cloud may be tracked, analyzed, and stored. Additionally, performance data may also be stored. Such information may be used in conjunction with the real-time status update information, preferences, and rankings to dynamically update the cloud computing network routing table (process block 532 ). Therefore, intelligent decision-making with regard to where to route each individual cloud-based application or service request can be realized.
  • a request for cloud-based applications or service may be received at the DNS level.
  • the request may be routed to an edge server or the like (process block 542 ).
  • the customer rules/policies may be applied to the incoming request (process block 544 ).
  • requests can be routed to a preferential cloud computing network. Accordingly, once an optimal cloud computing network is selected, the request is then routed to that cloud computing network (process block 546 ).
  • a response is received from the cloud computing network.
  • response time and other metrics may be collected and recorded for assisting in making future routing decisions. Then, the response may be routed back to the edge server and on to the requesting customer (process block 550 ).
  • FIG. 5D illustrates a method 505 in accordance with embodiments of the present invention.
  • a cloud-based application request may be received at the edge server.
  • the request may be analyzed with respect to the cloud computing network routing table and the policies associated with the customer originating the request (process block 562 ). Accordingly, based on the policies and cloud routing table, a cloud is determined to route the request (process block 564 ).
  • performance update and reports may be received regarding the request as well as the status of each cloud.
  • this information may be used to provide a centralized (or single) view for status and performance of each of the cloud computing networks. Accordingly, such a user interface may provide a dynamic view of each cloud's performance, status, applications, and service providers, etc. Accordingly, an administrator may be able to utilize such information to make routing decisions in real-time in order to provide the most optimal cloud-based application and service experience.
  • the routing table may also be updated (process block 568 ). As such, as subsequent requests are received by the edge server, these requests may be routed/re-routed to various clouds to reflect the changes to performance, cost, etc. of the clouds (process block 570 ). For example, a cloud based in India may have a higher latency for customers in New York than a cloud based in Atlanta. However, because of congestion (or other link conditions) the Indian cloud may be able to out-perform the Atlanta cloud despite the latency issues. As such, the routing table would be changed to reflect the change, and requests would be routed accordingly.
  • the routing table would be updated accordingly, and requests would then be routed back to the Atlanta-based cloud from the Indian-based cloud, and so forth.
  • System 600 may include a user device (or system) 605 in communication with a CDN 610 or PoP 410 .
  • the user device 605 is configured to direct cloud-based application requests to the edge server 415 within the CDN 610 or PoP 410 .
  • the edge server 415 may then direct the requests to a request analysis unit 615 .
  • the request analysis unit 615 may be configured to determine the application or service associated with the request, the customer associated with the request, etc.
  • the cloud control unit 620 may access a business rule/policy database 625 to determine the business rules and/or policies associated with the originating customer or destination application associated with the request. As discussed above, the customer preferences may be used to determine which cloud to route various requests. For example, a request from customer X for application Y may be routed differently for requests for application Z, and so forth. Furthermore, the cloud control unit 620 may also access network performance conditions for each of the cloud computing networks 630 a - 630 n . Thus, the cloud control unit 620 may utilize any combination of the request characteristics, the link performance conditions, customer preferences, business rules, policies, etc.
  • the cloud control unit 620 is able to change its routing determinations, thus providing a dynamic routing of requests to cloud computing networks 630 a - 630 n.
  • FIG. 7 a block diagram of an embodiment of a content distribution system 700 is shown in which a content originator 710 offloads delivery of content objects to a content delivery network (CDN).
  • the content originator 710 produces and/or distributes the content objects and may include one or more publishers 706 and content sites 708 .
  • the CDN delivers the content objects over the Internet 704 to end users 722 via corresponding end user devices 720 .
  • the CDN may include an origin server 712 , a policy server 716 , and various points of presence (PoPs) 718 .
  • PoPs 718 can be deployed throughout content distribution system 700 and may serve content objects to end user devices 720 in a particular geographic area and/or in connection with a particular service provider.
  • a PoP 718 may be designated to serve content objects over Internet 704 to end users 722 in a particular city, on a particular access network, etc. to promote efficient delivery and a good user experience.
  • the various CDN elements may be connected by a private network such as WAN 714 and/or a public network such as Internet 704 .
  • An end user 722 may browse for content objects at a content site 708 with its respective end user device 720 .
  • a content object can be any computer-accessible content and may include audio data, video data, images, etc. in any number of computer-accessible formats.
  • the terms content and content object may be used interchangeably wherever they appear.
  • End user devices 720 can include personal computers, media players, handheld computers, Internet appliances, smart phones, personal digital assistants, streaming radios, or any other device that receives and processes content objects.
  • the content site 708 could be a web page from which content is accessible via a web browser.
  • Links to content at the content site 708 may point to locations in the content delivery network.
  • the request may be assigned to a PoP 718 which, in turn, can deliver the requested content object to the end user device 720 . If the content object is not available at the assigned PoP location, the request may be propagated toward the core of the CDN and may ultimately be fulfilled from origin server 712 . Content may be cached at various points between the core CDN and edge locations to improve efficiency.
  • Distribution of content objects often represents an important source of revenue for publishers 706 .
  • content sites 708 may generate advertising revenue based on the number of times that a content object is viewed, clicked, or downloaded by end users 722 .
  • publishers 706 may seek to reach as many end users 722 with their content as possible while providing a good overall user experience.
  • end user devices 720 can vary widely in their respective capabilities and the manner in which they interact with content objects. Different end user devices 720 may support different collections of multimedia formats and different delivery schemes. For example, beginning with OS version 3.0, the iPhoneTM from Apple, Inc. supports M3U8 playlists and MPEG-2 segmented video with iPhoneTM HTTP Streaming (IHS) delivery, entirely over HTTP (Hypertext Transfer Protocol). On the other hand, the Blackberry StormTM from Research in Motion, Ltd. supports playback of multimedia content in Third Generation Partnership Project (3GPP) format, over RTSP (Real-Time Streaming Protocol).
  • 3GPP Third Generation Partnership Project
  • the manner in which delivery of a content object is initiated may vary from device to device. For example, some end user devices 720 may need help orchestrating a browser-to-player (B2P) handoff for certain types of content objects. Moreover, even when media formats and delivery methods are equally supported, the manner in which a content object is delivered may depend on the type of connection to Internet 704 available to the end user device 720 at a particular place and time. Thus, for example, the playback capabilities of the Blackberry StormTM may differ depending upon whether it is connected to the Internet 704 via a WIFI connection in a cybercafé, or via a cellular network in a remote location.
  • policy server 716 is coupled to content site 708 via Internet 704 and receives a notification when new content objects are available from publishers 706 .
  • a publisher 706 may upload its content to an origin server 712 and policy server 716 may receive notifications via WAN 714 when a new content object becomes available.
  • policy server 716 may be located within PoPs 718 , origin server 712 , or other parts of the content delivery network. Also, it will be recognized that the various operations of policy server 716 may be carried out by multiple individual servers such as decisioning servers, merge servers, assembly servers, etc.
  • policy server 716 determines how it should be made available to end users. This may involve generating a number of different versions of the content object optimized for use with different end user devices 720 , having different capabilities, and potentially used in different network environments.
  • the different versions of the content object may correspond to different production or encoding profiles maintained at policy server 716 .
  • the production profiles may be based upon a publisher's requirements for the distribution of its content objects. For example, a publisher may prefer to distribute its content in a specific media format or formats, to exploit device-specific capabilities (such as IHS streaming for iPhones), to optimize separately for high bitrate and low bitrate environments, to target specific operating systems and/or platforms such as WindowsTM or Mac OS, etc.
  • Policy server 716 may associate the different versions of a content object with a single network identifier such as a uniform resource locator (URL).
  • the single network identifier can then be returned to the publisher 706 which created the content.
  • the publisher 706 can add the network identifier to one or more content sites 708 which are accessible to end users 722 .
  • policy server 716 can determine a preferred version of the content object for the end user device 720 and can orchestrate its delivery to the requesting end user.
  • the preferred version and delivery method can be customized for hardware and software capabilities of the end user device 720 , bandwidth and connection quality, viewing habits, user preferences, or any combination of factors.
  • the preferred version may also include a selection of advertisements which are matched to information about the end user device and/or the end user.
  • policy server 716 provides publishers 706 with a one-to-many approach to optimized content delivery. Specifically, a single network identifier can point to multiple versions of a given content object from which policy server 716 selects a preferred version for use with a particular end user device. Policy server 716 thus relieves publishers 706 of the burden of staying up-to-date with technology. When a new platform emerges or device capabilities change, appropriate versions of the content object can be made available to end users 722 through an existing network identifier without further effort from the publisher 706 . Policy server 716 determines the preferred version of a content object in a manner that is transparent to the end user and thus avoids complicated configuration, specialized software, or manual selection. The end user experience is further improved by selecting a delivery method and sending the preferred version of the content object from a PoP 718 location with a fast response time for the user's location, network access, etc.
  • FIG. 8 provides a schematic illustration of one embodiment of a computer system 800 that can perform the methods of the invention, as described herein. It should be noted that FIG. 8 is meant only to provide a generalized illustration of various components, any or all of which may be utilized as appropriate. FIG. 8 , therefore, broadly illustrates how individual system elements may be implemented in a relatively separated or relatively more integrated manner.
  • the computer system 800 is shown comprising hardware elements that can be electrically coupled via a bus 805 (or may otherwise be in communication, as appropriate).
  • the hardware elements can include one or more processors 810 , including without limitation, one or more general purpose processors and/or one or more special purpose processors (such as digital signal processing chips, graphics acceleration chips, and/or the like); one or more input devices 815 , which can include without limitation a mouse, a keyboard and/or the like; and one or more output devices 820 , which can include without limitation a display device, a printer and/or the like.
  • the computer system 800 may further include (and/or be in communication with) one or more storage devices 825 , which can comprise, without limitation, local and/or network accessible storage and/or can include, without limitation, a disk drive, a drive array, an optical storage device, a solid state storage device such as a random access memory (“RAM”) and/or a read-only memory (“ROM”), which can be programmable, flash updateable and/or the like.
  • storage devices 825 can comprise, without limitation, local and/or network accessible storage and/or can include, without limitation, a disk drive, a drive array, an optical storage device, a solid state storage device such as a random access memory (“RAM”) and/or a read-only memory (“ROM”), which can be programmable, flash updateable and/or the like.
  • RAM random access memory
  • ROM read-only memory
  • the computer system 800 might also include a communications subsystem 830 , which can include without limitation a modem, a network card (wireless or wired), an infrared communication device, a wireless communication device and/or chipset (such as a BluetoothTM device, an 802.11 device, a WiFi device, a WiMax device, cellular communication facilities, etc.), and/or the like.
  • the communications subsystem 830 may permit data to be exchanged with a network (such as the network described below, to name one example), and/or any other devices described herein.
  • the computer system 800 will further comprise a working memory 835 , which can include a RAM or ROM device, as described above.
  • the computer system 800 also can comprise software elements, shown as being currently located within the working memory 835 , including an operating system 840 and/or other code, such as one or more application programs 845 , which may comprise computer programs of the invention, and/or may be designed to implement methods of the invention and/or configure systems of the invention, as described herein.
  • an operating system 840 and/or other code such as one or more application programs 845 , which may comprise computer programs of the invention, and/or may be designed to implement methods of the invention and/or configure systems of the invention, as described herein.
  • one or more procedures described with respect to the method(s) discussed above might be implemented as code and/or instructions executable by a computer (and/or a processor within a computer).
  • a set of these instructions and/or codes might be stored on a computer-readable storage medium, such as the storage device(s) 825 described above. In some cases, the storage medium might be incorporated within a computer system, such as the system 800 .
  • the storage medium might be separate from a computer system (i.e., a removable medium, such as a compact disc, etc.), and is provided in an installation package, such that the storage medium can be used to program a general purpose computer with the instructions/code stored thereon.
  • These instructions might take the form of executable code, which is executable by the computer system 800 and/or might take the form of source and/or installable code, which, upon compilation and/or installation on the computer system 800 (e.g., using any of a variety of generally available compilers, installation programs, compression/decompression utilities, etc.), then takes the form of executable code.
  • the computer or machine-readable medium may be non-transitory.
  • the invention employs a computer system (such as the computer system 800 ) to perform methods of the invention.
  • a computer system such as the computer system 800
  • some or all of the procedures of such methods are performed by the computer system 800 in response to processor 810 executing one or more sequences of one or more instructions (which might be incorporated into the operating system 840 and/or other code, such as an application program 845 ) contained in the working memory 835 .
  • Such instructions may be read into the working memory 835 from another machine-readable medium, such as one or more of the storage device(s) 825 .
  • execution of the sequences of instructions contained in the working memory 835 might cause the processor(s) 810 to perform one or more procedures of the methods described herein.
  • machine-readable medium and “computer readable medium,” as used herein, refer to any medium that participates in providing data that causes a machine to operate in a specific fashion.
  • various machine-readable media might be involved in providing instructions/code to processor(s) 810 for execution and/or might be used to store and/or carry such instructions/code (e.g., as signals).
  • a computer-readable medium is a physical and/or tangible storage medium. Such a medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media.
  • Non-volatile media includes, for example, optical or magnetic disks, such as the storage device(s) 825 .
  • Volatile media includes, without limitation, dynamic memory, such as the working memory 835 .
  • Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise the bus 805 , as well as the various components of the communications subsystem 830 (and/or the media by which the communications subsystem 830 provides communication with other devices).
  • transmission media can also take the form of waves (including without limitation radio, acoustic and/or light waves, such as those generated during radio wave and infrared data communications).
  • Common forms of physical and/or tangible computer-readable media include, for example, a floppy disk, a flexible disk, a hard disk, magnetic tape, or any other magnetic medium, a CD-ROM, any other optical medium, punchcards, papertape, any other physical medium with patterns of holes, a RAM, a PROM, an EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read instructions and/or code.
  • Various forms of machine-readable media may be involved in carrying one or more sequences of one or more instructions to the processor(s) 810 for execution.
  • the instructions may initially be carried on a magnetic disk and/or optical disc of a remote computer.
  • a remote computer might load the instructions into its dynamic memory and send the instructions as signals over a transmission medium to be received and/or executed by the computer system 800 .
  • These signals which might be in the form of electromagnetic signals, acoustic signals, optical signals and/or the like, are all examples of carrier waves on which instructions can be encoded, in accordance with various embodiments of the invention.
  • the communications subsystem 830 (and/or components thereof) generally will receive the signals, and the bus 805 then might carry the signals (and/or the data, instructions, etc., carried by the signals) to the working memory 835 , from which the processor(s) 810 retrieves and executes the instructions.
  • the instructions received by the working memory 835 may optionally be stored on a storage device 825 either before or after execution by the processor(s) 810 .
  • the present invention may be embodied in other specific forms.
  • the partial object cache as previously described, can be associated with a plurality of versions of programming structures.
  • Those skilled in the art will recognize, or be able to ascertain using no more than routine experimentation, many equivalents to the specific embodiments of the invention described herein. Such equivalents are intended to be encompassed by the following claims.

Abstract

Aspects of the present invention include a method of dynamically routing requests within multiple cloud computing networks. The method includes receiving a request for an application from a user device, forwarding the request to an edge server within a content delivery network (CDN), and analyzing the request to gather metrics about responsiveness provided by the multiple cloud computing networks running the application. The method further includes analyzing historical data for the multiple cloud computing networks regarding performance of the application, based on the performance metrics and the historical data, determining an optimal cloud computing network within the multiple cloud computing networks to route the request, routing the request to the optimal cloud computing network, and returning the response from the optimal cloud computing network to the user device.

Description

    CROSS-REFERENCES
  • This application is a continuation of U.S. patent application Ser. No. 13/245,582, filed Sep. 26, 2011, entitled “DYNAMIC ROUTE REQUESTS FOR MULTIPLE CLOUDS,” which is hereby incorporated by reference in its entirety for all purposes.
  • RELATED APPLICATION
  • This application is related to U.S. patent application Ser. No. 13/245,601, filed Sep. 26, 2011, entitled “EDGE-BASED RESOURCE SPIN-UP FOR CLOUD COMPUTING”, which is incorporated by reference in its entirety for any and all purposes.
  • BACKGROUND
  • Presently, compute resources (i.e., applications, etc.) within a cloud provider's network are spun-up in a cluster (e.g., servers which are aggregated in a centralized location, a datacenter, etc.). All requests are load-balanced back to that cluster. Unfortunately, such an implementation does not necessarily provide the best performance or experience for end users who may, for example, be located far away from the centralized cluster.
  • This problem is further compounded by the fact that applications provided within the “cloud” are becoming more robust and require additional resources and computing power as well as faster response times. Accordingly, the computations being preformed over the web are becoming increasingly more intensive. As such, with the centralized cluster approach, many of these computations are being routed away from the user which adds to or even causes delays and an unacceptable user experience.
  • One example of a current implementation is illustrated by method 100 of FIG. 1. A data center 105 includes a compute platform 110 which is in communication with devices which produce user requests 115. As such, user requests 115 are received by the data center 105, which includes the cloud resources. As requests increase, software and services within the data center 105 are spun-up by additional cloud resources using the compute platform 110. The distance between the compute platform 110 and the user requests 115 may be great, and therefore, responsiveness and user experience are diminished greatly.
  • Furthermore, in the current cloud-service environments, customers must deploy their applications to a single cloud, and utilize the elasticity of the cloud to determine additional resources and spin those up accordingly within the cloud environment. Unfortunately, if the cloud provider is experiencing difficulties (either regionally or globally), the customer has no way to re-route requests to another cloud, and thus performance is dramatically impacted. Thus, for at least these reasons, improvements in the art are needed.
  • BRIEF SUMMARY
  • In one embodiment, aspects of the present invention distribute new resources closer to end-users which are requesting the resource. As such, additional virtualized instances (as part of a cloud provisioning) are spun-up within servers that are physically near to the network equipment (i.e., web servers, switches, routers, load balancers) which are receiving the requests. Accordingly, by moving computational resources closer to the requesting users in cloud computing environments, the user experience is significantly enhanced.
  • Further areas of applicability of the present disclosure will become apparent from the detailed description provided hereinafter. It should be understood that the detailed description and specific examples, while indicating various embodiments, are intended for purposes of illustration only and are not intended to necessarily limit the scope of the disclosure.
  • Further aspects of the present invention include dynamically routing requests for applications to one of multiple cloud computing environments. Alternatively, the method may dynamically route an application request to an application that is hosted in multiple clouds (deployed within a management application) based upon a specified criteria. In one embodiment, the routing of requests for the application to a specific cloud in which the application is deployed may be based upon a criteria(s) that the application owner specifies. This may provide the application owner an ability to positively affect quality of service (QoS) for application delivery, ensure uninterrupted access to the application in the event of failure by one or more clouds, and provide more efficient application performance.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 shows a system for implementing cloud computing.
  • FIG. 2 shows a system for implementing edge-based resource spin-up for cloud computing.
  • FIG. 3 shows a method of implementing edge-based resource spin-up for cloud computing.
  • FIGS. 4A-4D show systems for implementing edge-based resource spin-up for cloud computing.
  • FIGS. 5A-5D show methods of implementing dynamic route requests for multiple clouds.
  • FIGS. 6A and 6B show systems for implementing dynamic route requests for multiple clouds.
  • FIG. 7 shows an embodiment of a content distribution system.
  • FIG. 8 shows an embodiment of a computer system.
  • In the figures, similar components and/or features may have the same reference label. In some cases, components of the same type are identified by following a first reference label with a dash and a second reference label that further distinguishes among the similar components. If only the first reference label is used, the description is applicable to any of the similar components designated by the first reference label.
  • DETAILED DESCRIPTION OF EMBODIMENTS
  • The ensuing description provides preferred exemplary embodiment(s) only, and such preferred exemplary embodiments are not intended to limit the scope or applicability of the present invention. Rather, the ensuing description will enable those who are skilled in the art to implement such preferred exemplary embodiment(s). Persons of skill in the art will recognize that various changes may be made in the function and arrangement of elements without departing from the spirit and scope of the invention as set forth in the appended claims.
  • FIG. 2 shows a system for implementing edge-based resource spin-up for cloud computing, in accordance with one embodiment of the present invention. In one embodiment, edge-based resource spin-up includes carrying out computational activities within a cloud computing environment closer to the end user. As such, an increase in responsiveness as well as a more efficient use of resources is realized. System 200 includes a data center 205 a and 205 b. In one embodiment, data centers 205 may be a facility used to house computer systems and associated components, such as telecommunications, networking systems, storage systems, etc. Furthermore, the data centers 205 may also be designated as points of presence (PoPs).
  • In one embodiment, the data centers 205 a and 205 b may include edge servers 210 a and 210 b, respectively. Further, edge servers 210 a and 210 b may include compute platforms 215 a and 215 b, respectively. It should be noted that one skilled in the art would conclude that any number of data centers, edge serves, and/or compute platforms may be included, and only two of each are shown for ease of explanation and illustration.
  • In a further embodiment, system 200 may include a load balancer 220 in communication with both data centers 205 a and 205 b, as well as user devices issuing user requests 225 a and 225 b. In a cloud computing environment such as the one depicted in system 200, many user requests may be received, and proper allocation and division of cloud resources should be allocated to handle the requests. Furthermore, many of the requests are time sensitive and latency sensitive (i.e., UI intensive applications, computation intensive applications, etc.), so ensuring fast response times to requests can be important. As such, in the confirmation of system 200, the load balancer 220 is configured to determine the “fastest” responding edge server/compute platform to direct the request. In one embodiment, fastest response time means the edge server closest physically to the requesting user device. Alternatively, fastest may mean the edge server with the lowest latency relative to the requesting device. In some instances, the closest and the lowest latency edge server may be the same server, but not always. For example, if the physically closest edge server is experiencing a heavy load of traffic and requests, the response time and/or network latency of the server may outweigh the physically close proximity to the requesting device.
  • In other words, the load balancer 220 is configured to ensure that the needed resources to respond to the user requests 225 a and 225 b are routed to the edge servers 210 a and 201 b which will provide the fastest response time for the request, which in many cases will be the edge server which is in the closest proximity to the requesting user device.
  • In one example, two groups of users make requests from two different geographical locations. The load balancer 220 then receives the requests and, based on the location of the request, distributes the request to the data center 205 a or 205 b closest to the user (alternatively, the request may be routed to the data center which will provide the faster response time). Once the request is routed, it is received by a “localized cloud instance” which is a de-centralized cloud computing environment with resources spun-up as physically close to the requesting device as possible. In one embodiment, such localized resources may be synchronized around the network to ensure that requests come to one localized cluster are treated in the same manner as other requests. Then, based on the request load that is delivered to that “localized cloud instance,” resources are spun-up in that locality based upon demand (i.e., subsequent user requests).
  • FIG. 3 shows a method 300 of implementing edge-based resource spin-up for cloud computing, in accordance with one embodiment of the present invention. At process block 305, a request for data or a service may be received at an edge server from a user device. In one embodiment, the closest edge server to the requesting device may be determined by using an enhanced anycast methodology. Accordingly, the edge server which provides the fastest response time relative to the requesting user device is selected.
  • In one embodiment, the request may be for an application, such as an enterprise application, a media application, etc. Alternatively, the request may be for data, such a video file, a music file, a document, etc. Each request may have associated information sent with the request which identifies the application and/or data used to service the request. The identification information may be embedded or attached to the request.
  • Furthermore, at process block 310, the edge server may extract the identification information. Then, based on the information, the edge server can identify the application/service used to process the request (process block 315). For example, the identification information may specifically identify the application by name or some other identifier, or alternatively the information may include an application type, etc.
  • Further, the selected edge server may be in communication with one or more compute platforms, which may be co-located or remotely-located with the edge server. Additionally, each of the compute platforms may have one or more containers running which provide a virtual construct for allocating resources. In one embodiment, these containers may be a type of virtualized resource which is different from a virtualized instanced, such as elastic computing cloud (EC2). The containers are then configured to execute and maintain applications needed to service the user requests. Hence, at decision block 320, a determination is made whether a container maintained by a compute platform in communication with the selected edge server is running (or capable of running) the application necessary for servicing the user request. In one embodiment, all of the edge-based compute platforms may include the “DNA” for running an application (e.g., an XML dataset that specifies instructions for each application to be run in a container), and the determination for being able to run the application based on the current levels of utilization. As such, the allocation of the compute platform becomes a predictive determination. In one embodiment, a compute platform is capable of running the application if the compute platform has sufficient unused resources, if the necessary application is loaded on the compute platform, etc.
  • If the application is not running on any of the containers within the compute platforms, then at process block 330, an available compute platform (or on other words, a compute platform which has available resources) is identified. Accordingly, it may not matter if the application is not currently running, as the application can be spun up; availability can be based on either a currently running application or the necessary capacity to support the application running, which could then be translated to actually spinning up a container, on demand, to support the requests. Then, one or more containers are spun-up by the identified compute platform to run the identified application or service (process block 335).
  • Alternatively, if there is a container identified as running the application, then a determination is made whether the container has sufficient resources to handle the increased load of the new request (decision block 325). If the container does not have sufficient resources to handle the increased load, then at process block 335, a container (or containers) may be spun-up to run the identified application. Alternatively, if the container has sufficient resources to handle the increased load, then at process block 340, the request is routed to the compute platform with the container already running the identified application. As such, the load is effectively balanced to the compute platform and container with available resources from the edge server with the closest physical proximity to the requesting device; thus, providing the most efficient user experience.
  • One example of an implementation of method 300 may be performed for the MediaTag™ application. In one embodiment, a user may click on a link/file that the user desires to purchase. The file includes an associated cookie which is used to point the request to the MediaTag application. The application makes a request of the cookie which has been stored on the user's machine by the website providing the music download; MediaTag then takes the cookie, explodes it, and carries out computational activity against the results. The edge server then upon receiving the request interprets the tag and identifies a compute platform which is capable of spinning-up resources for the MediaTag application. Alternatively, the determination may be based solely on geography—the closest POP with resources; there is sometimes a tradeoff between locality and capacity—as a system may chose to actually go to a more distant compute resource to carry out my request because the latency of serving the response is actually less than the latency caused in the local edge by the lack of capacity.
  • Then, the compute platform spins-up a container running the MediaTag application. The MediaTag application then creates a unique file based on the request (the file may include identification information, such as the username of the requester, the origination location, etc.). Again, alternatively, the choice may be based on both proximity and the current utilization level of that current proximal location; there is a tradeoff. Then, a response to the request is sent to the user (process block 345).
  • This entire process is implemented using the edge-based cloud computing solution of the present invention. At each step of the execution of the MediTag application, resources and servers are chosen based on their physical proximity to the requesting user device, thus increasing the efficiency and executing time of the MediaTag application. Other applications may be implemented in the same or similar way utilizing method 300.
  • Referring next to FIG. 4A, a system 400 for implementing edge-based resource spin-up for cloud computing is shown, in accordance with embodiments of the present invention. The system 400 may include a user device 405. In one embodiment, the user device 405 may be a mobile device, a cellular device, a SmartPhone, a mobile computing platform, a user terminal, etc. The user device 405 may be configured to send requests and access data, services, and applications. Further, the user device 405 may be in communication with a cloud computing network as shown in system 400.
  • In one embodiment, the user device 405 may be in communication with a point of presence (PoP) 410. PoP 410 may be configured as an access point to the Internet 430, a physical location that houses servers, routers, ATM switches, digital/analog call aggregators, etc. Further, PoP 410 may be either part of the facilities of a telecommunications provider that the Internet service provider (ISP) rents or a location separate from the telecommunications provider. Generally, PoPs are also located at Internet exchange points and collocation centers.
  • An edge server 415 may be located within PoP 410. The edge server 415 may be operated by a cloud computing provider, or the like. The edge server 415 may represent one of the cloud computing provider's closest connection points to the Internet 430. As such, the edge server 415 is uniquely qualified to provide the fastest and most efficient service to the user device 405, particularly with regard to spinning-up resources for use in a cloud computing environment. In many implementations there may be hundreds of edge servers ready to receive and process user requests.
  • Accordingly, as requests are generated from user device 405 and routed to edge server 415 via PoP 410, the edge server 415, in communication with a compute platform 420, may direct the compute platform 420 to handle the requests. For example, edge server 415 may direct compute platform 420 to spin-up a container 420 a to handle the incoming request. In one embodiment, spinning-up container 420 a includes allocating and assigning the necessary resources and applications to handle the request. For example, if the request is for application 420 b, then an instance of application 420 b (e.g., MediaTag, video player, gaming application, etc.) will be initiated in container 420 a. Likewise, if it is determined that application 420 b requires X amount of processing power, memory allocation, hard drive space, etc., then these needed resources will also be allocated in container 420 a. As such, container 420 a is spun-up such that container 420 a is equipped to handle the request from user device 405.
  • In a further embodiment, the PoP 410 may be selected as being the “closest” PoP to the user device 405. In one embodiment, closest means the closest in physical proximity to the user device 405, which in turn provides the fastest response time to requests, thus enhancing the cloud computing experience. For example, in a central server cloud computing configuration, computations are routed away from the user device; however, in an edge-based cloud computing environment as in system 400, the computations are performed as close as possible to the user device 405. As such, rendering of cloud applications can be done without diminishing the experience (i.e., the application can be rendered as though the application is run “locally” on the user's device, or in a local area network or the like). This may be particularly important with regard to mobile users in that—as the user moves, so too may the “closest” available edge server “move”.
  • Furthermore, such a cloud computing configuration as in system 400 can provide a more efficient use of resources. For example, instead of implementing a large expensive resource intensive centralized cloud computing platform, the de-centralized model allows for resources to be spun-up in order to handle specific user requests. Furthermore, system 400 (additionally, systems 401-403) provides a scalable solution, such that as additional requests are received by compute platform 420, the platform can direct additional containers to be spun-up to dynamically handle the increased load. In other words, the containers allow for dynamically creating instances of the systems' operating system which has been “tuned” for a specific purpose. Therefore, the necessary resources are provided in response to the received requests, and the resources are provided at the closest location to the requesting device.
  • Alternatively, other factors can be considered when determining the appropriate edge server to route the user device 405's request. For example, “effective latency” or “effective distance” may be considered. In one embodiment, effective latency or distance may be defined as the accrual speed of a response to a user device request. For example, edge server 415 may be physically closer to user device 405, but edge server 415 may be heavily congested. Thus, edge server X (which is physically further from user device 405) may have a better effective distance or latency, and may ultimately be selected to process the request, despite its distance from user device 405. Similarly, outages and other issues may be considered when determining which edge serve will actually (or effectively) provide the fastest response time and ultimately the most desirable user experience. Such a cloud computing environment uniquely provides what a centralized cloud computing environment is unable to provide: scalability, efficiency, and faster response times.
  • Additionally, configuration files (or the like) may be used to determine the edge server and/or the compute platform to handle certain user requests. For example, a certain edge server may be closest to the user device 405, but because the edge server does not have access to the requested application or other resource, a further edge server may need to be used. Therefore, the configuration file may provide such designation and mappings, such that requests are routed to edge servers and compute platforms that are actually equipped to handle the request. Additionally, the configuration files may also provide an accounting of the physical as well as the virtual resources available to each edge server, which can assist in routing decisions by not overloading edge servers above their resource capability capacity, etc.
  • An additional advantage of edge-based cloud computing may be that requests are able to be routed using URLs. URLs uniquely allow for information to be appended to the URL which can provide the necessary information to the edge server for more efficient routing and resource allocation. In addition, URL-based routing enables a variety of systems (e.g., anything capable of dealing with HTTP) the ability to forward the request along. This makes for a very flexible application architecture and a distributed computing environment in that individual application components, all making URL (i.e., HTTP-based) requests, can be completed from different compute resources, not necessarily all the same resource—the resources used for these requests can be fanned out.
  • FIG. 4B shows an alternative embodiment of an edge-based cloud computing configuration. In one embodiment, system 401 may have the compute platform 420 co-located with edge server 415 at PoP 410. Furthermore, multiple compute platforms 420 a, 420 b to 420 n may be provided. As such, the additional compute platforms may provide the edge server with access to additional resources in order to handle additional requests. Furthermore, since the compute platforms 420 a to 420 n are co-located with the edge server 415 at the PoP 410, latency can be significantly reduced. Accordingly, computations for user requests are pushed even closer to the user device 405 originating the request. Accordingly, the edge is “super-charged” with readily available compute resources to meet specific types of requests. In addition, new intelligence is added to the edge to better route request traffic (i.e., URLS) to other edge resources, different POPs, etc., depending upon performance and/or availability factors that reflect customer preferences.
  • In a further embodiment, each of the compute platforms 420 a through 420 n are capable of spinning-up multiple containers 422 a through 422 n. Thus, each of the containers 422 a through 422 n can also provide instances of applications 423 a through 423 n. Accordingly, each compute platform 420 can expand or shrink to effectively and efficiently accommodate an increase or decrease in user requests. The dynamic nature of resource allocation coupled with the relative closeness in proximity to the user device 405 provide for an optimal user experience in a cloud computing environment.
  • FIG. 4C shows another embodiment of an edge-based cloud computing environment. System 402 shows an alternative configuration in which the compute platforms 420 a through 420 n are located at a compute server 435, which may be remotely located from the edge server 415 and PoP 410. For example, since server space may be expensive within the PoP 410, it may be economical to place the compute platforms at a location in close proximity to the PoP 410 which is less expensive. As such, the diminished response time (which is minimal) may be outweighed by the reduced cost of non-PoP space (in particular, if the same edge resources are being utilized for other purposes (i.e., streaming)). Additionally, at the reduced cost of implementing compute server 435, the computational power at the edge server 415 can be significantly increased, and as such is able to handle an increased amount of requests; thus increasing the scalability and efficiency of the cloud computing environment.
  • FIG. 4D shows a further embodiment of an edge-based cloud computing environment. In this embodiment, shown as system 403, instead of the edge server 415 being located at the PoP 410, a router (or the like) 412 is located at the PoP 410. The router 412 may be configured to route requests from the PoP 410 to the edge server 415. Then, the edge server 415 can direct the requests to the appropriate compute platform 420 within the compute server 435. In one embodiment, the compute server 435 may be co-located with edge server 415, or alternatively the compute server 435 may be remotely-located from the edge server 415.
  • Further aspects of this invention include dynamically routing requests for applications to one of multiple cloud computing environments. Alternatively, the method may dynamically route an application request to an application that is hosted in multiple clouds (deployed within a management application) based upon a specified criteria. In one embodiment, the routing of requests for the application to a specific cloud in which the application is deployed may be based upon a criteria(s) that the application owner specifies. This may provide the application owner an ability to positively affect quality of service (QoS) for application delivery, ensure uninterrupted access to the application in the event of failure by one or more clouds, and provide more efficient application performance.
  • As Web applications are becoming increasingly more complex and resource intensive, (in some cases requiring multiple coded elements, multiple data stores, external and internal system integration, etc.), even a few milliseconds of latency between the user requesting an element of the application (via, for example, a URL) and the response to the user, can cause a user to utilize a competitors' offering and thereby materially impact business for the Web application owner. In addition, there are a variety of factors that may impact (negatively or positively) the ability for a specific cloud to respond favorably (based on, for example, business rules) to a user request. This can include proximity of the cloud assets to the end user, peering relationships between the cloud service provider and ISPs on which users are accessing the cloud resources, etc. Thus, aspects of the present invention create a picture of that responsiveness to eventually enable a dynamic adjustment of business rules based on analysis of the data provided as part of the overall system.
  • One embodiment of the present invention may provide that a customer signs up for a “multi-cloud application deployment” which entails provisioning for a content delivery network (CDN) account (i.e., providing HTTP service or the like), as well as enabling the customer access to a portal (or other web-based UI) that allows the customer to specify the cloud-based locations of their applications, the URLs to those applications, the business rules the customer wants the applications to follow when shuttling requests to different clouds, etc. Each cloud that the customer specifies may require a unique hostname provided by the CDN.
  • Turning now to FIG. 5A, which illustrates that once the multi-cloud application deployment has been properly configured, method 500 may be executed. At process block 502, a request may be received by the end user. This request may then be passed to the edge of the CDN via a CNAME or designated hostname provided by the CDN to the customer that configures their application URLs accordingly (process block 504). At process block 506, the request may then be passed through a request analysis unit (or similar module) to gather metrics on the response provided by the application to which the request is being used. For example, a historical picture of specific cloud responses may be generated and developed.
  • At process block 508, based upon historical analysis (explained below) and business rules, a cloud control unit (or similar module) may then determine the cloud to which to direct the request. At process block 510, the response from the cloud for the request is captured by a request analysis unit (or similar module) and the response is returned to the user (process block 512). As such, a historical picture of cloud responsiveness for the application is developed through the following two ways: 1) Actual data—the request analysis unit captures ongoing data about the responses from clouds to user requests and develops a histogram (or the like) to depict overall cloud responsiveness (which may be provided to the customer), or 2) Analytics—based upon the activities carried out in the actual data, the system develops an overall picture of cloud performance (by day, by time, by geographic request, etc.) using a systematic “pinging” of the cloud application over a period of time. Thus, the aggregation of these “pings” to the application may then be utilized to further shape the data picture of the overall responsiveness of a specific cloud.
  • Furthermore, aspects of the invention also provide customers with the ability to specify business rules that determine when and why a request should be routed to a specific cloud. These business rules may be dynamically adjusted within parameters (rather than, for example, hard and fast thresholds) based upon analysis provided by the data gathered from actual and analytics data.
  • Referring next to FIG. 5B a method 501 is illustrated in accordance with embodiments of the present invention. At process block 520, customer cloud computing network routing preferences may be received. In one embodiment, the preferences may include cost, performance, applications provided, network conditions, outages, business relationships, peering relationships, proximity, etc. For example, the customer may prefer to optimize cloud usage to be most cost efficient as possible. As such, the least expensive cloud computing network may be selected for routing requests for this particular customer, even at the expense of performance. Similarly, if the customer places a high importance on performance, then cost may be secondary to providing the highest level of performance. At any rate, the determination of which of the cloud computing networks to route requests may be based in whole or in part on the customer's preferences.
  • An additional consideration may be driven by providing benefits to the network provider. For example, one or more cloud computing networks may be underutilized and so requests may be routed to the underutilized cloud computing networks in order to balance the load among the various clouds. Furthermore, certain clouds may provide services and applications and, as such, requests may be directed to the clouds which align with the requested service or application.
  • In addition to the preferences received from the customers, the customers may also be able to provide weighting and/or ranking for each of the preferences (process block 522). For example, the customer may have preferences set for cost, performance, and quality of service, and each of these categories may have a weight associated with it. In one embodiment, performance may be, for example, weighted at a first value while cost may be weighted at a second (lower) value. Accordingly, when analyzing which cloud to route requests, the weights of each preferences may guide and direct the decision making process.
  • Utilizing the preference information and weighting information in connection with performance data for each of the clouds, a cloud computing network routing table (or similar construction) may be generated (process block 524). Accordingly, in one embodiment, the routing table may be used to provide real-time routing changes and decisions for directing requests to the most optimal and favorable cloud computing networks. The routing table may be dynamically updated as preferences and weighting change and in response to changes in the network conditions and performance. The cloud and network conditions may be determined based in part on status requests (pinging, multicast, and the like) sent to each of the cloud computing networks (process block 526). As such, the status for each cloud computing network can be updated based on the response received form the status requests (process block 528).
  • At process block 530, in addition to the performance updates, historical data for each cloud may be tracked, analyzed, and stored. Additionally, performance data may also be stored. Such information may be used in conjunction with the real-time status update information, preferences, and rankings to dynamically update the cloud computing network routing table (process block 532). Therefore, intelligent decision-making with regard to where to route each individual cloud-based application or service request can be realized.
  • Referring now to FIG. 5C, a method 503 is illustrated in accordance with embodiments of the present invention. At process block 540, a request for cloud-based applications or service may be received at the DNS level. In response, the request may be routed to an edge server or the like (process block 542). At the edge server, the customer rules/policies may be applied to the incoming request (process block 544). As such, as discussed above, requests can be routed to a preferential cloud computing network. Accordingly, once an optimal cloud computing network is selected, the request is then routed to that cloud computing network (process block 546).
  • At process block 548, a response is received from the cloud computing network. In one embodiment, response time and other metrics may be collected and recorded for assisting in making future routing decisions. Then, the response may be routed back to the edge server and on to the requesting customer (process block 550).
  • FIG. 5D illustrates a method 505 in accordance with embodiments of the present invention. At process block 560, a cloud-based application request may be received at the edge server. The request may be analyzed with respect to the cloud computing network routing table and the policies associated with the customer originating the request (process block 562). Accordingly, based on the policies and cloud routing table, a cloud is determined to route the request (process block 564).
  • At process block 566, performance update and reports may be received regarding the request as well as the status of each cloud. In one embodiment, this information may be used to provide a centralized (or single) view for status and performance of each of the cloud computing networks. Accordingly, such a user interface may provide a dynamic view of each cloud's performance, status, applications, and service providers, etc. Accordingly, an administrator may be able to utilize such information to make routing decisions in real-time in order to provide the most optimal cloud-based application and service experience.
  • Furthermore, based on the updated information for each of the cloud computing networks, the routing table may also be updated (process block 568). As such, as subsequent requests are received by the edge server, these requests may be routed/re-routed to various clouds to reflect the changes to performance, cost, etc. of the clouds (process block 570). For example, a cloud based in India may have a higher latency for customers in New York than a cloud based in Atlanta. However, because of congestion (or other link conditions) the Indian cloud may be able to out-perform the Atlanta cloud despite the latency issues. As such, the routing table would be changed to reflect the change, and requests would be routed accordingly. Similarly, as the Atlanta-based cloud congestion subsides, the routing table would be updated accordingly, and requests would then be routed back to the Atlanta-based cloud from the Indian-based cloud, and so forth. Essentially, the dynamic nature of the routing of requests to multiple cloud computing networks for vides customers with the ability to have requests dynamically routed in the most efficient way possible based on current conditions and preferences.
  • Turning now to FIGS. 6A and 6B, systems 600 and 601 are illustrated in accordance with embodiments of the present invention. System 600 may include a user device (or system) 605 in communication with a CDN 610 or PoP 410. The user device 605 is configured to direct cloud-based application requests to the edge server 415 within the CDN 610 or PoP 410. The edge server 415 may then direct the requests to a request analysis unit 615. In one embodiment, the request analysis unit 615 may be configured to determine the application or service associated with the request, the customer associated with the request, etc.
  • Accordingly, such information about the request may be passed to a cloud control unit 620. The cloud control unit 620 may access a business rule/policy database 625 to determine the business rules and/or policies associated with the originating customer or destination application associated with the request. As discussed above, the customer preferences may be used to determine which cloud to route various requests. For example, a request from customer X for application Y may be routed differently for requests for application Z, and so forth. Furthermore, the cloud control unit 620 may also access network performance conditions for each of the cloud computing networks 630 a-630 n. Thus, the cloud control unit 620 may utilize any combination of the request characteristics, the link performance conditions, customer preferences, business rules, policies, etc. to determine which of cloud computing networks 630 a-630 n to route each request. Furthermore, as conditions and preferences change, the cloud control unit 620 is able to change its routing determinations, thus providing a dynamic routing of requests to cloud computing networks 630 a-630 n.
  • Referring first to FIG. 7, a block diagram of an embodiment of a content distribution system 700 is shown in which a content originator 710 offloads delivery of content objects to a content delivery network (CDN). The content originator 710 produces and/or distributes the content objects and may include one or more publishers 706 and content sites 708. The CDN delivers the content objects over the Internet 704 to end users 722 via corresponding end user devices 720.
  • As shown, the CDN may include an origin server 712, a policy server 716, and various points of presence (PoPs) 718. PoPs 718 can be deployed throughout content distribution system 700 and may serve content objects to end user devices 720 in a particular geographic area and/or in connection with a particular service provider. For example, a PoP 718 may be designated to serve content objects over Internet 704 to end users 722 in a particular city, on a particular access network, etc. to promote efficient delivery and a good user experience. The various CDN elements may be connected by a private network such as WAN 714 and/or a public network such as Internet 704.
  • An end user 722 may browse for content objects at a content site 708 with its respective end user device 720. As used herein, a content object can be any computer-accessible content and may include audio data, video data, images, etc. in any number of computer-accessible formats. The terms content and content object may be used interchangeably wherever they appear. End user devices 720 can include personal computers, media players, handheld computers, Internet appliances, smart phones, personal digital assistants, streaming radios, or any other device that receives and processes content objects. The content site 708 could be a web page from which content is accessible via a web browser.
  • Links to content at the content site 708 may point to locations in the content delivery network. When an end user requests delivery of a particular content object, the request may be assigned to a PoP 718 which, in turn, can deliver the requested content object to the end user device 720. If the content object is not available at the assigned PoP location, the request may be propagated toward the core of the CDN and may ultimately be fulfilled from origin server 712. Content may be cached at various points between the core CDN and edge locations to improve efficiency.
  • Distribution of content objects often represents an important source of revenue for publishers 706. For example, content sites 708 may generate advertising revenue based on the number of times that a content object is viewed, clicked, or downloaded by end users 722. Thus, to maximize their revenue, publishers 706 may seek to reach as many end users 722 with their content as possible while providing a good overall user experience.
  • Unfortunately, end user devices 720 can vary widely in their respective capabilities and the manner in which they interact with content objects. Different end user devices 720 may support different collections of multimedia formats and different delivery schemes. For example, beginning with OS version 3.0, the iPhone™ from Apple, Inc. supports M3U8 playlists and MPEG-2 segmented video with iPhone™ HTTP Streaming (IHS) delivery, entirely over HTTP (Hypertext Transfer Protocol). On the other hand, the Blackberry Storm™ from Research in Motion, Ltd. supports playback of multimedia content in Third Generation Partnership Project (3GPP) format, over RTSP (Real-Time Streaming Protocol).
  • To further complicate matters, the manner in which delivery of a content object is initiated may vary from device to device. For example, some end user devices 720 may need help orchestrating a browser-to-player (B2P) handoff for certain types of content objects. Moreover, even when media formats and delivery methods are equally supported, the manner in which a content object is delivered may depend on the type of connection to Internet 704 available to the end user device 720 at a particular place and time. Thus, for example, the playback capabilities of the Blackberry Storm™ may differ depending upon whether it is connected to the Internet 704 via a WIFI connection in a cybercafé, or via a cellular network in a remote location.
  • In the present embodiment, policy server 716 is coupled to content site 708 via Internet 704 and receives a notification when new content objects are available from publishers 706. Alternatively, a publisher 706 may upload its content to an origin server 712 and policy server 716 may receive notifications via WAN 714 when a new content object becomes available. Although shown separately, policy server 716 may be located within PoPs 718, origin server 712, or other parts of the content delivery network. Also, it will be recognized that the various operations of policy server 716 may be carried out by multiple individual servers such as decisioning servers, merge servers, assembly servers, etc.
  • When a new content object is ready for processing, policy server 716 determines how it should be made available to end users. This may involve generating a number of different versions of the content object optimized for use with different end user devices 720, having different capabilities, and potentially used in different network environments. The different versions of the content object may correspond to different production or encoding profiles maintained at policy server 716. The production profiles, in turn, may be based upon a publisher's requirements for the distribution of its content objects. For example, a publisher may prefer to distribute its content in a specific media format or formats, to exploit device-specific capabilities (such as IHS streaming for iPhones), to optimize separately for high bitrate and low bitrate environments, to target specific operating systems and/or platforms such as Windows™ or Mac OS, etc.
  • Policy server 716 may associate the different versions of a content object with a single network identifier such as a uniform resource locator (URL). The single network identifier can then be returned to the publisher 706 which created the content. The publisher 706 can add the network identifier to one or more content sites 708 which are accessible to end users 722. When a request for the content object is received from an end user device 720, it can be sent to policy server 716 for analysis. Using all available information, policy server 716 can determine a preferred version of the content object for the end user device 720 and can orchestrate its delivery to the requesting end user. The preferred version and delivery method can be customized for hardware and software capabilities of the end user device 720, bandwidth and connection quality, viewing habits, user preferences, or any combination of factors. The preferred version may also include a selection of advertisements which are matched to information about the end user device and/or the end user.
  • As described herein, policy server 716 provides publishers 706 with a one-to-many approach to optimized content delivery. Specifically, a single network identifier can point to multiple versions of a given content object from which policy server 716 selects a preferred version for use with a particular end user device. Policy server 716 thus relieves publishers 706 of the burden of staying up-to-date with technology. When a new platform emerges or device capabilities change, appropriate versions of the content object can be made available to end users 722 through an existing network identifier without further effort from the publisher 706. Policy server 716 determines the preferred version of a content object in a manner that is transparent to the end user and thus avoids complicated configuration, specialized software, or manual selection. The end user experience is further improved by selecting a delivery method and sending the preferred version of the content object from a PoP 718 location with a fast response time for the user's location, network access, etc.
  • FIG. 8 provides a schematic illustration of one embodiment of a computer system 800 that can perform the methods of the invention, as described herein. It should be noted that FIG. 8 is meant only to provide a generalized illustration of various components, any or all of which may be utilized as appropriate. FIG. 8, therefore, broadly illustrates how individual system elements may be implemented in a relatively separated or relatively more integrated manner.
  • The computer system 800 is shown comprising hardware elements that can be electrically coupled via a bus 805 (or may otherwise be in communication, as appropriate). The hardware elements can include one or more processors 810, including without limitation, one or more general purpose processors and/or one or more special purpose processors (such as digital signal processing chips, graphics acceleration chips, and/or the like); one or more input devices 815, which can include without limitation a mouse, a keyboard and/or the like; and one or more output devices 820, which can include without limitation a display device, a printer and/or the like.
  • The computer system 800 may further include (and/or be in communication with) one or more storage devices 825, which can comprise, without limitation, local and/or network accessible storage and/or can include, without limitation, a disk drive, a drive array, an optical storage device, a solid state storage device such as a random access memory (“RAM”) and/or a read-only memory (“ROM”), which can be programmable, flash updateable and/or the like. The computer system 800 might also include a communications subsystem 830, which can include without limitation a modem, a network card (wireless or wired), an infrared communication device, a wireless communication device and/or chipset (such as a Bluetooth™ device, an 802.11 device, a WiFi device, a WiMax device, cellular communication facilities, etc.), and/or the like. The communications subsystem 830 may permit data to be exchanged with a network (such as the network described below, to name one example), and/or any other devices described herein.
  • In many embodiments, the computer system 800 will further comprise a working memory 835, which can include a RAM or ROM device, as described above.
  • The computer system 800 also can comprise software elements, shown as being currently located within the working memory 835, including an operating system 840 and/or other code, such as one or more application programs 845, which may comprise computer programs of the invention, and/or may be designed to implement methods of the invention and/or configure systems of the invention, as described herein. Merely by way of example, one or more procedures described with respect to the method(s) discussed above might be implemented as code and/or instructions executable by a computer (and/or a processor within a computer). A set of these instructions and/or codes might be stored on a computer-readable storage medium, such as the storage device(s) 825 described above. In some cases, the storage medium might be incorporated within a computer system, such as the system 800. In other embodiments, the storage medium might be separate from a computer system (i.e., a removable medium, such as a compact disc, etc.), and is provided in an installation package, such that the storage medium can be used to program a general purpose computer with the instructions/code stored thereon. These instructions might take the form of executable code, which is executable by the computer system 800 and/or might take the form of source and/or installable code, which, upon compilation and/or installation on the computer system 800 (e.g., using any of a variety of generally available compilers, installation programs, compression/decompression utilities, etc.), then takes the form of executable code. In one embodiment, the computer or machine-readable medium may be non-transitory.
  • It will be apparent to those skilled in the art that substantial variations may be made in accordance with specific requirements. For example, customized hardware might also be used, and/or particular elements might be implemented in hardware, software (including portable software, such as applets, etc.), or both. Further, connection to other computing devices such as network input/output devices may be employed.
  • In one aspect, the invention employs a computer system (such as the computer system 800) to perform methods of the invention. According to a set of embodiments, some or all of the procedures of such methods are performed by the computer system 800 in response to processor 810 executing one or more sequences of one or more instructions (which might be incorporated into the operating system 840 and/or other code, such as an application program 845) contained in the working memory 835. Such instructions may be read into the working memory 835 from another machine-readable medium, such as one or more of the storage device(s) 825. Merely by way of example, execution of the sequences of instructions contained in the working memory 835 might cause the processor(s) 810 to perform one or more procedures of the methods described herein.
  • The terms “machine-readable medium” and “computer readable medium,” as used herein, refer to any medium that participates in providing data that causes a machine to operate in a specific fashion. In an embodiment implemented using the computer system 800, various machine-readable media might be involved in providing instructions/code to processor(s) 810 for execution and/or might be used to store and/or carry such instructions/code (e.g., as signals). In many implementations, a computer-readable medium is a physical and/or tangible storage medium. Such a medium may take many forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media includes, for example, optical or magnetic disks, such as the storage device(s) 825. Volatile media includes, without limitation, dynamic memory, such as the working memory 835. Transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise the bus 805, as well as the various components of the communications subsystem 830 (and/or the media by which the communications subsystem 830 provides communication with other devices). Hence, transmission media can also take the form of waves (including without limitation radio, acoustic and/or light waves, such as those generated during radio wave and infrared data communications).
  • Common forms of physical and/or tangible computer-readable media include, for example, a floppy disk, a flexible disk, a hard disk, magnetic tape, or any other magnetic medium, a CD-ROM, any other optical medium, punchcards, papertape, any other physical medium with patterns of holes, a RAM, a PROM, an EPROM, a FLASH-EPROM, any other memory chip or cartridge, a carrier wave as described hereinafter, or any other medium from which a computer can read instructions and/or code.
  • Various forms of machine-readable media may be involved in carrying one or more sequences of one or more instructions to the processor(s) 810 for execution. Merely by way of example, the instructions may initially be carried on a magnetic disk and/or optical disc of a remote computer. A remote computer might load the instructions into its dynamic memory and send the instructions as signals over a transmission medium to be received and/or executed by the computer system 800. These signals, which might be in the form of electromagnetic signals, acoustic signals, optical signals and/or the like, are all examples of carrier waves on which instructions can be encoded, in accordance with various embodiments of the invention.
  • The communications subsystem 830 (and/or components thereof) generally will receive the signals, and the bus 805 then might carry the signals (and/or the data, instructions, etc., carried by the signals) to the working memory 835, from which the processor(s) 810 retrieves and executes the instructions. The instructions received by the working memory 835 may optionally be stored on a storage device 825 either before or after execution by the processor(s) 810.
  • As will be understood by those skilled in the art, the present invention may be embodied in other specific forms. In one particular embodiment of the partial object cache, as previously described, can be associated with a plurality of versions of programming structures. Those skilled in the art will recognize, or be able to ascertain using no more than routine experimentation, many equivalents to the specific embodiments of the invention described herein. Such equivalents are intended to be encompassed by the following claims.

Claims (20)

1. A method of dynamically routing of requests within a plurality of cloud computing networks, the method comprising:
receiving a request for an application from a user device;
forwarding the request to an edge server within a content delivery network (CDN), the edge server being at a point of presence (POP) that is one of a plurality of POPs within the CDN;
analyzing the request to gather metrics for the application configured to run on the plurality of cloud computing networks;
analyzing, within the CDN, historical data for the plurality of cloud computing networks regarding performance of the application;
based on the performance metrics and the historical data, determining an optimal cloud computing network within the plurality of cloud computing networks to route the request;
routing the request to the optimal cloud computing network; and
returning the response from the optimal cloud computing network to the user device.
2. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 1, wherein the determination of the optimal cloud computing network comprises determining a customer's cost preferences and routing the request to a cloud computing network which corresponds to the cost preferences.
3. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 1, where the performance metrics are determined in part by sending ping requests to the plurality of cloud computing networks.
4. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 3, wherein responsiveness is based in part on round trip time of the ping requests.
5. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 4, further comprising storing the performance metrics in a performance database.
6. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 5, further comprising:
dynamically updating the performance database; and
determining the optimal cloud computing network based on data stored within the updated performance database.
7. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 1, wherein the historical data includes logging information for each request processed by each of the plurality of cloud computing networks.
8. A method of dynamically routing of requests within a plurality of clouds computing networks, the method comprising:
receiving a plurality of cloud computing network routing preferences from a customer;
receiving rankings for each of the plurality of cloud computing network routing preferences;
based on the plurality of cloud computing network routing preferences and the rankings, establishing a cloud computing network routing table;
transmitting a status request to each of the plurality of cloud computing networks, wherein the status requests are configured to determine cloud resource and network resource conditions;
receiving cloud resource and network resource status updates;
storing historical information for each of the plurality of cloud computing networks;
based at least in part on the plurality of cloud computing network routing preferences and ranking, the status updates, and the historical information, dynamically updating the cloud computing network routing table, the dynamically updating being performed within a content delivery network (CDN);
receiving at least one application request;
forwarding the request to an edge server within the CDN, the edge server being at a point of presence (POP) that is one of a plurality of POPs within the CDN; and
based on the cloud computing network routine table and the requested at least one application, routine the at least one application request to at least one of the plurality of cloud computing networks.
9. (canceled)
10. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 8, further comprising receiving updates to the plurality of cloud computing network routing preferences.
11. The method of dynamically routing of requests within the plurality of cloud computing networks as in claim 8, further comprising receiving updates to the rankings for each of the plurality of cloud computing network routing preferences.
12. A system for dynamically routing of requests within a plurality of cloud computing networks, the system comprising:
a memory storage device; and
one or more processors coupled with the memory storage device, wherein the memory storage device has sets of instructions stored thereon which, when executed by the one or more processors, cause the one or more processors to:
receive a request for an application from an edge server, the edge server being at a point of presence (POP) that is one of a plurality of POPs within a content delivery network (CDN);
analyze the request which respect to a cloud computing network routing table;
based on the cloud computing network routing table and the requested application, determine a cloud computing network from the plurality of cloud computing networks to route the request;
receive performance information for the application configured to run on each of the plurality of cloud computing networks;
based on the performance information, update, within the CDN, the cloud computing network routing table; and
based on the updated cloud computing network routing table, re-route subsequent requests to one or more of the plurality of the cloud computing networks.
13. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 12, further comprising the CDN.
14. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 13, wherein the CDN includes a request analysis unit, a cloud control unit, a policy database, and the edge server.
15. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 14, wherein the request analysis unit is configured to analyze requests to determine which of the plurality of cloud computing networks to route the requests.
16. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 15, wherein the cloud control unit is configured to receive the analysis from the request analysis unit and route the requests in accordance with the analysis.
17. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 12, wherein the re-routing of the subsequent requests is based on part on one or more of the plurality of cloud computing networks being down or partially down.
18. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 12, wherein sets of instructions stored on the memory storage device, when further executed by the one or more processors, cause the one or more processors to store the performance information in a performance database.
19. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 18, wherein sets of instructions stored on the memory storage device, when further executed by the one or more processors, cause the one or more processors to:
dynamically update the performance database; and
determine an optimal cloud computing network based on data stored within the updated performance database.
20. The system for dynamically routing of requests within the plurality of cloud computing networks as in claim 19, wherein the performance database includes logging information for each request processed by each of the plurality of cloud computing networks.
US13/325,352 2011-09-26 2011-12-14 Dynamic route requests for multiple clouds Abandoned US20130080623A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/325,352 US20130080623A1 (en) 2011-09-26 2011-12-14 Dynamic route requests for multiple clouds

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/245,582 US8849976B2 (en) 2011-09-26 2011-09-26 Dynamic route requests for multiple clouds
US13/325,352 US20130080623A1 (en) 2011-09-26 2011-12-14 Dynamic route requests for multiple clouds

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/245,582 Continuation US8849976B2 (en) 2011-09-26 2011-09-26 Dynamic route requests for multiple clouds

Publications (1)

Publication Number Publication Date
US20130080623A1 true US20130080623A1 (en) 2013-03-28

Family

ID=47912494

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/245,582 Active 2032-01-23 US8849976B2 (en) 2011-09-26 2011-09-26 Dynamic route requests for multiple clouds
US13/325,352 Abandoned US20130080623A1 (en) 2011-09-26 2011-12-14 Dynamic route requests for multiple clouds

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/245,582 Active 2032-01-23 US8849976B2 (en) 2011-09-26 2011-09-26 Dynamic route requests for multiple clouds

Country Status (1)

Country Link
US (2) US8849976B2 (en)

Cited By (85)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130111063A1 (en) * 2011-11-01 2013-05-02 Electronics And Telecommunications Research Institute Routing method in content-centric network
US20130159392A1 (en) * 2011-12-19 2013-06-20 Intellectual Discovery Co., Ltd. System and method for providing virtual device
US20130268674A1 (en) * 2012-04-06 2013-10-10 International Business Machines Corporation Dynamic allocation of workload deployment units across a plurality of clouds
US20150046574A1 (en) * 2013-08-12 2015-02-12 International Business Machines Corporation System to enhance performance, throughput and reliability of an existing cloud offering
US20150052247A1 (en) * 2013-08-14 2015-02-19 Verizon Patent And Licensing Inc. Private cloud topology management system
US20150172354A1 (en) * 2013-12-17 2015-06-18 Limelight Networks, Inc. Content-delivery transfer for cooperative delivery systems
US9086929B2 (en) 2012-04-06 2015-07-21 International Business Machines Corporation Dynamic allocation of a workload across a plurality of clouds
US20150222535A1 (en) * 2014-02-04 2015-08-06 Fastly, Inc. Communication path selection for content delivery
US20150281111A1 (en) * 2014-03-28 2015-10-01 Amazon Technologies, Inc. Implementation of a service that coordinates the placement and execution of containers
US20160182639A1 (en) * 2014-12-17 2016-06-23 University-Industry Cooperation Group Of Kyung-Hee University Internet of things network system using fog computing network
US20160191651A1 (en) * 2014-12-29 2016-06-30 Akamai Technologies, Inc. Managing mobile device user subscription and service preferences to predictively pre-fetch content
US20170041342A1 (en) * 2015-08-04 2017-02-09 AO Kaspersky Lab System and method of utilizing a dedicated computer security service
US20170257429A1 (en) * 2016-03-02 2017-09-07 International Business Machines Corporation Optimization of integration flows in cloud environments
US9860303B1 (en) * 2013-09-27 2018-01-02 Amazon Technologies, Inc. Data center growth control
US9894168B2 (en) 2008-03-31 2018-02-13 Amazon Technologies, Inc. Locality based content distribution
US9893957B2 (en) 2009-10-02 2018-02-13 Amazon Technologies, Inc. Forward-based resource delivery network management techniques
US9912740B2 (en) 2008-06-30 2018-03-06 Amazon Technologies, Inc. Latency measurement in resource requests
US9929959B2 (en) 2013-06-04 2018-03-27 Amazon Technologies, Inc. Managing network computing components utilizing request routing
US9930131B2 (en) 2010-11-22 2018-03-27 Amazon Technologies, Inc. Request routing processing
US9954934B2 (en) 2008-03-31 2018-04-24 Amazon Technologies, Inc. Content delivery reconciliation
CN107995042A (en) * 2017-12-14 2018-05-04 北京云端智度科技有限公司 A kind of network quality appraisal procedure
US9979780B1 (en) * 2012-06-25 2018-05-22 EMC IP Holding Company LLC Method and apparatus for selection between multiple candidate clouds for job processing
US9985927B2 (en) 2008-11-17 2018-05-29 Amazon Technologies, Inc. Managing content delivery network service providers by a content broker
US9992086B1 (en) 2016-08-23 2018-06-05 Amazon Technologies, Inc. External health checking of virtual private cloud network environments
US9992303B2 (en) 2007-06-29 2018-06-05 Amazon Technologies, Inc. Request routing utilizing client location information
US10015241B2 (en) 2012-09-20 2018-07-03 Amazon Technologies, Inc. Automated profiling of resource usage
US10015237B2 (en) 2010-09-28 2018-07-03 Amazon Technologies, Inc. Point of presence management in request routing
US10027582B2 (en) 2007-06-29 2018-07-17 Amazon Technologies, Inc. Updating routing information based on client location
US10033627B1 (en) 2014-12-18 2018-07-24 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10033691B1 (en) 2016-08-24 2018-07-24 Amazon Technologies, Inc. Adaptive resolution of domain name requests in virtual private cloud network environments
US10049051B1 (en) 2015-12-11 2018-08-14 Amazon Technologies, Inc. Reserved cache space in content delivery networks
US10075551B1 (en) 2016-06-06 2018-09-11 Amazon Technologies, Inc. Request management for hierarchical cache
US10079742B1 (en) 2010-09-28 2018-09-18 Amazon Technologies, Inc. Latency measurement in resource requests
US10091096B1 (en) 2014-12-18 2018-10-02 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10097566B1 (en) 2015-07-31 2018-10-09 Amazon Technologies, Inc. Identifying targets of network attacks
US10097398B1 (en) 2010-09-28 2018-10-09 Amazon Technologies, Inc. Point of presence management in request routing
US10097448B1 (en) 2014-12-18 2018-10-09 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10110694B1 (en) 2016-06-29 2018-10-23 Amazon Technologies, Inc. Adaptive transfer rate for retrieving content from a server
US10116584B2 (en) 2008-11-17 2018-10-30 Amazon Technologies, Inc. Managing content delivery network service providers
US10135620B2 (en) 2009-09-04 2018-11-20 Amazon Technologis, Inc. Managing secure content in a content delivery network
US10157135B2 (en) 2008-03-31 2018-12-18 Amazon Technologies, Inc. Cache optimization
US10162753B2 (en) 2009-06-16 2018-12-25 Amazon Technologies, Inc. Managing resources using resource expiration data
US10180993B2 (en) 2015-05-13 2019-01-15 Amazon Technologies, Inc. Routing based request correlation
US10200402B2 (en) 2015-09-24 2019-02-05 Amazon Technologies, Inc. Mitigating network attacks
US10225362B2 (en) 2012-06-11 2019-03-05 Amazon Technologies, Inc. Processing DNS queries to identify pre-processing information
US10225322B2 (en) 2010-09-28 2019-03-05 Amazon Technologies, Inc. Point of presence management in request routing
US10225326B1 (en) 2015-03-23 2019-03-05 Amazon Technologies, Inc. Point of presence based data uploading
US10230819B2 (en) 2009-03-27 2019-03-12 Amazon Technologies, Inc. Translation of resource identifiers using popularity information upon client request
US10257307B1 (en) 2015-12-11 2019-04-09 Amazon Technologies, Inc. Reserved cache space in content delivery networks
US10264062B2 (en) 2009-03-27 2019-04-16 Amazon Technologies, Inc. Request routing using a popularity identifier to identify a cache component
US10270878B1 (en) * 2015-11-10 2019-04-23 Amazon Technologies, Inc. Routing for origin-facing points of presence
US10305797B2 (en) 2008-03-31 2019-05-28 Amazon Technologies, Inc. Request routing based on class
US10331462B1 (en) * 2018-11-06 2019-06-25 Cloudflare, Inc. Cloud computing platform that executes third-party code in a distributed cloud computing network
US10348639B2 (en) 2015-12-18 2019-07-09 Amazon Technologies, Inc. Use of virtual endpoints to improve data transmission rates
US10372499B1 (en) 2016-12-27 2019-08-06 Amazon Technologies, Inc. Efficient region selection system for executing request-driven code
US20190288934A1 (en) * 2018-03-14 2019-09-19 International Business Machines Corporation Dynamically redirecting affiliated data to an edge computing device
US10447648B2 (en) 2017-06-19 2019-10-15 Amazon Technologies, Inc. Assignment of a POP to a DNS resolver based on volume of communications over a link between client devices and the POP
US10469513B2 (en) 2016-10-05 2019-11-05 Amazon Technologies, Inc. Encrypted network addresses
US10469355B2 (en) 2015-03-30 2019-11-05 Amazon Technologies, Inc. Traffic surge management for points of presence
US10491534B2 (en) 2009-03-27 2019-11-26 Amazon Technologies, Inc. Managing resources and entries in tracking information in resource cache components
US10506029B2 (en) 2010-01-28 2019-12-10 Amazon Technologies, Inc. Content distribution network
US10503613B1 (en) 2017-04-21 2019-12-10 Amazon Technologies, Inc. Efficient serving of resources during server unavailability
US10505864B1 (en) 2013-09-30 2019-12-10 EMC IP Holding Company LLC Method for running HPC jobs
US10511567B2 (en) 2008-03-31 2019-12-17 Amazon Technologies, Inc. Network resource identification
US10554748B2 (en) 2008-03-31 2020-02-04 Amazon Technologies, Inc. Content management
US10592578B1 (en) 2018-03-07 2020-03-17 Amazon Technologies, Inc. Predictive content push-enabled content delivery network
US10601718B2 (en) 2017-04-03 2020-03-24 Bank Of America Corporation Data transfer, over session or connection, and between computing device and server associated with a routing network for modifying one or more parameters of the routing network
US10601934B2 (en) 2017-04-03 2020-03-24 Bank Of America Corporation Data transfer, over session or connection, and between computing device and one or more servers for transmitting data to a third party computing device
US10608918B2 (en) 2017-04-03 2020-03-31 Bank Of America Corporation Data transfer, over session or connection, and between computing device and one or more servers to determine likelihood of user device using a routing network
US10609156B2 (en) * 2017-04-03 2020-03-31 Bank Of America Corporation Data transfer, over session or connection, and between computing device and server associated with one or more routing networks in response to detecting activity
US10623408B1 (en) 2012-04-02 2020-04-14 Amazon Technologies, Inc. Context sensitive object management
US10645056B2 (en) 2012-12-19 2020-05-05 Amazon Technologies, Inc. Source-dependent address resolution
US20200186476A1 (en) * 2018-12-05 2020-06-11 Citrix Systems, Inc. System and methods to filter out noisy application signatures to improve precision of first packet classification
US10716060B2 (en) 2017-04-03 2020-07-14 Bank Of America Corporation Data transfer between computing device and user device at different locations and over session or connection to display one or more routing networks to use
US10831549B1 (en) 2016-12-27 2020-11-10 Amazon Technologies, Inc. Multi-region request-driven code execution system
US10862852B1 (en) 2018-11-16 2020-12-08 Amazon Technologies, Inc. Resolution of domain name requests in heterogeneous network environments
US10938884B1 (en) 2017-01-30 2021-03-02 Amazon Technologies, Inc. Origin server cloaking using virtual private cloud network environments
US10958501B1 (en) 2010-09-28 2021-03-23 Amazon Technologies, Inc. Request routing information based on client IP groupings
US11025747B1 (en) 2018-12-12 2021-06-01 Amazon Technologies, Inc. Content request pattern-based routing system
US11075987B1 (en) 2017-06-12 2021-07-27 Amazon Technologies, Inc. Load estimating content delivery network
US11108729B2 (en) 2010-09-28 2021-08-31 Amazon Technologies, Inc. Managing request routing information utilizing client identifiers
US11290418B2 (en) 2017-09-25 2022-03-29 Amazon Technologies, Inc. Hybrid content request routing system
US20220224770A1 (en) * 2021-01-13 2022-07-14 Dell Products, L.P. Client driven cloud network access system and method
US11604667B2 (en) 2011-04-27 2023-03-14 Amazon Technologies, Inc. Optimized deployment based upon customer locality
US20240089218A1 (en) * 2022-09-14 2024-03-14 At&T Intellectual Property I, L.P. System and method of software defined network enabled slicing as a service utilizing artificial intelligence

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104012049B (en) * 2011-10-13 2016-12-07 交互数字专利控股公司 For the method and apparatus providing interface between content delivery network
US10148762B2 (en) * 2011-10-18 2018-12-04 Facebook, Inc. Platform-specific notification delivery channel
US9313299B2 (en) * 2013-12-18 2016-04-12 Verizon Patent And Licensing Inc. Cloud edge appliances to extend cloud applications to customer premises
US20150263983A1 (en) * 2013-12-30 2015-09-17 Stratus Technologies Bermuda Ltd. System and Method for Allocating Resources and Managing a Cloud Based Computer System
US9391917B2 (en) * 2014-04-04 2016-07-12 Xerox Corporation Methods and systems for recommending computational resources
CN104038842B (en) * 2014-06-18 2018-09-18 百视通网络电视技术发展有限责任公司 A kind of method and apparatus prefetching request program information in CDN network
US10965608B2 (en) * 2014-06-24 2021-03-30 Keepsayk LLC Mobile supercloud computing system and method
US10296391B2 (en) 2014-06-30 2019-05-21 Microsoft Technology Licensing, Llc Assigning a player to a machine
GB2528473B (en) 2014-07-23 2016-06-22 Ibm Effective roaming for software-as-a-service infrastructure
US9888127B2 (en) 2015-07-30 2018-02-06 At&T Intellectual Property I, L.P. Methods, systems, and computer readable storage devices for adjusting the use of virtual resources providing communication services based on load
US10277736B2 (en) 2015-07-30 2019-04-30 At&T Intellectual Property I, L.P. Methods, systems, and computer readable storage devices for determining whether to handle a request for communication services by a physical telephone number mapping service or a virtual telephone number mapping service
US9851999B2 (en) 2015-07-30 2017-12-26 At&T Intellectual Property I, L.P. Methods, systems, and computer readable storage devices for handling virtualization of a physical telephone number mapping service
US9866521B2 (en) 2015-07-30 2018-01-09 At&T Intellectual Property L.L.P. Methods, systems, and computer readable storage devices for determining whether to forward requests from a physical telephone number mapping service server to a virtual telephone number mapping service server
US10063666B2 (en) 2016-06-14 2018-08-28 Futurewei Technologies, Inc. Modular telecommunication edge cloud system
US10616315B2 (en) 2016-07-20 2020-04-07 International Business Machines Corporation Load balancing system
CN106776005B (en) * 2016-11-23 2019-12-13 华中科技大学 Resource management system and method for containerized application
CN111656754B (en) 2018-07-13 2023-01-13 三星电子株式会社 Method for edge computing service and electronic device thereof
WO2020013677A1 (en) * 2018-07-13 2020-01-16 삼성전자 주식회사 Method and electronic device for edge computing service
US10834222B1 (en) 2019-05-09 2020-11-10 Akamai Technologies Inc. Server utilizing multiple object retrieval candidates
US11233768B1 (en) 2020-09-30 2022-01-25 Akamai Technologies, Inc. CDN configuration tuning based on domain scan analysis

Citations (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6052718A (en) * 1997-01-07 2000-04-18 Sightpath, Inc Replica routing
US20020055989A1 (en) * 2000-11-08 2002-05-09 Stringer-Calvert David W.J. Methods and apparatus for scalable, distributed management of virtual private networks
US20030065816A1 (en) * 2001-09-28 2003-04-03 Intel Corporation User-preferred network interface switching using route table manipulation
US20030079005A1 (en) * 2001-05-29 2003-04-24 61C Networks, Inc. System and method for efficient wide area network routing
US6785704B1 (en) * 1999-12-20 2004-08-31 Fastforward Networks Content distribution system for operation over an internetwork including content peering arrangements
US20040215816A1 (en) * 2002-12-16 2004-10-28 Hayes Stephen T Apparatus and methods for communication among devices
US6981055B1 (en) * 2000-08-22 2005-12-27 Internap Network Services Corporation Method and system for optimizing routing through multiple available internet route providers
US20070078960A1 (en) * 2005-10-04 2007-04-05 International Business Machines Corporation Grid computing accounting and statistics management system
US7245640B2 (en) * 2002-12-18 2007-07-17 Intel Corporation Packet origination
US7275103B1 (en) * 2002-12-18 2007-09-25 Veritas Operating Corporation Storage path optimization for SANs
US7281041B2 (en) * 2002-10-31 2007-10-09 Hewlett-Packard Development Company, L.P. Method and apparatus for providing a baselining and auto-thresholding framework
US7472192B2 (en) * 2000-05-15 2008-12-30 Arbinet-Thexchange Ltd. Communication system and method for connecting to multiple ISP's
US7599323B2 (en) * 2002-10-17 2009-10-06 Alcatel-Lucent Usa Inc. Multi-interface mobility client
US20100088205A1 (en) * 2008-10-02 2010-04-08 Verizon Business Network Services Inc. Methods, Systems and Computer Program Products for a Cloud Computing Spot Market Platform
US20100179987A1 (en) * 2009-01-13 2010-07-15 Viasat, Inc. Content set based pre-positioning
US7760742B2 (en) * 2005-03-15 2010-07-20 Hewlett-Packard Development Company, L.P. Systems and methods of using pseudo-routes for data transmissions
US20100223378A1 (en) * 2009-02-27 2010-09-02 Yottaa Inc System and method for computer cloud management
US20100250710A1 (en) * 2009-03-25 2010-09-30 Limelight Networks, Inc. Publishing-point management for content delivery network
US20100333116A1 (en) * 2009-06-30 2010-12-30 Anand Prahlad Cloud gateway system for managing data storage to cloud storage sites
US20110282982A1 (en) * 2010-05-13 2011-11-17 Microsoft Corporation Dynamic application placement based on cost and availability of energy in datacenters
US20120005359A1 (en) * 2010-07-01 2012-01-05 Scott Wayne Seago System and method for aggregation across cloud providers
US20120089726A1 (en) * 2010-10-06 2012-04-12 Infosys Technologies Ltd. Automated service level management of applications in cloud computing environment
US20120317291A1 (en) * 2010-02-17 2012-12-13 Alexander Wolfe Content and application delivery network aggregation

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6421726B1 (en) 1997-03-14 2002-07-16 Akamai Technologies, Inc. System and method for selection and retrieval of diverse types of video data on a computer network
WO2002071242A1 (en) 2001-03-01 2002-09-12 Akamai Technologies, Inc. Optimal route selection in a content delivery network
US7860081B2 (en) 2007-05-31 2010-12-28 International Business Machines Corporation Optimization process and system for multiplexed gateway architecture
US8072951B2 (en) * 2007-06-15 2011-12-06 Silver Spring Networks, Inc. Method and system for providing routing protocols in a frequency hopping spread spectrum network
GB2456026A (en) 2007-12-26 2009-07-01 Contendo Inc CDN balancing and sharing platform
US8180896B2 (en) 2008-08-06 2012-05-15 Edgecast Networks, Inc. Global load balancing on a content delivery network
EP2415207B1 (en) * 2009-03-31 2014-12-03 Coach Wei System and method for access management and security protection for network accessible computer services
WO2011071850A2 (en) * 2009-12-07 2011-06-16 Coach Wei System and method for website performance optimization and internet traffic processing
US8402139B2 (en) * 2010-02-26 2013-03-19 Red Hat, Inc. Methods and systems for matching resource requests with cloud computing environments
US8904005B2 (en) * 2010-11-23 2014-12-02 Red Hat, Inc. Indentifying service dependencies in a cloud deployment
US9015469B2 (en) * 2011-07-28 2015-04-21 Cloudflare, Inc. Supporting secure sessions in a cloud-based proxy service

Patent Citations (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7149771B1 (en) * 1997-01-07 2006-12-12 Cisco Technology, Inc. Replica routing
US6052718A (en) * 1997-01-07 2000-04-18 Sightpath, Inc Replica routing
US6785704B1 (en) * 1999-12-20 2004-08-31 Fastforward Networks Content distribution system for operation over an internetwork including content peering arrangements
US7472192B2 (en) * 2000-05-15 2008-12-30 Arbinet-Thexchange Ltd. Communication system and method for connecting to multiple ISP's
US6981055B1 (en) * 2000-08-22 2005-12-27 Internap Network Services Corporation Method and system for optimizing routing through multiple available internet route providers
US20020055989A1 (en) * 2000-11-08 2002-05-09 Stringer-Calvert David W.J. Methods and apparatus for scalable, distributed management of virtual private networks
US20030079005A1 (en) * 2001-05-29 2003-04-24 61C Networks, Inc. System and method for efficient wide area network routing
US20030065816A1 (en) * 2001-09-28 2003-04-03 Intel Corporation User-preferred network interface switching using route table manipulation
US7599323B2 (en) * 2002-10-17 2009-10-06 Alcatel-Lucent Usa Inc. Multi-interface mobility client
US7281041B2 (en) * 2002-10-31 2007-10-09 Hewlett-Packard Development Company, L.P. Method and apparatus for providing a baselining and auto-thresholding framework
US20040215816A1 (en) * 2002-12-16 2004-10-28 Hayes Stephen T Apparatus and methods for communication among devices
US7245640B2 (en) * 2002-12-18 2007-07-17 Intel Corporation Packet origination
US7275103B1 (en) * 2002-12-18 2007-09-25 Veritas Operating Corporation Storage path optimization for SANs
US7760742B2 (en) * 2005-03-15 2010-07-20 Hewlett-Packard Development Company, L.P. Systems and methods of using pseudo-routes for data transmissions
US20070078960A1 (en) * 2005-10-04 2007-04-05 International Business Machines Corporation Grid computing accounting and statistics management system
US20100088205A1 (en) * 2008-10-02 2010-04-08 Verizon Business Network Services Inc. Methods, Systems and Computer Program Products for a Cloud Computing Spot Market Platform
US20100179987A1 (en) * 2009-01-13 2010-07-15 Viasat, Inc. Content set based pre-positioning
US8489673B2 (en) * 2009-01-13 2013-07-16 Viasat, Inc. Content set based pre-positioning
US20100223378A1 (en) * 2009-02-27 2010-09-02 Yottaa Inc System and method for computer cloud management
US20100250710A1 (en) * 2009-03-25 2010-09-30 Limelight Networks, Inc. Publishing-point management for content delivery network
US20100333116A1 (en) * 2009-06-30 2010-12-30 Anand Prahlad Cloud gateway system for managing data storage to cloud storage sites
US20120317291A1 (en) * 2010-02-17 2012-12-13 Alexander Wolfe Content and application delivery network aggregation
US20110282982A1 (en) * 2010-05-13 2011-11-17 Microsoft Corporation Dynamic application placement based on cost and availability of energy in datacenters
US20120005359A1 (en) * 2010-07-01 2012-01-05 Scott Wayne Seago System and method for aggregation across cloud providers
US20120089726A1 (en) * 2010-10-06 2012-04-12 Infosys Technologies Ltd. Automated service level management of applications in cloud computing environment

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Li et al., "CloubCmp: comparing public cloud providers", IMC 2010 Proceedings of the 10th ACM SIGCOMM conference on Internet measurement, pg.1-14, ISBN: 9781-4503-0483-2 *
Rosberg et al. ("Architecture for Open Content Delivery Networks", ARC Special Research Centre for Ultra-Broadband Information Networks, The University of Melbourne, 03/15/2002) *
Rosberg et al., "Architecture for Open Content Delivery Networks", ARC Special Research Centre for Ultra-Broadband Information Networks, The University of Melbourne, 03/15/2002. *
Salekul et al., "Network Edge Intelligence for the Emerging Next-Generation Internet", future internet, ISSN 1999-5903, Published November 5, 2010 *
Salekul et all., "Network Edge Intelligence for Emerging Next-Generation Internet", 11/05/2010, Future Internet ISSN 1999-5903 *
Wieder et al., "Conductor: orchestrating the clouds", LADIS '10 Proceedings of the 4th International Workshop on Large Scale Distributed Systems and Middleware, pg.44-48, ISBN: 978-1-4503-0406-1, 2010 Article *

Cited By (162)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10027582B2 (en) 2007-06-29 2018-07-17 Amazon Technologies, Inc. Updating routing information based on client location
US9992303B2 (en) 2007-06-29 2018-06-05 Amazon Technologies, Inc. Request routing utilizing client location information
US10511567B2 (en) 2008-03-31 2019-12-17 Amazon Technologies, Inc. Network resource identification
US10771552B2 (en) 2008-03-31 2020-09-08 Amazon Technologies, Inc. Content management
US10554748B2 (en) 2008-03-31 2020-02-04 Amazon Technologies, Inc. Content management
US9894168B2 (en) 2008-03-31 2018-02-13 Amazon Technologies, Inc. Locality based content distribution
US10797995B2 (en) 2008-03-31 2020-10-06 Amazon Technologies, Inc. Request routing based on class
US10645149B2 (en) 2008-03-31 2020-05-05 Amazon Technologies, Inc. Content delivery reconciliation
US10305797B2 (en) 2008-03-31 2019-05-28 Amazon Technologies, Inc. Request routing based on class
US11194719B2 (en) 2008-03-31 2021-12-07 Amazon Technologies, Inc. Cache optimization
US9954934B2 (en) 2008-03-31 2018-04-24 Amazon Technologies, Inc. Content delivery reconciliation
US10530874B2 (en) 2008-03-31 2020-01-07 Amazon Technologies, Inc. Locality based content distribution
US10157135B2 (en) 2008-03-31 2018-12-18 Amazon Technologies, Inc. Cache optimization
US11245770B2 (en) 2008-03-31 2022-02-08 Amazon Technologies, Inc. Locality based content distribution
US11909639B2 (en) 2008-03-31 2024-02-20 Amazon Technologies, Inc. Request routing based on class
US11451472B2 (en) 2008-03-31 2022-09-20 Amazon Technologies, Inc. Request routing based on class
US10158729B2 (en) 2008-03-31 2018-12-18 Amazon Technologies, Inc. Locality based content distribution
US9912740B2 (en) 2008-06-30 2018-03-06 Amazon Technologies, Inc. Latency measurement in resource requests
US9985927B2 (en) 2008-11-17 2018-05-29 Amazon Technologies, Inc. Managing content delivery network service providers by a content broker
US10523783B2 (en) 2008-11-17 2019-12-31 Amazon Technologies, Inc. Request routing utilizing client location information
US11115500B2 (en) 2008-11-17 2021-09-07 Amazon Technologies, Inc. Request routing utilizing client location information
US11283715B2 (en) 2008-11-17 2022-03-22 Amazon Technologies, Inc. Updating routing information based on client location
US11811657B2 (en) 2008-11-17 2023-11-07 Amazon Technologies, Inc. Updating routing information based on client location
US10742550B2 (en) 2008-11-17 2020-08-11 Amazon Technologies, Inc. Updating routing information based on client location
US10116584B2 (en) 2008-11-17 2018-10-30 Amazon Technologies, Inc. Managing content delivery network service providers
US10574787B2 (en) 2009-03-27 2020-02-25 Amazon Technologies, Inc. Translation of resource identifiers using popularity information upon client request
US10491534B2 (en) 2009-03-27 2019-11-26 Amazon Technologies, Inc. Managing resources and entries in tracking information in resource cache components
US10230819B2 (en) 2009-03-27 2019-03-12 Amazon Technologies, Inc. Translation of resource identifiers using popularity information upon client request
US10264062B2 (en) 2009-03-27 2019-04-16 Amazon Technologies, Inc. Request routing using a popularity identifier to identify a cache component
US10783077B2 (en) 2009-06-16 2020-09-22 Amazon Technologies, Inc. Managing resources using resource expiration data
US10521348B2 (en) 2009-06-16 2019-12-31 Amazon Technologies, Inc. Managing resources using resource expiration data
US10162753B2 (en) 2009-06-16 2018-12-25 Amazon Technologies, Inc. Managing resources using resource expiration data
US10135620B2 (en) 2009-09-04 2018-11-20 Amazon Technologis, Inc. Managing secure content in a content delivery network
US10785037B2 (en) 2009-09-04 2020-09-22 Amazon Technologies, Inc. Managing secure content in a content delivery network
US9893957B2 (en) 2009-10-02 2018-02-13 Amazon Technologies, Inc. Forward-based resource delivery network management techniques
US10218584B2 (en) 2009-10-02 2019-02-26 Amazon Technologies, Inc. Forward-based resource delivery network management techniques
US11205037B2 (en) 2010-01-28 2021-12-21 Amazon Technologies, Inc. Content distribution network
US10506029B2 (en) 2010-01-28 2019-12-10 Amazon Technologies, Inc. Content distribution network
US10931738B2 (en) 2010-09-28 2021-02-23 Amazon Technologies, Inc. Point of presence management in request routing
US10225322B2 (en) 2010-09-28 2019-03-05 Amazon Technologies, Inc. Point of presence management in request routing
US11336712B2 (en) 2010-09-28 2022-05-17 Amazon Technologies, Inc. Point of presence management in request routing
US11632420B2 (en) 2010-09-28 2023-04-18 Amazon Technologies, Inc. Point of presence management in request routing
US10079742B1 (en) 2010-09-28 2018-09-18 Amazon Technologies, Inc. Latency measurement in resource requests
US10015237B2 (en) 2010-09-28 2018-07-03 Amazon Technologies, Inc. Point of presence management in request routing
US10958501B1 (en) 2010-09-28 2021-03-23 Amazon Technologies, Inc. Request routing information based on client IP groupings
US10097398B1 (en) 2010-09-28 2018-10-09 Amazon Technologies, Inc. Point of presence management in request routing
US10778554B2 (en) 2010-09-28 2020-09-15 Amazon Technologies, Inc. Latency measurement in resource requests
US11108729B2 (en) 2010-09-28 2021-08-31 Amazon Technologies, Inc. Managing request routing information utilizing client identifiers
US9930131B2 (en) 2010-11-22 2018-03-27 Amazon Technologies, Inc. Request routing processing
US10951725B2 (en) 2010-11-22 2021-03-16 Amazon Technologies, Inc. Request routing processing
US11604667B2 (en) 2011-04-27 2023-03-14 Amazon Technologies, Inc. Optimized deployment based upon customer locality
US20130111063A1 (en) * 2011-11-01 2013-05-02 Electronics And Telecommunications Research Institute Routing method in content-centric network
US20130159392A1 (en) * 2011-12-19 2013-06-20 Intellectual Discovery Co., Ltd. System and method for providing virtual device
US9207989B2 (en) * 2011-12-19 2015-12-08 Intellectual Discovery Co., Ltd. System and method for providing virtual device
US10623408B1 (en) 2012-04-02 2020-04-14 Amazon Technologies, Inc. Context sensitive object management
US10069761B2 (en) * 2012-04-06 2018-09-04 International Business Machines Corporation Dynamic allocation of workload deployment units across a plurality of clouds
US20130268674A1 (en) * 2012-04-06 2013-10-10 International Business Machines Corporation Dynamic allocation of workload deployment units across a plurality of clouds
US10554740B2 (en) 2012-04-06 2020-02-04 International Business Machines Corporation Dynamic allocation of a workload across a plurality of clouds
US9071613B2 (en) * 2012-04-06 2015-06-30 International Business Machines Corporation Dynamic allocation of workload deployment units across a plurality of clouds
US9086929B2 (en) 2012-04-06 2015-07-21 International Business Machines Corporation Dynamic allocation of a workload across a plurality of clouds
US20150244646A1 (en) * 2012-04-06 2015-08-27 International Business Machines Corporation Dynamic allocation of workload deployment units across a plurality of clouds
US11431651B2 (en) 2012-04-06 2022-08-30 International Business Machines Corporation Dynamic allocation of workload deployment units across a plurality of clouds
US10225362B2 (en) 2012-06-11 2019-03-05 Amazon Technologies, Inc. Processing DNS queries to identify pre-processing information
US11729294B2 (en) 2012-06-11 2023-08-15 Amazon Technologies, Inc. Processing DNS queries to identify pre-processing information
US11303717B2 (en) 2012-06-11 2022-04-12 Amazon Technologies, Inc. Processing DNS queries to identify pre-processing information
US9979780B1 (en) * 2012-06-25 2018-05-22 EMC IP Holding Company LLC Method and apparatus for selection between multiple candidate clouds for job processing
US10542079B2 (en) 2012-09-20 2020-01-21 Amazon Technologies, Inc. Automated profiling of resource usage
US10015241B2 (en) 2012-09-20 2018-07-03 Amazon Technologies, Inc. Automated profiling of resource usage
US10645056B2 (en) 2012-12-19 2020-05-05 Amazon Technologies, Inc. Source-dependent address resolution
US9929959B2 (en) 2013-06-04 2018-03-27 Amazon Technologies, Inc. Managing network computing components utilizing request routing
US10374955B2 (en) 2013-06-04 2019-08-06 Amazon Technologies, Inc. Managing network computing components utilizing request routing
US9246778B2 (en) * 2013-08-12 2016-01-26 International Business Machines Corporation System to enhance performance, throughput and reliability of an existing cloud offering
US20150046574A1 (en) * 2013-08-12 2015-02-12 International Business Machines Corporation System to enhance performance, throughput and reliability of an existing cloud offering
US20150046583A1 (en) * 2013-08-12 2015-02-12 International Business Machines Corporation System to enhance performance, throughput and reliability of an existing cloud offering
US9253056B2 (en) * 2013-08-12 2016-02-02 International Business Machines Corporation System to enhance performance, throughput and reliability of an existing cloud offering
US20150052247A1 (en) * 2013-08-14 2015-02-19 Verizon Patent And Licensing Inc. Private cloud topology management system
US9338223B2 (en) * 2013-08-14 2016-05-10 Verizon Patent And Licensing Inc. Private cloud topology management system
US9860303B1 (en) * 2013-09-27 2018-01-02 Amazon Technologies, Inc. Data center growth control
US10505864B1 (en) 2013-09-30 2019-12-10 EMC IP Holding Company LLC Method for running HPC jobs
US20150172354A1 (en) * 2013-12-17 2015-06-18 Limelight Networks, Inc. Content-delivery transfer for cooperative delivery systems
US9887914B2 (en) * 2014-02-04 2018-02-06 Fastly, Inc. Communication path selection for content delivery
US10536372B2 (en) 2014-02-04 2020-01-14 Fastly Inc. Communication path selection for content delivery
US20150222535A1 (en) * 2014-02-04 2015-08-06 Fastly, Inc. Communication path selection for content delivery
US10218633B2 (en) * 2014-03-28 2019-02-26 Amazon Technologies, Inc. Implementation of a service that coordinates the placement and execution of containers
US20150281111A1 (en) * 2014-03-28 2015-10-01 Amazon Technologies, Inc. Implementation of a service that coordinates the placement and execution of containers
US20160182639A1 (en) * 2014-12-17 2016-06-23 University-Industry Cooperation Group Of Kyung-Hee University Internet of things network system using fog computing network
US11381487B2 (en) 2014-12-18 2022-07-05 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10097448B1 (en) 2014-12-18 2018-10-09 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10728133B2 (en) 2014-12-18 2020-07-28 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10091096B1 (en) 2014-12-18 2018-10-02 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10033627B1 (en) 2014-12-18 2018-07-24 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US11863417B2 (en) 2014-12-18 2024-01-02 Amazon Technologies, Inc. Routing mode and point-of-presence selection service
US10771583B2 (en) * 2014-12-29 2020-09-08 Akamai Technologies, Inc. Managing mobile device user subscription and service preferences to predictively pre-fetch content
US20200404068A1 (en) * 2014-12-29 2020-12-24 Akamai Technologies, Inc. Managing mobile device user subscription and service preferences to predictively pre-fetch content
US20160191651A1 (en) * 2014-12-29 2016-06-30 Akamai Technologies, Inc. Managing mobile device user subscription and service preferences to predictively pre-fetch content
US11509741B2 (en) * 2014-12-29 2022-11-22 Akamai Technologies, Inc. Managing mobile device user subscription and service preferences to predictively pre-fetch content
US10225326B1 (en) 2015-03-23 2019-03-05 Amazon Technologies, Inc. Point of presence based data uploading
US11297140B2 (en) 2015-03-23 2022-04-05 Amazon Technologies, Inc. Point of presence based data uploading
US10469355B2 (en) 2015-03-30 2019-11-05 Amazon Technologies, Inc. Traffic surge management for points of presence
US11461402B2 (en) 2015-05-13 2022-10-04 Amazon Technologies, Inc. Routing based request correlation
US10180993B2 (en) 2015-05-13 2019-01-15 Amazon Technologies, Inc. Routing based request correlation
US10691752B2 (en) 2015-05-13 2020-06-23 Amazon Technologies, Inc. Routing based request correlation
US10097566B1 (en) 2015-07-31 2018-10-09 Amazon Technologies, Inc. Identifying targets of network attacks
US9667657B2 (en) * 2015-08-04 2017-05-30 AO Kaspersky Lab System and method of utilizing a dedicated computer security service
US20170041342A1 (en) * 2015-08-04 2017-02-09 AO Kaspersky Lab System and method of utilizing a dedicated computer security service
US10200402B2 (en) 2015-09-24 2019-02-05 Amazon Technologies, Inc. Mitigating network attacks
US10270878B1 (en) * 2015-11-10 2019-04-23 Amazon Technologies, Inc. Routing for origin-facing points of presence
US11134134B2 (en) 2015-11-10 2021-09-28 Amazon Technologies, Inc. Routing for origin-facing points of presence
US10049051B1 (en) 2015-12-11 2018-08-14 Amazon Technologies, Inc. Reserved cache space in content delivery networks
US10257307B1 (en) 2015-12-11 2019-04-09 Amazon Technologies, Inc. Reserved cache space in content delivery networks
US10348639B2 (en) 2015-12-18 2019-07-09 Amazon Technologies, Inc. Use of virtual endpoints to improve data transmission rates
US20170257429A1 (en) * 2016-03-02 2017-09-07 International Business Machines Corporation Optimization of integration flows in cloud environments
US11032360B2 (en) 2016-03-02 2021-06-08 International Business Machines Corporation Optimization of integration flows in cloud environments
US10432707B2 (en) * 2016-03-02 2019-10-01 International Business Machines Corporation Optimization of integration flows in cloud environments
US10666756B2 (en) 2016-06-06 2020-05-26 Amazon Technologies, Inc. Request management for hierarchical cache
US11463550B2 (en) 2016-06-06 2022-10-04 Amazon Technologies, Inc. Request management for hierarchical cache
US10075551B1 (en) 2016-06-06 2018-09-11 Amazon Technologies, Inc. Request management for hierarchical cache
US11457088B2 (en) 2016-06-29 2022-09-27 Amazon Technologies, Inc. Adaptive transfer rate for retrieving content from a server
US10110694B1 (en) 2016-06-29 2018-10-23 Amazon Technologies, Inc. Adaptive transfer rate for retrieving content from a server
US9992086B1 (en) 2016-08-23 2018-06-05 Amazon Technologies, Inc. External health checking of virtual private cloud network environments
US10516590B2 (en) 2016-08-23 2019-12-24 Amazon Technologies, Inc. External health checking of virtual private cloud network environments
US10033691B1 (en) 2016-08-24 2018-07-24 Amazon Technologies, Inc. Adaptive resolution of domain name requests in virtual private cloud network environments
US10469442B2 (en) 2016-08-24 2019-11-05 Amazon Technologies, Inc. Adaptive resolution of domain name requests in virtual private cloud network environments
US10505961B2 (en) 2016-10-05 2019-12-10 Amazon Technologies, Inc. Digitally signed network address
US10469513B2 (en) 2016-10-05 2019-11-05 Amazon Technologies, Inc. Encrypted network addresses
US11330008B2 (en) 2016-10-05 2022-05-10 Amazon Technologies, Inc. Network addresses with encoded DNS-level information
US10616250B2 (en) 2016-10-05 2020-04-07 Amazon Technologies, Inc. Network addresses with encoded DNS-level information
US10831549B1 (en) 2016-12-27 2020-11-10 Amazon Technologies, Inc. Multi-region request-driven code execution system
US10372499B1 (en) 2016-12-27 2019-08-06 Amazon Technologies, Inc. Efficient region selection system for executing request-driven code
US20210042163A1 (en) * 2016-12-27 2021-02-11 Amazon Technologies, Inc. Multi-region request-driven code execution system
US11762703B2 (en) * 2016-12-27 2023-09-19 Amazon Technologies, Inc. Multi-region request-driven code execution system
US10938884B1 (en) 2017-01-30 2021-03-02 Amazon Technologies, Inc. Origin server cloaking using virtual private cloud network environments
US10608918B2 (en) 2017-04-03 2020-03-31 Bank Of America Corporation Data transfer, over session or connection, and between computing device and one or more servers to determine likelihood of user device using a routing network
US10798007B2 (en) 2017-04-03 2020-10-06 Bank Of America Corporation Data transfer, over session or connection, and between computing device and server associated with a routing network for modifying one or more parameters of the routing network
US10716060B2 (en) 2017-04-03 2020-07-14 Bank Of America Corporation Data transfer between computing device and user device at different locations and over session or connection to display one or more routing networks to use
US10609156B2 (en) * 2017-04-03 2020-03-31 Bank Of America Corporation Data transfer, over session or connection, and between computing device and server associated with one or more routing networks in response to detecting activity
US10601934B2 (en) 2017-04-03 2020-03-24 Bank Of America Corporation Data transfer, over session or connection, and between computing device and one or more servers for transmitting data to a third party computing device
US10601718B2 (en) 2017-04-03 2020-03-24 Bank Of America Corporation Data transfer, over session or connection, and between computing device and server associated with a routing network for modifying one or more parameters of the routing network
US10503613B1 (en) 2017-04-21 2019-12-10 Amazon Technologies, Inc. Efficient serving of resources during server unavailability
US11075987B1 (en) 2017-06-12 2021-07-27 Amazon Technologies, Inc. Load estimating content delivery network
US10447648B2 (en) 2017-06-19 2019-10-15 Amazon Technologies, Inc. Assignment of a POP to a DNS resolver based on volume of communications over a link between client devices and the POP
US11290418B2 (en) 2017-09-25 2022-03-29 Amazon Technologies, Inc. Hybrid content request routing system
CN107995042A (en) * 2017-12-14 2018-05-04 北京云端智度科技有限公司 A kind of network quality appraisal procedure
US10592578B1 (en) 2018-03-07 2020-03-17 Amazon Technologies, Inc. Predictive content push-enabled content delivery network
US20190288934A1 (en) * 2018-03-14 2019-09-19 International Business Machines Corporation Dynamically redirecting affiliated data to an edge computing device
US10567269B2 (en) * 2018-03-14 2020-02-18 International Business Machines Corporation Dynamically redirecting affiliated data to an edge computing device
US10860340B2 (en) * 2018-11-06 2020-12-08 Cloudflare, Inc. Cloud computing platform that executes third-party code in a distributed cloud computing network
US10331462B1 (en) * 2018-11-06 2019-06-25 Cloudflare, Inc. Cloud computing platform that executes third-party code in a distributed cloud computing network
US11853776B2 (en) * 2018-11-06 2023-12-26 Cloudflare, Inc. Cloud computing platform that executes third-party code in a distributed cloud computing network
US20210089328A1 (en) * 2018-11-06 2021-03-25 Cloudflare, Inc. Cloud Computing Platform That Executes Third-Party Code in A Distributed Cloud Computing Network
US11561805B2 (en) * 2018-11-06 2023-01-24 Cloudflare, Inc. Cloud computing platform that executes third-party code in a distributed cloud computing network
US20200142711A1 (en) * 2018-11-06 2020-05-07 Cloudflare, Inc. Cloud Computing Platform That Executes Third-Party Code in A Distributed Cloud Computing Network
US10862852B1 (en) 2018-11-16 2020-12-08 Amazon Technologies, Inc. Resolution of domain name requests in heterogeneous network environments
US11362986B2 (en) 2018-11-16 2022-06-14 Amazon Technologies, Inc. Resolution of domain name requests in heterogeneous network environments
US20210152481A1 (en) * 2018-12-05 2021-05-20 Citrix Systems, Inc. System and methods to filter out noisy application signatures to improve precision of first packet classification
US10911362B2 (en) * 2018-12-05 2021-02-02 Citrix Systems, Inc. System and methods to filter out noisy application signatures to improve precision of first packet classification
US11582158B2 (en) * 2018-12-05 2023-02-14 Citrix Systems, Inc. System and methods to filter out noisy application signatures to improve precision of first packet classification
US20200186476A1 (en) * 2018-12-05 2020-06-11 Citrix Systems, Inc. System and methods to filter out noisy application signatures to improve precision of first packet classification
US11025747B1 (en) 2018-12-12 2021-06-01 Amazon Technologies, Inc. Content request pattern-based routing system
US20220224770A1 (en) * 2021-01-13 2022-07-14 Dell Products, L.P. Client driven cloud network access system and method
US11496601B2 (en) * 2021-01-13 2022-11-08 Dell Products, L.P. Client driven cloud network access system and method
US20240089218A1 (en) * 2022-09-14 2024-03-14 At&T Intellectual Property I, L.P. System and method of software defined network enabled slicing as a service utilizing artificial intelligence

Also Published As

Publication number Publication date
US20130080613A1 (en) 2013-03-28
US8849976B2 (en) 2014-09-30

Similar Documents

Publication Publication Date Title
US8539079B2 (en) Edge-based resource spin-up for cloud computing
US8849976B2 (en) Dynamic route requests for multiple clouds
WO2013049079A2 (en) Dynamic request handling using cloud-computing capabilities
US11863417B2 (en) Routing mode and point-of-presence selection service
US8745239B2 (en) Edge-based resource spin-up for cloud computing
US10091096B1 (en) Routing mode and point-of-presence selection service
US10033627B1 (en) Routing mode and point-of-presence selection service
US10250525B2 (en) Intent-based services orchestration
Gomes et al. Edge caching with mobility prediction in virtualized LTE mobile networks
US11924650B2 (en) System, method and service product for content delivery
EP3296870B1 (en) Cdn-based content management system
US9634922B2 (en) Apparatus, system, and method for cloud-assisted routing
US10601698B2 (en) Techniques for managing telemetry data for content delivery and/or data transfer networks
JP7037511B2 (en) Base stations, access request response methods, equipment and systems
Sardis et al. On the investigation of cloud-based mobile media environments with service-populating and QoS-aware mechanisms
US20150046591A1 (en) Dynamic edge server allocation
KR20150121713A (en) Allocating resources between network nodes for providing a network node function
Mukerjee et al. The impact of brokers on the future of content delivery
KR20110053906A (en) Method and system for optimization of multimedia service over ims network
Ibn-Khedher et al. OPAC: An optimal placement algorithm for virtual CDN
CN106209952B (en) Service node distribution method and device, CDN management server and system
Meng et al. Elastic caching solutions for content dissemination services of ip-based internet technologies prospective
CN115514981A (en) Video stream scheduling system and related product
EP2913979A1 (en) A method and system to process traffic optimization requests
Frank Dynamic content delivery infrastructure deployment using network cloud resources

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION