[go: nahoru, domu]

US8930543B2 - Dynamically building a set of compute nodes to host the user's workload - Google Patents

Dynamically building a set of compute nodes to host the user's workload Download PDF

Info

Publication number
US8930543B2
US8930543B2 US13/858,849 US201313858849A US8930543B2 US 8930543 B2 US8930543 B2 US 8930543B2 US 201313858849 A US201313858849 A US 201313858849A US 8930543 B2 US8930543 B2 US 8930543B2
Authority
US
United States
Prior art keywords
virtual machines
cloud group
cloud
compute nodes
workload
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US13/858,849
Other versions
US20130227131A1 (en
Inventor
Rohith K. Ashok
Roy F. Brabson
Hugh E. Hockett
Matt R. Hogstrom
Wolfgang Segmuller
Matthew J. Sheard
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US13/858,849 priority Critical patent/US8930543B2/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BRABSON, ROY F., SEGMULLER, WOLFGANG, HOCKETT, HUGH E., HOGSTROM, MATT R., SHEARD, MATTHEW J., ASHOK, ROHITH K.
Publication of US20130227131A1 publication Critical patent/US20130227131A1/en
Application granted granted Critical
Publication of US8930543B2 publication Critical patent/US8930543B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1031Controlling of the operation of servers by a load balancer, e.g. adding or removing servers that serve requests
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5072Grid computing

Definitions

  • the present invention relates to cloud computing, and more particularly to dynamically building a set of compute nodes to host a user's workload.
  • computing is delivered as a service rather than a product, whereby shared resources, software and information are provided to computers and other devices as a metered service over a network, such as the Internet.
  • a network such as the Internet.
  • computation, software, data access and storage services are provided to users that do not require knowledge of the physical location and configuration of the system that delivers the services.
  • cloud groups Users may divide the cloud computing environment into one or more “cloud groups,” where each cloud group may include a group of physical compute nodes (e.g., servers in racks in a data center) that contain similar hypervisor capabilities.
  • a hypervisor also called a virtual machine manager
  • the hypervisor presents to the guest operating systems a virtual operating platform and manages the execution of the guest operating systems.
  • a user may group the compute nodes to form a cloud group by any logic the user chooses. For example, the user may create a cloud group for ESX® hypervisors and create another cloud group for z/VM® hypervisors. In another example, a user may create a cloud group for their development organization and create another cloud group to run production workload.
  • a user manually assigns the compute nodes to form a cloud group, which requires the user to possess an understanding of the cloud computing environment and its composition. For example, a user may want to select compute nodes that reside in different parts of the cloud computing environment to create a cloud group that provides high availability, such as to ensure a prearranged level of operational performance will be met during a contractual measurement period (e.g., having a backup compute node in case one of the compute nodes fails).
  • a contractual measurement period e.g., having a backup compute node in case one of the compute nodes fails.
  • users may not possess such an understanding of the cloud computing environment and its composition.
  • a method for dynamically building a set of compute nodes to host a user's workload comprises receiving workload definitions comprising types of workloads that are to be run in a cloud group as well as a number of instances of each workload the cloud group should support. The method further comprises using the workload definitions to determine virtual machine demands that the cloud group will place on a cloud computing environment. Additionally, the method comprises receiving demand constraints on the cloud group. Furthermore, the method comprises receiving placement constraints on the cloud group. In addition, the method comprises identifying, by a processor, the set of compute nodes to host the user's workload based on the virtual machine demands, the demand constraints and the placement constraints.
  • FIG. 1 illustrates a network system configured in accordance with an embodiment of the present invention
  • FIG. 2 illustrates a cloud computing environment in accordance with an embodiment of the present invention.
  • FIG. 3 illustrates a schematic of racks of compute nodes of the cloud computing node(s) that are managed by an administrative server in accordance with an embodiment of the present invention
  • FIG. 4 illustrates a virtualization environment for a compute node in accordance with an embodiment of the present invention
  • FIG. 5 illustrates a hardware configuration of an administrative server configured in accordance with an embodiment of the present invention.
  • FIG. 6 is a flowchart of a method for dynamically building a set of compute nodes to host the user's workload in accordance with an embodiment of the present invention.
  • FIGS. 7A-7B are a flowchart of a method for monitoring the cloud group and potentially rebalancing the computing resources, if needed, in accordance with an embodiment of the present invention.
  • the present invention comprises a method, system and computer program product for dynamically building a set of compute nodes to host a user's workload.
  • an administrative server receives workload definitions that include the types of workloads (e.g., purchase orders, online banking) that are to be run in a cloud group as well as a number of instances of each workload the cloud group should support. These workload definitions may be used to determine the virtual machine demands that the cloud group will place on the cloud computing environment.
  • the administrative server further receives the demand constraints (constraints being applied to the computing resources used to provide the requested services), placement constraints (constraints on the location of instances of the virtual machines) and license enforcement policies (limitations on the usages of the virtual machines via a software license).
  • the administrative server identifies a set of compute nodes to host the user's workload based on the virtual machines demands, the demand constraints, the placement constraints and the license enforcement policies. In this manner, a set of compute nodes is dynamically built for consideration in forming a cloud group without the user requiring knowledge of the cloud's composition.
  • Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction.
  • This cloud model is composed of five essential characteristics, three service models, and four deployment models.
  • On-Demand Self-Service A consumer can unilaterally provision computing capabilities, such as server time and network storage, as needed, automatically without requiring human interaction with each service's provider.
  • Capabilities are available over a network and accessed through standard mechanisms that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones, tablets, laptops and workstations).
  • heterogeneous thin or thick client platforms e.g., mobile phones, tablets, laptops and workstations.
  • Resource Pooling The provider's computing resources are pooled to serve multiple consumers using a multi-tenant model, with different physical and virtual resources dynamically assigned and reassigned according to consumer demand. There is a sense of location independence in that the consumer generally has no control or knowledge over the exact location of the provided resources but may be able to specify location at a higher level of abstraction (e.g., country, state or data center). Examples of resources include storage, processing, memory and network bandwidth.
  • Rapid Elasticity Capabilities can be elastically provisioned and released, in some cases automatically, to scale rapidly outward and inward commensurate with demand. To the consumer, the capabilities available for provisioning often appear to be unlimited and can be purchased in any quantity at any time.
  • Measured Service Cloud systems automatically control and optimize resource use by leveraging a metering capability at some level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth and active user accounts). Resource usage can be monitored, controlled and reported providing transparency for both the provider and consumer of the utilized service.
  • level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth and active user accounts).
  • SaaS Software as a Service: The capability provided to the consumer is to use the provider's applications running on a cloud infrastructure. The applications are accessible from various client devices through either a thin client interface, such as a web browser (e.g., web-based e-mail) or a program interface. The consumer does not manage or control the underlying cloud infrastructure including network, servers, operating systems, storage, or even individual application capabilities, with the possible exception of limited user-specific application configuration settings.
  • PaaS Platform as a Service
  • the capability provided to the consumer is to deploy onto the cloud infrastructure consumer-created or acquired applications created using programming languages, libraries, services and tools supported by the provider.
  • the consumer does not manage or control the underlying cloud infrastructure including networks, servers, operating systems or storage, but has control over the deployed applications and possibly configuration settings for the application-hosting environment.
  • IaaS Infrastructure as a Service
  • the consumer does not manage or control the underlying cloud infrastructure but has control over operating systems, storage and deployed applications; and possibly limited control of select networking components (e.g., host firewalls).
  • Private Cloud The cloud infrastructure is provisioned for exclusive use by a single organization comprising multiple consumers (e.g., business units). It may be owned, managed and operated by the organization, a third party or some combination of them, and it may exist on or off premises.
  • Public Cloud The cloud infrastructure is provisioned for open use by the general public. It may be owned, managed and operated by a business, academic or government organization, or some combination of them. It exists on the premises of the cloud provider.
  • Hybrid Cloud The cloud infrastructure is a composition of two or more distinct cloud infrastructures (private, community or public) that remain unique entities, but are bound together by standardized or proprietary technology that enables data and application portability (e.g., cloud bursting for load balancing between clouds).
  • FIG. 1 illustrates a network system 100 configured in accordance with an embodiment of the present invention.
  • Network system 100 includes a client device 101 connected to a cloud computing environment 102 via a network 103 .
  • Client device 101 may be any type of computing device (e.g., portable computing unit, personal digital assistant (PDA), smartphone, laptop computer, mobile phone, navigation device, game console, desktop computer system, workstation, Internet appliance and the like) configured with the capability of connecting to cloud computing environment 102 via network 103 .
  • PDA personal digital assistant
  • Network 103 may be, for example, a local area network, a wide area network, a wireless wide area network, a circuit-switched telephone network, a Global System for Mobile Communications (GSM) network, Wireless Application Protocol (WAP) network, a WiFi network, an IEEE 802.11 standards network, various combinations thereof, etc.
  • GSM Global System for Mobile Communications
  • WAP Wireless Application Protocol
  • WiFi Wireless Fidelity
  • IEEE 802.11 standards network
  • Cloud computing environment 102 is used to deliver computing as a service to client device 101 implementing the model discussed above.
  • An embodiment of cloud computing environment 102 is discussed below in connection with FIG. 2 .
  • FIG. 2 illustrates cloud computing environment 102 in accordance with an embodiment of the present invention.
  • cloud computing environment 102 includes one or more cloud computing nodes 201 with which local computing devices used by cloud consumers, such as, for example, personal digital assistant (PDA) or cellular telephone 202 , desktop computer 203 , laptop computer 204 , and/or automobile computer system 205 may communicate.
  • Nodes 201 may communicate with one another. They may be grouped (not shown) physically or virtually, in one or more networks, such as Private, Community, Public, or Hybrid clouds as described hereinabove, or a combination thereof.
  • This allows cloud computing environment 102 to offer infrastructure, platforms and/or software as services for which a cloud consumer does not need to maintain resources on a local computing device.
  • Cloud computing nodes 201 may include one or more racks of compute nodes (e.g., servers) that are managed by a server (referred to herein as the “administrative server”) in cloud computing environment 102 as discussed below in greater detail in connection with FIG. 3 .
  • computing devices 202 , 203 , 204 , 205 shown in FIG. 2 which may represent client device 101 of FIG. 1
  • cloud computing nodes 201 and cloud computing environment 102 can communicate with any type of computerized device over any type of network and/or network addressable connection (e.g., using a web browser).
  • Program code located on one of nodes 201 may be stored on a computer recordable storage medium in one of nodes 201 and downloaded to computing devices 202 , 203 , 204 , 205 over a network for use in these computing devices.
  • a server computer in computing nodes 201 may store program code on a computer readable storage medium on the server computer.
  • the server computer may download the program code to computing device 202 , 203 , 204 , 205 for use on the computing device.
  • FIG. 3 illustrates a schematic of racks of compute nodes (e.g., servers) of cloud computing node(s) 201 that are managed by an administrative server in accordance with an embodiment of the present invention.
  • compute nodes e.g., servers
  • cloud computing node(s) 201 that are managed by an administrative server in accordance with an embodiment of the present invention.
  • cloud computing node(s) 201 may include one or more racks 301 A- 301 C of hardware components or “compute nodes,” such as servers or other electronic devices.
  • rack 301 A houses compute nodes 302 A- 302 E
  • rack 301 B houses compute nodes 302 F- 302 J
  • rack 301 C houses compute nodes 302 K- 302 O.
  • Racks 301 A- 301 C may collectively be referred to as racks 301 or rack 301 , respectively.
  • compute nodes 302 A- 3020 may collectively be referred to as compute nodes 302 or compute node 302 , respectively.
  • FIG. 3 is not to be limited in scope to the number of racks 301 or compute nodes 302 depicted.
  • cloud computing node 201 may be comprised of any number of racks 301 which may house any number of compute nodes 302 .
  • FIG. 3 illustrates racks 301 housing compute nodes 302
  • racks 301 may house any type of computing component that is used by cloud computing node 201 .
  • compute nodes 302 may be distributed across cloud computing environment 102 ( FIGS. 1 and 2 ).
  • the set of compute nodes 302 dynamically built to host the user's workload may include compute nodes 302 that are distributed across cloud computing environment 102 in various cloud computing nodes 201 .
  • racks 301 are each coupled to an administrative server 303 configured to provide data center-level functions.
  • Administrative server 303 supports a module, referred to herein as the management software 304 , that can be used to manage all the compute nodes 302 of cloud computing nodes 201 , monitor system utilization, intelligently deploy images of data and optimize the operations of cloud computing environment 102 .
  • management software 304 can be used to dynamically build a set of compute nodes 302 to be considered to form a cloud group as well as to monitor a cloud group and rebalance system resources, if necessary, as discussed further below.
  • a cloud group refers to a group of compute nodes 302 that are used in combination to run user designated workloads.
  • a description of the hardware configuration of administrative server 303 is provided further below in connection with FIG. 5 .
  • FIG. 4 illustrates a virtualization environment for compute node 302 ( FIG. 3 ) in accordance with an embodiment of the present invention.
  • Compute node 302 includes a virtual operating system 401 .
  • Operating system 401 executes on a real or physical computer 402 .
  • Real computer 402 includes one or more processors 403 , a memory 404 (also referred to herein as the host physical memory), one or more disk drives 405 and the like.
  • Other components of real computer 402 are not discussed herein for the sake of brevity.
  • Virtual operating system 401 further includes user portions 406 A- 406 B (identified as “Guest 1 and Guest 2 ,” respectively, in FIG. 4 ), referred to herein as “guests.”
  • Each guest 406 A, 406 B is capable of functioning as a separate system. That is, each guest 406 A- 406 B can be independently reset, host a guest operating system 407 A- 407 B, respectively, (identified as “Guest 1 O/S” and “Guest 2 O/S,” respectively, in FIG. 4 ) and operate with different programs.
  • An operating system or application program running in guest 406 A, 406 B appears to have access to a full and complete system, but in reality, only a portion of it is available.
  • Guests 406 A- 406 B may collectively or individually be referred to as guests 406 or guest 406 , respectively.
  • Guest operating systems 407 A- 407 B may collectively or individually be referred to as guest operating systems 407 or guest operating system 407 , respectively.
  • Each guest operating system 407 A, 407 B may host one or more virtual machine applications 408 A- 408 C (identified as “VM 1 ,” “VM 2 ” and “VM 3 ,” respectively, in FIG. 4 ), such as JavaTM virtual machines.
  • virtual machine applications 408 A- 408 C such as “VM 1 ,” “VM 2 ” and “VM 3 ,” respectively, in FIG. 4
  • guest operating system 407 A hosts virtual machine applications 408 A- 408 B.
  • Guest operating system 407 B hosts virtual machine application 408 C.
  • Virtual machines 408 A- 408 C may collectively or individually be referred to as virtual machines 408 or virtual machine 408 , respectively.
  • Virtual operating system 401 further includes a common base portion 409 , referred to herein as a hypervisor.
  • Hypervisor 409 may be implemented in microcode running on processor 403 or it may be implemented in software as part of virtual operating system 401 .
  • Hypervisor 409 is configured to manage and enable guests 406 to run on a single host.
  • virtual operating system 401 and its components execute on physical or real computer 402 . These software components may be loaded into memory 404 for execution by processor 403 .
  • the virtualization environment for compute node 302 is not to be limited in scope to the elements depicted in FIG. 4 .
  • the virtualization environment for compute node 302 may include other components that were not discussed herein for the sake of brevity.
  • FIG. 5 illustrates a hardware configuration of administrative server 303 ( FIG. 3 ) which is representative of a hardware environment for practicing the present invention.
  • Administrative server 303 has a processor 501 coupled to various other components by system bus 502 .
  • An operating system 503 runs on processor 501 and provides control and coordinates the functions of the various components of FIG. 5 .
  • An application 504 in accordance with the principles of the present invention runs in conjunction with operating system 503 and provides calls to operating system 503 where the calls implement the various functions or services to be performed by application 504 .
  • Application 504 may include, for example, a program for dynamically building a set of compute nodes 302 ( FIG. 3 ) to host the user's workload as well as monitoring a cloud group and rebalancing system resources, if necessary, as discussed further below in association with FIGS. 6 and 7 A- 7 B.
  • ROM 505 is coupled to system bus 502 and includes a basic input/output system (“BIOS”) that controls certain basic functions of administrative server 303 .
  • RAM random access memory
  • Disk adapter 507 are also coupled to system bus 502 .
  • software components including operating system 503 and application 504 may be loaded into RAM 506 , which may be administrative server's 303 main memory for execution.
  • Disk adapter 507 may be an integrated drive electronics (“IDE”) adapter that communicates with a disk unit 508 , e.g., disk drive.
  • IDE integrated drive electronics
  • the program for dynamically building a set of compute nodes 302 ( FIG. 3 ) to host the user's workload as well as monitoring a cloud group and rebalancing system resources, if necessary, as discussed further below in association with FIGS. 6 and 7 A- 7 B, may reside in disk unit 508 or in application 504 .
  • Administrative server 303 may further include a communications adapter 509 coupled to bus 502 .
  • Communications adapter 509 interconnects bus 502 with an outside network (e.g., network 103 of FIG. 1 ).
  • aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” ‘module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
  • a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
  • a computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus or device.
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
  • Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the C programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the function/acts specified in the flowchart and/or block diagram block or blocks.
  • a user manually assigns the compute nodes to form a cloud group, which requires the user to possess an understanding of the cloud computing environment and its composition. For example, a user may want to select compute nodes that reside in different parts of the cloud computing environment to create a cloud group that provides high availability, such as to ensure a prearranged level of operational performance will be met during a contractual measurement period (e.g., having a backup compute node in case one of the compute nodes fails).
  • a contractual measurement period e.g., having a backup compute node in case one of the compute nodes fails.
  • users may not possess such an understanding of the cloud computing environment and its composition. As a result, users may not be able to optimally select a group of compute nodes to form a cloud group that meets their desired needs, let alone in an efficient manner.
  • FIG. 6 is a flowchart of a method for dynamically building a set of compute nodes to host the user's workload.
  • administrative server 303 FIGS. 3 and 5
  • FIGS. 7A-7B are a flowchart of a method for monitoring the cloud group and potentially rebalancing the computing resources, if needed.
  • FIG. 6 is a flowchart of a method 600 for dynamically building a set of compute nodes 302 ( FIGS. 3 and 4 ) to host the user's workload in accordance with an embodiment of the present invention.
  • administrative server 303 receives workload definitions from the user, which may include the types of workloads (e.g., purchase orders, online banking, serving web pages) that are to be run in the newly created cloud group as well as the number of instances of each workload the cloud group should support.
  • workload definitions may include the types of workloads (e.g., purchase orders, online banking, serving web pages) that are to be run in the newly created cloud group as well as the number of instances of each workload the cloud group should support.
  • step 602 administrative server 303 uses the received workload definitions to determine virtual machine demands (e.g., number of Central Processing Units (CPUs), memory required, ingress/egress bandwidth required for data and storage networks) that the cloud group will place on cloud computing environment 102 .
  • virtual machine demands e.g., number of Central Processing Units (CPUs), memory required, ingress/egress bandwidth required for data and storage networks
  • step 603 administrative server 303 receives the demand constraints on the cloud group to be created.
  • Demand constraints refer to user specified constraints being applied to computing resources used to provide the requested services.
  • demand constraints may include processor, memory, storage, network I/O, storage I/O and bandwidth constraints that each virtual machine 408 of the virtual machines 408 deployed in the cloud group requires.
  • placement constraints refer to user specified constraints on the location of instances of virtual machines 408 .
  • placement constraints may include deployment policies, such as high availability and energy conservation.
  • High availability refers to ensuring a prearranged level of operational performance will be met during a contractual measurement period (e.g., having a backup compute node 302 in case one of the compute nodes 302 fails).
  • a high availability policy may spread the workloads across compute nodes 302 .
  • An energy conservation policy may attempt to place the workloads on the same physical compute node 302 so as to conserve energy use.
  • Placement constraints may further include consolidation.
  • a user may specify a constraint that a particular cloud group needs physical separation of their workloads.
  • administrative server 303 would ensure that workloads for that cloud group are only placed on compute nodes 302 on which no workloads from other cloud groups would run.
  • a license enforcement policy refers to limitations on the usages of virtual machines 408 via a software license, such as in a service-level agreement. For example, due to license enforcement policies, it may be preferable to place a virtual machine 408 on a compute node 302 that has a higher usage than another compute node 302 . In another example, due to the way licenses are counted, it may be possible to run additional instances of a virtual machine 408 on compute node 302 without additional license charges, where the licenses are counted based on the number of compute nodes 302 and not on the number of virtual machines 408 running on compute nodes 302 .
  • step 606 administrative server 303 identifies a set of compute nodes 302 to be considered for forming a cloud group based on the determined virtual machine demands, demand constraints, placement constraints and license enforcement policies discussed above while being able to host the user's workload. That is, administrative server 303 dynamically builds a set of compute nodes 302 to be considered for forming a cloud group that satisfies a “pattern.”
  • a pattern may refer to the user's workload requirements, demand constraints, placement constraints and license enforcement policies.
  • step 607 administrative server 303 deploys virtual machines 408 to the identified set of compute nodes 302 in connection with meeting the determined virtual machine demands, demand constraints, placement constraints and license enforcement policies while being able to host the user's workload.
  • step 608 administrative server 303 deploys the workloads to the identified set of compute nodes 302 and starts a new instance of the workload on each appropriate compute node 302 in the identified set of compute nodes 302 .
  • a set of compute nodes 302 is dynamically built for consideration in forming a cloud group without the user requiring knowledge of the cloud's composition.
  • method 600 may include other and/or additional steps that, for clarity, are not depicted. Further, in some implementations, method 600 may be executed in a different order presented and that the order presented in the discussion of FIG. 6 is illustrative. Additionally, in some implementations, certain steps in method 600 may be executed in a substantially simultaneous manner or may be omitted.
  • administrative server 303 monitors the cloud group to determine if the computing resources (e.g., virtual machines 408 ) need to be rebalanced as discussed further below in connection with FIGS. 7A-7B .
  • FIGS. 7A-7B are a flowchart of a method 700 for monitoring the cloud group and rebalancing the computing resources, if necessary, after the deployment of virtual machines 408 to the set of compute nodes 302 in accordance with an embodiment of the present invention.
  • step 701 administrative server 303 monitors for changes in the workload demands as well as for changes in the demand constraints and placement constraints after the deployment of virtual machines 408 .
  • administrative server 303 monitors for hardware failures and for predicted hardware failures. For example, administrative server 303 may treat a large number of retransmits or bad packets sent over a network port as indicating a problem with that network connection. In another example, administrative server 303 may treat the failure of more than one dual in-line memory module as indicating that additional dual in-line memory modules are likely to fail. In one embodiment, administrative server 303 treats the “predicted” failure as though it has already occurred (i.e., administrative server 303 treats a detected failure the same as a predicted failure).
  • step 703 administrative server 303 monitors demands on the deployed virtual machines 408 .
  • administrative server 303 monitors the set of compute nodes 302 to ensure that all virtual machines 408 deployed in the set of compute nodes 302 are receiving their required demands (e.g., meeting processor, memory, storage, network I/O and storage I/O requirements).
  • step 704 administrative server 303 monitors the cloud group for hardware usage. For example, if multiple servers are not being fully utilized, then it may be desirable to remove those servers from the cloud group.
  • step 705 administrative server 303 monitors the cloud group for hardware (e.g., servers, switches) additions/subtractions. For example, a new server may be added to the cloud group.
  • hardware e.g., servers, switches
  • step 706 administrative server 303 monitors the cloud group for license usage.
  • step 707 a determination is made by administrative server 303 as to whether workload demands and/or demand constraints and/or placement constraints have changed after the deployment of virtual machines 408 .
  • step 708 administrative server 303 rebalances the computing resources used by the cloud group accordingly.
  • Rebalancing may refer to relocating one or more of virtual machines 408 to other compute nodes 302 of the cloud group. Such relocating may include what is referred to herein as “chained relocations,” which involve other required relocations in order to relocate a single virtual machine 408 . For example, in order to move virtual machine 408 A on compute node 302 A to compute node 302 B, virtual machine 408 B on compute node 302 A may need to be moved to compute node 302 C.
  • Rebalancing may also refer to adding one or more virtual machines 408 to the set of computing nodes 302 . Additionally, rebalancing may refer to relocating one or more virtual machines 408 in the cloud group to other computing nodes 302 outside of the cloud group. “Rebalancing,” as used herein, refers to any of the examples listed above, including a combination of them, in order to rebalance the computing resources so as to meet the required workload demands and constraints.
  • administrative server 303 determines whether it detected any hardware failures or detected any predicted hardware failures. If administrative server 303 detected any hardware failures or detected any predicted hardware failures, then, in step 710 , administrative server 303 determines if the remaining hardware (e.g., computing nodes 302 ) in the cloud group are sufficient to handle the workloads that use the cloud group.
  • the remaining hardware e.g., computing nodes 302
  • step 708 administrative server 303 rebalances the computing resources used by the cloud group accordingly.
  • step 711 administrative server 303 determines if virtual machines 408 in the deployed virtual machines 408 are not receiving their required demands. If virtual machines 408 in the deployed virtual machines 408 are not receiving their required demands, then, in step 708 , administrative server 303 rebalances the computing resources used by the cloud group accordingly.
  • step 712 administrative server 303 determines if computing resources used by the cloud group need to be rebalanced because the hardware is not being fully utilized. If the hardware is not being fully utilized, then, in step 708 , administrative server 303 rebalances the computing resources used by the cloud group accordingly.
  • administrative server 303 determines if there were any hardware (e.g., switches, servers) additions/subtractions in the cloud group. If there were hardware additions/subtractions in the cloud group, then, in step 708 , administrative server 303 rebalances the computing resources used by the cloud group accordingly.
  • hardware e.g., switches, servers
  • step 714 administrative server 303 determines if there are any available licenses due to a virtual machine 408 that was previously consuming a license (i.e., previously counted as using a license) no longer having a need to consume the license. If there are available licenses, then, in step 708 , administrative server 303 rebalances the computing resources used by the cloud group accordingly.
  • administrative server 303 continues to monitor for changes in the workload demands as well as for changes in the demand constraints and placement constraints after the deployment of virtual machines 408 in step 701 .
  • method 700 may include other and/or additional steps that, for clarity, are not depicted. Further, in some implementations, method 700 may be executed in a different order presented and that the order presented in the discussion of FIGS. 7A-7B is illustrative. Additionally, in some implementations, certain steps in method 700 may be executed in a substantially simultaneous manner or may be omitted.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Debugging And Monitoring (AREA)

Abstract

A method, system and computer program product for dynamically building a set of compute nodes to host a user's workload. An administrative server receives workload definitions that include the types of workloads that are to be run in a cloud group as well as a number of instances of each workload the cloud group should support. These workload definitions are used to determine the virtual machine demands that the cloud group will place on the cloud environment. The administrative server further receives the demand constraints, placement constraints and license enforcement policies. The administrative server identifies a set of compute nodes to host the user's workload based on the virtual machines demands, the demand constraints, the placement constraints and the license enforcement policies. In this manner, a set of compute nodes is dynamically built for consideration in forming a cloud group without the user requiring knowledge of the cloud's composition.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
The present application is a continuation application of pending U.S. patent application Ser. No. 13/356,427, which was filed on Jan. 23, 2012, which is assigned to the assignee of the present invention. The present application claims priority benefits to U.S. patent application Ser. No. 13/356,427.
TECHNICAL FIELD
The present invention relates to cloud computing, and more particularly to dynamically building a set of compute nodes to host a user's workload.
BACKGROUND
In a cloud computing environment, computing is delivered as a service rather than a product, whereby shared resources, software and information are provided to computers and other devices as a metered service over a network, such as the Internet. In such an environment, computation, software, data access and storage services are provided to users that do not require knowledge of the physical location and configuration of the system that delivers the services.
Users may divide the cloud computing environment into one or more “cloud groups,” where each cloud group may include a group of physical compute nodes (e.g., servers in racks in a data center) that contain similar hypervisor capabilities. A hypervisor (also called a virtual machine manager) is one of many virtualization techniques allowing multiple operating systems, termed guests, to run concurrently on a host computer. The hypervisor presents to the guest operating systems a virtual operating platform and manages the execution of the guest operating systems.
A user may group the compute nodes to form a cloud group by any logic the user chooses. For example, the user may create a cloud group for ESX® hypervisors and create another cloud group for z/VM® hypervisors. In another example, a user may create a cloud group for their development organization and create another cloud group to run production workload.
Currently, a user manually assigns the compute nodes to form a cloud group, which requires the user to possess an understanding of the cloud computing environment and its composition. For example, a user may want to select compute nodes that reside in different parts of the cloud computing environment to create a cloud group that provides high availability, such as to ensure a prearranged level of operational performance will be met during a contractual measurement period (e.g., having a backup compute node in case one of the compute nodes fails). However, users may not possess such an understanding of the cloud computing environment and its composition.
As a result, users may not be able to optimally select a group of compute nodes to form a cloud group that meets their desired needs, let alone in an efficient manner.
BRIEF SUMMARY
In one embodiment of the present invention, a method for dynamically building a set of compute nodes to host a user's workload comprises receiving workload definitions comprising types of workloads that are to be run in a cloud group as well as a number of instances of each workload the cloud group should support. The method further comprises using the workload definitions to determine virtual machine demands that the cloud group will place on a cloud computing environment. Additionally, the method comprises receiving demand constraints on the cloud group. Furthermore, the method comprises receiving placement constraints on the cloud group. In addition, the method comprises identifying, by a processor, the set of compute nodes to host the user's workload based on the virtual machine demands, the demand constraints and the placement constraints.
Other forms of the embodiment of the method described above are in a system and in a computer program product.
The foregoing has outlined rather generally the features and technical advantages of one or more embodiments of the present invention in order that the detailed description of the present invention that follows may be better understood. Additional features and advantages of the present invention will be described hereinafter which may form the subject of the claims of the present invention.
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
A better understanding of the present invention can be obtained when the following detailed description is considered in conjunction with the following drawings, in which:
FIG. 1 illustrates a network system configured in accordance with an embodiment of the present invention;
FIG. 2 illustrates a cloud computing environment in accordance with an embodiment of the present invention.
FIG. 3 illustrates a schematic of racks of compute nodes of the cloud computing node(s) that are managed by an administrative server in accordance with an embodiment of the present invention;
FIG. 4 illustrates a virtualization environment for a compute node in accordance with an embodiment of the present invention;
FIG. 5 illustrates a hardware configuration of an administrative server configured in accordance with an embodiment of the present invention; and
FIG. 6 is a flowchart of a method for dynamically building a set of compute nodes to host the user's workload in accordance with an embodiment of the present invention; and
FIGS. 7A-7B are a flowchart of a method for monitoring the cloud group and potentially rebalancing the computing resources, if needed, in accordance with an embodiment of the present invention.
DETAILED DESCRIPTION
The present invention comprises a method, system and computer program product for dynamically building a set of compute nodes to host a user's workload. In one embodiment of the present invention, an administrative server receives workload definitions that include the types of workloads (e.g., purchase orders, online banking) that are to be run in a cloud group as well as a number of instances of each workload the cloud group should support. These workload definitions may be used to determine the virtual machine demands that the cloud group will place on the cloud computing environment. The administrative server further receives the demand constraints (constraints being applied to the computing resources used to provide the requested services), placement constraints (constraints on the location of instances of the virtual machines) and license enforcement policies (limitations on the usages of the virtual machines via a software license). The administrative server identifies a set of compute nodes to host the user's workload based on the virtual machines demands, the demand constraints, the placement constraints and the license enforcement policies. In this manner, a set of compute nodes is dynamically built for consideration in forming a cloud group without the user requiring knowledge of the cloud's composition.
In the following description, numerous specific details are set forth to provide a thorough understanding of the present invention. However, it will be apparent to those skilled in the art that the present invention may be practiced without such specific details. In other instances, well-known circuits have been shown in block diagram form in order not to obscure the present invention in unnecessary detail. For the most part, details considering timing considerations and the like have been omitted inasmuch as such details are not necessary to obtain a complete understanding of the present invention and are within the skills of persons of ordinary skill in the relevant art.
It is understood in advance that although this disclosure includes a detailed description on cloud computing, implementation of the teachings recited herein are not limited to a cloud computing environment. Rather, the embodiments of the present invention are capable of being implemented in conjunction with any type of clustered computing environment now known or later developed.
In any event, the following definitions have been derived from the “The NIST Definition of Cloud Computing” by Peter Mell and Timothy Grance, dated September 2011, which is cited on an Information Disclosure Statement filed herewith, and a copy of which is provided to the U.S. Patent and Trademark Office.
Cloud computing is a model for enabling ubiquitous, convenient, on-demand network access to a shared pool of configurable computing resources (e.g., networks, servers, storage, applications, and services) that can be rapidly provisioned and released with minimal management effort or service provider interaction. This cloud model is composed of five essential characteristics, three service models, and four deployment models.
Characteristics are as follows:
On-Demand Self-Service: A consumer can unilaterally provision computing capabilities, such as server time and network storage, as needed, automatically without requiring human interaction with each service's provider.
Broad Network Access: Capabilities are available over a network and accessed through standard mechanisms that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones, tablets, laptops and workstations).
Resource Pooling: The provider's computing resources are pooled to serve multiple consumers using a multi-tenant model, with different physical and virtual resources dynamically assigned and reassigned according to consumer demand. There is a sense of location independence in that the consumer generally has no control or knowledge over the exact location of the provided resources but may be able to specify location at a higher level of abstraction (e.g., country, state or data center). Examples of resources include storage, processing, memory and network bandwidth.
Rapid Elasticity: Capabilities can be elastically provisioned and released, in some cases automatically, to scale rapidly outward and inward commensurate with demand. To the consumer, the capabilities available for provisioning often appear to be unlimited and can be purchased in any quantity at any time.
Measured Service: Cloud systems automatically control and optimize resource use by leveraging a metering capability at some level of abstraction appropriate to the type of service (e.g., storage, processing, bandwidth and active user accounts). Resource usage can be monitored, controlled and reported providing transparency for both the provider and consumer of the utilized service.
Service Models are as follows:
Software as a Service (SaaS): The capability provided to the consumer is to use the provider's applications running on a cloud infrastructure. The applications are accessible from various client devices through either a thin client interface, such as a web browser (e.g., web-based e-mail) or a program interface. The consumer does not manage or control the underlying cloud infrastructure including network, servers, operating systems, storage, or even individual application capabilities, with the possible exception of limited user-specific application configuration settings.
Platform as a Service (PaaS): The capability provided to the consumer is to deploy onto the cloud infrastructure consumer-created or acquired applications created using programming languages, libraries, services and tools supported by the provider. The consumer does not manage or control the underlying cloud infrastructure including networks, servers, operating systems or storage, but has control over the deployed applications and possibly configuration settings for the application-hosting environment.
Infrastructure as a Service (IaaS): The capability provided to the consumer is to provision processing, storage, networks and other fundamental computing resources where the consumer is able to deploy and run arbitrary software, which can include operating systems and applications. The consumer does not manage or control the underlying cloud infrastructure but has control over operating systems, storage and deployed applications; and possibly limited control of select networking components (e.g., host firewalls).
Deployment Models are as follows:
Private Cloud: The cloud infrastructure is provisioned for exclusive use by a single organization comprising multiple consumers (e.g., business units). It may be owned, managed and operated by the organization, a third party or some combination of them, and it may exist on or off premises.
Community Cloud: The cloud infrastructure is provisioned for exclusive use by a specific community of consumers from organizations that have shared concerns (e.g., mission, security requirements, policy and compliance considerations). It may be owned, managed and operated by one or more of the organizations in the community, a third party, or some combination of them, and it may exist on or off premises.
Public Cloud: The cloud infrastructure is provisioned for open use by the general public. It may be owned, managed and operated by a business, academic or government organization, or some combination of them. It exists on the premises of the cloud provider.
Hybrid Cloud: The cloud infrastructure is a composition of two or more distinct cloud infrastructures (private, community or public) that remain unique entities, but are bound together by standardized or proprietary technology that enables data and application portability (e.g., cloud bursting for load balancing between clouds).
Referring now to the Figures in detail, FIG. 1 illustrates a network system 100 configured in accordance with an embodiment of the present invention. Network system 100 includes a client device 101 connected to a cloud computing environment 102 via a network 103. Client device 101 may be any type of computing device (e.g., portable computing unit, personal digital assistant (PDA), smartphone, laptop computer, mobile phone, navigation device, game console, desktop computer system, workstation, Internet appliance and the like) configured with the capability of connecting to cloud computing environment 102 via network 103.
Network 103 may be, for example, a local area network, a wide area network, a wireless wide area network, a circuit-switched telephone network, a Global System for Mobile Communications (GSM) network, Wireless Application Protocol (WAP) network, a WiFi network, an IEEE 802.11 standards network, various combinations thereof, etc. Other networks, whose descriptions are omitted here for brevity, may also be used in conjunction with system 100 of FIG. 1 without departing from the scope of the present invention.
Cloud computing environment 102 is used to deliver computing as a service to client device 101 implementing the model discussed above. An embodiment of cloud computing environment 102 is discussed below in connection with FIG. 2.
FIG. 2 illustrates cloud computing environment 102 in accordance with an embodiment of the present invention. As shown, cloud computing environment 102 includes one or more cloud computing nodes 201 with which local computing devices used by cloud consumers, such as, for example, personal digital assistant (PDA) or cellular telephone 202, desktop computer 203, laptop computer 204, and/or automobile computer system 205 may communicate. Nodes 201 may communicate with one another. They may be grouped (not shown) physically or virtually, in one or more networks, such as Private, Community, Public, or Hybrid clouds as described hereinabove, or a combination thereof. This allows cloud computing environment 102 to offer infrastructure, platforms and/or software as services for which a cloud consumer does not need to maintain resources on a local computing device. Cloud computing nodes 201 may include one or more racks of compute nodes (e.g., servers) that are managed by a server (referred to herein as the “administrative server”) in cloud computing environment 102 as discussed below in greater detail in connection with FIG. 3.
It is understood that the types of computing devices 202, 203, 204, 205 shown in FIG. 2, which may represent client device 101 of FIG. 1, are intended to be illustrative and that cloud computing nodes 201 and cloud computing environment 102 can communicate with any type of computerized device over any type of network and/or network addressable connection (e.g., using a web browser). Program code located on one of nodes 201 may be stored on a computer recordable storage medium in one of nodes 201 and downloaded to computing devices 202, 203, 204, 205 over a network for use in these computing devices. For example, a server computer in computing nodes 201 may store program code on a computer readable storage medium on the server computer. The server computer may download the program code to computing device 202, 203, 204, 205 for use on the computing device.
Referring now to FIG. 3, FIG. 3 illustrates a schematic of racks of compute nodes (e.g., servers) of cloud computing node(s) 201 that are managed by an administrative server in accordance with an embodiment of the present invention.
As shown, cloud computing node(s) 201 (also may be referred to as a “cluster”) may include one or more racks 301A-301C of hardware components or “compute nodes,” such as servers or other electronic devices. For example, rack 301A houses compute nodes 302A-302E, rack 301B houses compute nodes 302F-302J and rack 301C houses compute nodes 302K-302O. Racks 301A-301C may collectively be referred to as racks 301 or rack 301, respectively. Furthermore, compute nodes 302A-3020 may collectively be referred to as compute nodes 302 or compute node 302, respectively. An illustrative virtualization environment for compute node 302 is discussed further below in connection with FIG. 4. FIG. 3 is not to be limited in scope to the number of racks 301 or compute nodes 302 depicted. For example, cloud computing node 201 may be comprised of any number of racks 301 which may house any number of compute nodes 302. Furthermore, while FIG. 3 illustrates racks 301 housing compute nodes 302, racks 301 may house any type of computing component that is used by cloud computing node 201. Furthermore, while the following discusses compute nodes 302 being confined in designated racks 301, it is noted for clarity that compute nodes 302 may be distributed across cloud computing environment 102 (FIGS. 1 and 2). It is further noted that the set of compute nodes 302 dynamically built to host the user's workload, as discussed further herein, may include compute nodes 302 that are distributed across cloud computing environment 102 in various cloud computing nodes 201.
As further shown in FIG. 3, racks 301 are each coupled to an administrative server 303 configured to provide data center-level functions. Administrative server 303 supports a module, referred to herein as the management software 304, that can be used to manage all the compute nodes 302 of cloud computing nodes 201, monitor system utilization, intelligently deploy images of data and optimize the operations of cloud computing environment 102. Furthermore, management software 304 can be used to dynamically build a set of compute nodes 302 to be considered to form a cloud group as well as to monitor a cloud group and rebalance system resources, if necessary, as discussed further below. A cloud group refers to a group of compute nodes 302 that are used in combination to run user designated workloads. A description of the hardware configuration of administrative server 303 is provided further below in connection with FIG. 5.
Referring now to FIG. 4, FIG. 4 illustrates a virtualization environment for compute node 302 (FIG. 3) in accordance with an embodiment of the present invention. Compute node 302 includes a virtual operating system 401. Operating system 401 executes on a real or physical computer 402. Real computer 402 includes one or more processors 403, a memory 404 (also referred to herein as the host physical memory), one or more disk drives 405 and the like. Other components of real computer 402 are not discussed herein for the sake of brevity.
Virtual operating system 401 further includes user portions 406A-406B (identified as “Guest 1 and Guest 2,” respectively, in FIG. 4), referred to herein as “guests.” Each guest 406A, 406B is capable of functioning as a separate system. That is, each guest 406A-406B can be independently reset, host a guest operating system 407A-407B, respectively, (identified as “Guest 1 O/S” and “Guest 2 O/S,” respectively, in FIG. 4) and operate with different programs. An operating system or application program running in guest 406A, 406B appears to have access to a full and complete system, but in reality, only a portion of it is available. Guests 406A-406B may collectively or individually be referred to as guests 406 or guest 406, respectively. Guest operating systems 407A-407B may collectively or individually be referred to as guest operating systems 407 or guest operating system 407, respectively.
Each guest operating system 407A, 407B may host one or more virtual machine applications 408A-408C (identified as “VM 1,” “VM 2” and “VM 3,” respectively, in FIG. 4), such as Java™ virtual machines. For example, guest operating system 407A hosts virtual machine applications 408A-408B. Guest operating system 407B hosts virtual machine application 408C. Virtual machines 408A-408C may collectively or individually be referred to as virtual machines 408 or virtual machine 408, respectively.
Virtual operating system 401 further includes a common base portion 409, referred to herein as a hypervisor. Hypervisor 409 may be implemented in microcode running on processor 403 or it may be implemented in software as part of virtual operating system 401. Hypervisor 409 is configured to manage and enable guests 406 to run on a single host.
As discussed above, virtual operating system 401 and its components execute on physical or real computer 402. These software components may be loaded into memory 404 for execution by processor 403.
The virtualization environment for compute node 302 is not to be limited in scope to the elements depicted in FIG. 4. The virtualization environment for compute node 302 may include other components that were not discussed herein for the sake of brevity.
Referring now to FIG. 5, FIG. 5 illustrates a hardware configuration of administrative server 303 (FIG. 3) which is representative of a hardware environment for practicing the present invention. Administrative server 303 has a processor 501 coupled to various other components by system bus 502. An operating system 503 runs on processor 501 and provides control and coordinates the functions of the various components of FIG. 5. An application 504 in accordance with the principles of the present invention runs in conjunction with operating system 503 and provides calls to operating system 503 where the calls implement the various functions or services to be performed by application 504. Application 504 may include, for example, a program for dynamically building a set of compute nodes 302 (FIG. 3) to host the user's workload as well as monitoring a cloud group and rebalancing system resources, if necessary, as discussed further below in association with FIGS. 6 and 7A-7B.
Referring again to FIG. 5, read-only memory (“ROM”) 505 is coupled to system bus 502 and includes a basic input/output system (“BIOS”) that controls certain basic functions of administrative server 303. Random access memory (“RAM”) 506 and disk adapter 507 are also coupled to system bus 502. It should be noted that software components including operating system 503 and application 504 may be loaded into RAM 506, which may be administrative server's 303 main memory for execution. Disk adapter 507 may be an integrated drive electronics (“IDE”) adapter that communicates with a disk unit 508, e.g., disk drive. It is noted that the program for dynamically building a set of compute nodes 302 (FIG. 3) to host the user's workload as well as monitoring a cloud group and rebalancing system resources, if necessary, as discussed further below in association with FIGS. 6 and 7A-7B, may reside in disk unit 508 or in application 504.
Administrative server 303 may further include a communications adapter 509 coupled to bus 502. Communications adapter 509 interconnects bus 502 with an outside network (e.g., network 103 of FIG. 1).
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” ‘module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus or device.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the C programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the present invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the function/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the function/acts specified in the flowchart and/or block diagram block or blocks.
As stated in the Background section, currently, a user manually assigns the compute nodes to form a cloud group, which requires the user to possess an understanding of the cloud computing environment and its composition. For example, a user may want to select compute nodes that reside in different parts of the cloud computing environment to create a cloud group that provides high availability, such as to ensure a prearranged level of operational performance will be met during a contractual measurement period (e.g., having a backup compute node in case one of the compute nodes fails). However, users may not possess such an understanding of the cloud computing environment and its composition. As a result, users may not be able to optimally select a group of compute nodes to form a cloud group that meets their desired needs, let alone in an efficient manner.
The principles of the present invention provide a means for dynamically building a set of compute nodes to host the user's workload based on the user's workload requirements, the demand and placement constraints as well as the license enforcement policies without the user requiring knowledge of the cloud's composition as discussed further below in connection FIG. 6. FIG. 6 is a flowchart of a method for dynamically building a set of compute nodes to host the user's workload. After dynamically building the set of compute nodes to host the user's workload, administrative server 303 (FIGS. 3 and 5) monitors the cloud group and can rebalance the computing resources accordingly, if necessary, as discussed further below in connection with FIGS. 7A-7B. FIGS. 7A-7B are a flowchart of a method for monitoring the cloud group and potentially rebalancing the computing resources, if needed.
Referring to FIG. 6, FIG. 6 is a flowchart of a method 600 for dynamically building a set of compute nodes 302 (FIGS. 3 and 4) to host the user's workload in accordance with an embodiment of the present invention.
Referring to FIG. 6, in conjunction with FIGS. 1-5, in step 601, administrative server 303 receives workload definitions from the user, which may include the types of workloads (e.g., purchase orders, online banking, serving web pages) that are to be run in the newly created cloud group as well as the number of instances of each workload the cloud group should support.
In step 602, administrative server 303 uses the received workload definitions to determine virtual machine demands (e.g., number of Central Processing Units (CPUs), memory required, ingress/egress bandwidth required for data and storage networks) that the cloud group will place on cloud computing environment 102.
In step 603, administrative server 303 receives the demand constraints on the cloud group to be created. Demand constraints refer to user specified constraints being applied to computing resources used to provide the requested services. For example, demand constraints may include processor, memory, storage, network I/O, storage I/O and bandwidth constraints that each virtual machine 408 of the virtual machines 408 deployed in the cloud group requires.
In step 604, administrative server 303 receives the placement constraints on the cloud group to be created. Placement constraints refer to user specified constraints on the location of instances of virtual machines 408. For instance, placement constraints may include deployment policies, such as high availability and energy conservation. High availability refers to ensuring a prearranged level of operational performance will be met during a contractual measurement period (e.g., having a backup compute node 302 in case one of the compute nodes 302 fails). As a result, a high availability policy may spread the workloads across compute nodes 302. An energy conservation policy, on the other hand, may attempt to place the workloads on the same physical compute node 302 so as to conserve energy use. Placement constraints may further include consolidation. In another example of placement constraints, a user may specify a constraint that a particular cloud group needs physical separation of their workloads. As a result, administrative server 303 would ensure that workloads for that cloud group are only placed on compute nodes 302 on which no workloads from other cloud groups would run.
In step 605, administrative server 303 receives license enforcement policies. A license enforcement policy refers to limitations on the usages of virtual machines 408 via a software license, such as in a service-level agreement. For example, due to license enforcement policies, it may be preferable to place a virtual machine 408 on a compute node 302 that has a higher usage than another compute node 302. In another example, due to the way licenses are counted, it may be possible to run additional instances of a virtual machine 408 on compute node 302 without additional license charges, where the licenses are counted based on the number of compute nodes 302 and not on the number of virtual machines 408 running on compute nodes 302.
In step 606, administrative server 303 identifies a set of compute nodes 302 to be considered for forming a cloud group based on the determined virtual machine demands, demand constraints, placement constraints and license enforcement policies discussed above while being able to host the user's workload. That is, administrative server 303 dynamically builds a set of compute nodes 302 to be considered for forming a cloud group that satisfies a “pattern.” A pattern may refer to the user's workload requirements, demand constraints, placement constraints and license enforcement policies.
In step 607, administrative server 303 deploys virtual machines 408 to the identified set of compute nodes 302 in connection with meeting the determined virtual machine demands, demand constraints, placement constraints and license enforcement policies while being able to host the user's workload.
In step 608, administrative server 303 deploys the workloads to the identified set of compute nodes 302 and starts a new instance of the workload on each appropriate compute node 302 in the identified set of compute nodes 302.
In this manner, a set of compute nodes 302 is dynamically built for consideration in forming a cloud group without the user requiring knowledge of the cloud's composition.
In some implementations, method 600 may include other and/or additional steps that, for clarity, are not depicted. Further, in some implementations, method 600 may be executed in a different order presented and that the order presented in the discussion of FIG. 6 is illustrative. Additionally, in some implementations, certain steps in method 600 may be executed in a substantially simultaneous manner or may be omitted.
After dynamically building the set of compute nodes 302 for hosting the user's workload, administrative server 303 monitors the cloud group to determine if the computing resources (e.g., virtual machines 408) need to be rebalanced as discussed further below in connection with FIGS. 7A-7B.
FIGS. 7A-7B are a flowchart of a method 700 for monitoring the cloud group and rebalancing the computing resources, if necessary, after the deployment of virtual machines 408 to the set of compute nodes 302 in accordance with an embodiment of the present invention.
Referring to FIG. 7A, in conjunction with FIGS. 1-5, in step 701, administrative server 303 monitors for changes in the workload demands as well as for changes in the demand constraints and placement constraints after the deployment of virtual machines 408.
In step 702, administrative server 303 monitors for hardware failures and for predicted hardware failures. For example, administrative server 303 may treat a large number of retransmits or bad packets sent over a network port as indicating a problem with that network connection. In another example, administrative server 303 may treat the failure of more than one dual in-line memory module as indicating that additional dual in-line memory modules are likely to fail. In one embodiment, administrative server 303 treats the “predicted” failure as though it has already occurred (i.e., administrative server 303 treats a detected failure the same as a predicted failure).
In step 703, administrative server 303 monitors demands on the deployed virtual machines 408. For example, administrative server 303 monitors the set of compute nodes 302 to ensure that all virtual machines 408 deployed in the set of compute nodes 302 are receiving their required demands (e.g., meeting processor, memory, storage, network I/O and storage I/O requirements).
In step 704, administrative server 303 monitors the cloud group for hardware usage. For example, if multiple servers are not being fully utilized, then it may be desirable to remove those servers from the cloud group.
In step 705, administrative server 303 monitors the cloud group for hardware (e.g., servers, switches) additions/subtractions. For example, a new server may be added to the cloud group.
In step 706, administrative server 303 monitors the cloud group for license usage.
In step 707, a determination is made by administrative server 303 as to whether workload demands and/or demand constraints and/or placement constraints have changed after the deployment of virtual machines 408.
If there has been a change in the workload demands and/or demand constraints and/or placement constraints, then, in step 708, administrative server 303 rebalances the computing resources used by the cloud group accordingly. Rebalancing may refer to relocating one or more of virtual machines 408 to other compute nodes 302 of the cloud group. Such relocating may include what is referred to herein as “chained relocations,” which involve other required relocations in order to relocate a single virtual machine 408. For example, in order to move virtual machine 408A on compute node 302A to compute node 302B, virtual machine 408B on compute node 302A may need to be moved to compute node 302C. Rebalancing may also refer to adding one or more virtual machines 408 to the set of computing nodes 302. Additionally, rebalancing may refer to relocating one or more virtual machines 408 in the cloud group to other computing nodes 302 outside of the cloud group. “Rebalancing,” as used herein, refers to any of the examples listed above, including a combination of them, in order to rebalance the computing resources so as to meet the required workload demands and constraints.
If, however, there has not been a change in the workload demands and/or demand constraints and/or placement constraints, then, in step 709, administrative server 303 determines whether it detected any hardware failures or detected any predicted hardware failures. If administrative server 303 detected any hardware failures or detected any predicted hardware failures, then, in step 710, administrative server 303 determines if the remaining hardware (e.g., computing nodes 302) in the cloud group are sufficient to handle the workloads that use the cloud group.
If the remaining hardware in the cloud group is not sufficient to handle the workloads that use the cloud group, then, in step 708, administrative server 303 rebalances the computing resources used by the cloud group accordingly.
Referring to FIG. 7B, in conjunction with FIGS. 1-5, if, however, administrative server 303 did not detect a hardware failure or detect a predicted hardware failure, or if the remaining hardware in the cloud group is sufficient to handle the workloads that use the cloud group, then, in step 711, administrative server 303 determines if virtual machines 408 in the deployed virtual machines 408 are not receiving their required demands. If virtual machines 408 in the deployed virtual machines 408 are not receiving their required demands, then, in step 708, administrative server 303 rebalances the computing resources used by the cloud group accordingly.
If, however, virtual machines 408 in the deployed virtual machines 408 are receiving their required demands, then, in step 712, administrative server 303 determines if computing resources used by the cloud group need to be rebalanced because the hardware is not being fully utilized. If the hardware is not being fully utilized, then, in step 708, administrative server 303 rebalances the computing resources used by the cloud group accordingly.
If, however, hardware is being fully utilized, then, in step 713, administrative server 303 determines if there were any hardware (e.g., switches, servers) additions/subtractions in the cloud group. If there were hardware additions/subtractions in the cloud group, then, in step 708, administrative server 303 rebalances the computing resources used by the cloud group accordingly.
If, however, there were no hardware additions/subtractions in the cloud group, then, in step 714, administrative server 303 determines if there are any available licenses due to a virtual machine 408 that was previously consuming a license (i.e., previously counted as using a license) no longer having a need to consume the license. If there are available licenses, then, in step 708, administrative server 303 rebalances the computing resources used by the cloud group accordingly.
If, however, there are no available licenses, then administrative server 303 continues to monitor for changes in the workload demands as well as for changes in the demand constraints and placement constraints after the deployment of virtual machines 408 in step 701.
In some implementations, method 700 may include other and/or additional steps that, for clarity, are not depicted. Further, in some implementations, method 700 may be executed in a different order presented and that the order presented in the discussion of FIGS. 7A-7B is illustrative. Additionally, in some implementations, certain steps in method 700 may be executed in a substantially simultaneous manner or may be omitted.
The descriptions of the various embodiments of the present invention have been presented for purposes of illustration, but are not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement over technologies found in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.

Claims (8)

The invention claimed is:
1. A method for dynamically building a set of compute nodes to host a user's workload, the method comprising:
receiving workload definitions comprising types of workloads that are to be run in a cloud group as well as a number of instances of each workload said cloud group should support;
using said workload definitions to determine virtual machine demands that said cloud group will place on a cloud computing environment;
receiving demand constraints on said cloud group;
receiving placement constraints on said cloud group; and
identifying, by a processor, said set of compute nodes to host said user's workload based on said virtual machine demands, said demand constraints and said placement constraints.
2. The method as recited in claim 1, wherein said demand constraints on said cloud group comprise one or more of the following: processor, memory, storage, network I/O, storage I/O and bandwidth constraints.
3. The method as recited in claim 1, wherein said placement constraints on said cloud group comprise one or more of the following: high availability, consolidation and energy conservation constraints.
4. The method as recited in claim 1 further comprising:
receiving license enforcement policies; and
identifying said set of compute nodes to host said user's workload based on said virtual machine demands, said demand constraints, said placement constraints and said license enforcement policies.
5. The method as recited in claim 1 further comprising:
deploying virtual machines to said set of compute nodes; and
deploying workloads to said set of compute nodes.
6. The method as recited in claim 5 further comprising one or more of the following:
monitoring for changes in workload demands, said demand constraints and said placement constraints after said deployment of said virtual machines;
monitoring for hardware failures and for predicted hardware failures after said deployment of said virtual machines;
monitoring demands on said virtual machines after said deployment of said virtual machines;
monitoring for hardware usage after said deployment of said virtual machines;
monitoring for additions or subtractions of hardware in said cloud group after said deployment of said virtual machines; and
monitoring for license usage after said deployment of said virtual machines.
7. The method as recited in claim 6 further comprising one or more of the following:
rebalancing computing resources in said cloud group in response to one or more of said workload demands, said demand constraints and said placement constraints changing after said deployment of said virtual machines;
said rebalancing of said computing resources in said cloud group in response to detecting said hardware failure or said predicted hardware failure;
said rebalancing of said computing resources in said cloud group in response to one or more virtual machines of said deployed virtual machines not receiving required demands;
said rebalancing of said computing resources in said cloud group in response to hardware not being fully utilized;
said rebalancing of said computing resources in said cloud group in response to said additions or subtractions of hardware in said cloud group; and
said rebalancing of said computing resources in said cloud group in response to a license being available.
8. The method as recited in claim 7, wherein said rebalancing of said computing resources comprises one or more of the following:
relocating one or more virtual machines in said deployed virtual machines, wherein said relocating comprises chained relocations;
adding one or more virtual machines to said deployed virtual machines in said set of compute nodes; and
relocating one or more virtual machines in said deployed virtual machines in said set of compute nodes of said cloud group to other compute nodes outside of said cloud group.
US13/858,849 2012-01-23 2013-04-08 Dynamically building a set of compute nodes to host the user's workload Active US8930543B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/858,849 US8930543B2 (en) 2012-01-23 2013-04-08 Dynamically building a set of compute nodes to host the user's workload

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US13/356,427 US8930542B2 (en) 2012-01-23 2012-01-23 Dynamically building a set of compute nodes to host the user's workload
US13/858,849 US8930543B2 (en) 2012-01-23 2013-04-08 Dynamically building a set of compute nodes to host the user's workload

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/356,427 Continuation US8930542B2 (en) 2012-01-23 2012-01-23 Dynamically building a set of compute nodes to host the user's workload

Publications (2)

Publication Number Publication Date
US20130227131A1 US20130227131A1 (en) 2013-08-29
US8930543B2 true US8930543B2 (en) 2015-01-06

Family

ID=48798168

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/356,427 Active 2033-01-18 US8930542B2 (en) 2012-01-23 2012-01-23 Dynamically building a set of compute nodes to host the user's workload
US13/858,849 Active US8930543B2 (en) 2012-01-23 2013-04-08 Dynamically building a set of compute nodes to host the user's workload

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/356,427 Active 2033-01-18 US8930542B2 (en) 2012-01-23 2012-01-23 Dynamically building a set of compute nodes to host the user's workload

Country Status (3)

Country Link
US (2) US8930542B2 (en)
CN (1) CN104067260B (en)
WO (1) WO2013110188A1 (en)

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140122706A1 (en) * 2012-10-26 2014-05-01 International Business Machines Corporation Method for determining system topology graph changes in a distributed computing system
US20140196039A1 (en) * 2013-01-08 2014-07-10 Commvault Systems, Inc. Virtual machine categorization system and method
US9286086B2 (en) 2012-12-21 2016-03-15 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US9286110B2 (en) 2013-01-14 2016-03-15 Commvault Systems, Inc. Seamless virtual machine recall in a data storage system
US9417968B2 (en) 2014-09-22 2016-08-16 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US9436555B2 (en) 2014-09-22 2016-09-06 Commvault Systems, Inc. Efficient live-mount of a backed up virtual machine in a storage management system
US9495404B2 (en) 2013-01-11 2016-11-15 Commvault Systems, Inc. Systems and methods to process block-level backup for selective file restoration for virtual machines
US9710465B2 (en) 2014-09-22 2017-07-18 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US9740702B2 (en) 2012-12-21 2017-08-22 Commvault Systems, Inc. Systems and methods to identify unprotected virtual machines
US9823977B2 (en) 2014-11-20 2017-11-21 Commvault Systems, Inc. Virtual machine change block tracking
US9939981B2 (en) 2013-09-12 2018-04-10 Commvault Systems, Inc. File manager integration with virtualization in an information management system with an enhanced storage manager, including user control and storage management of virtual machines
US10152251B2 (en) 2016-10-25 2018-12-11 Commvault Systems, Inc. Targeted backup of virtual machine
US10162528B2 (en) 2016-10-25 2018-12-25 Commvault Systems, Inc. Targeted snapshot based on virtual machine location
US10210022B2 (en) 2016-10-14 2019-02-19 International Business Machines Corporation Feedback mechanism for controlling dispatching work tasks in a multi-tier storage environment
US10387073B2 (en) 2017-03-29 2019-08-20 Commvault Systems, Inc. External dynamic virtual machine synchronization
US10417102B2 (en) 2016-09-30 2019-09-17 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, including virtual machine distribution logic
US10423455B2 (en) 2017-02-03 2019-09-24 Microsoft Technology Licensing, Llc Method for deploying virtual machines in cloud computing systems based on predicted lifetime
US10474542B2 (en) 2017-03-24 2019-11-12 Commvault Systems, Inc. Time-based virtual machine reversion
US10565067B2 (en) 2016-03-09 2020-02-18 Commvault Systems, Inc. Virtual server cloud file system for virtual machine backup from cloud operations
US10650057B2 (en) 2014-07-16 2020-05-12 Commvault Systems, Inc. Volume or virtual machine level backup and generating placeholders for virtual machine files
US10678758B2 (en) 2016-11-21 2020-06-09 Commvault Systems, Inc. Cross-platform virtual machine data and memory backup and replication
US10768971B2 (en) 2019-01-30 2020-09-08 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data
US10776209B2 (en) 2014-11-10 2020-09-15 Commvault Systems, Inc. Cross-platform virtual machine backup and replication
US10877928B2 (en) 2018-03-07 2020-12-29 Commvault Systems, Inc. Using utilities injected into cloud-based virtual machines for speeding up virtual machine backup operations
US10996974B2 (en) 2019-01-30 2021-05-04 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data, including management of cache storage for virtual machine data
US11321189B2 (en) 2014-04-02 2022-05-03 Commvault Systems, Inc. Information management by a media agent in the absence of communications with a storage manager
US11436210B2 (en) 2008-09-05 2022-09-06 Commvault Systems, Inc. Classification of virtualization data
US11442768B2 (en) 2020-03-12 2022-09-13 Commvault Systems, Inc. Cross-hypervisor live recovery of virtual machines
US11449394B2 (en) 2010-06-04 2022-09-20 Commvault Systems, Inc. Failover systems and methods for performing backup operations, including heterogeneous indexing and load balancing of backup and indexing resources
US11467753B2 (en) 2020-02-14 2022-10-11 Commvault Systems, Inc. On-demand restore of virtual machine data
US11500669B2 (en) 2020-05-15 2022-11-15 Commvault Systems, Inc. Live recovery of virtual machines in a public cloud computing environment
US11550680B2 (en) 2018-12-06 2023-01-10 Commvault Systems, Inc. Assigning backup resources in a data storage management system based on failover of partnered data storage resources
US11656951B2 (en) 2020-10-28 2023-05-23 Commvault Systems, Inc. Data loss vulnerability detection
US11663099B2 (en) 2020-03-26 2023-05-30 Commvault Systems, Inc. Snapshot-based disaster recovery orchestration of virtual machine failover and failback operations
US20230168943A1 (en) * 2021-11-29 2023-06-01 Red Hat, Inc. Aggregating host machines into a single cloud node for workloads requiring excessive resources
US12124338B2 (en) 2023-04-10 2024-10-22 Commvault Systems, Inc. Data loss vulnerability detection

Families Citing this family (80)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130263208A1 (en) * 2012-04-02 2013-10-03 Narsimha Reddy Challa Managing virtual machines in a cloud computing system
WO2013184133A1 (en) * 2012-06-08 2013-12-12 Hewlett-Packard Development Company, L.P. Cloud application deployment portability
US9712375B2 (en) * 2012-12-12 2017-07-18 Microsoft Technology Licensing, Llc Workload deployment with infrastructure management agent provisioning
US9514387B2 (en) 2013-09-17 2016-12-06 Google Inc. System and method of monitoring and measuring cluster performance hosted by an IAAS provider by means of outlier detection
US9648040B1 (en) * 2013-09-19 2017-05-09 Amazon Technologies, Inc. Authorization check using a web service request
US9882825B2 (en) * 2014-03-14 2018-01-30 Citrix Systems, Inc. Method and system for securely transmitting volumes into cloud
US10142192B2 (en) 2014-04-09 2018-11-27 International Business Machines Corporation Management of virtual machine resources in computing environments
US10129105B2 (en) 2014-04-09 2018-11-13 International Business Machines Corporation Management of virtual machine placement in computing environments
US10129112B2 (en) 2014-08-14 2018-11-13 At&T Intellectual Property I, L.P. Workflow-based resource management
US10693946B2 (en) 2014-09-16 2020-06-23 Amazon Technologies, Inc. Instance backed mobile devices
US9256467B1 (en) * 2014-11-11 2016-02-09 Amazon Technologies, Inc. System for managing and scheduling containers
US9959148B2 (en) * 2015-02-11 2018-05-01 Wipro Limited Method and device for estimating optimal resources for server virtualization
US9720763B2 (en) 2015-03-09 2017-08-01 Seagate Technology Llc Proactive cloud orchestration
US9916233B1 (en) 2015-03-27 2018-03-13 Amazon Technologies, Inc. Using containers for update deployment
US9998150B1 (en) 2015-06-16 2018-06-12 Amazon Technologies, Inc. Layered data redundancy coding techniques for layer-local data recovery
US10298259B1 (en) 2015-06-16 2019-05-21 Amazon Technologies, Inc. Multi-layered data redundancy coding techniques
US10270475B1 (en) 2015-06-16 2019-04-23 Amazon Technologies, Inc. Layered redundancy coding for encoded parity data
US10270476B1 (en) 2015-06-16 2019-04-23 Amazon Technologies, Inc. Failure mode-sensitive layered redundancy coding techniques
US10977128B1 (en) 2015-06-16 2021-04-13 Amazon Technologies, Inc. Adaptive data loss mitigation for redundancy coding systems
US11061969B1 (en) 2015-06-29 2021-07-13 Amazon Technologies, Inc. Instance backed mobile devices with multiple instances
US10609122B1 (en) 2015-06-29 2020-03-31 Amazon Technologies, Inc. Instance backed building or place
US10198311B1 (en) 2015-07-01 2019-02-05 Amazon Technologies, Inc. Cross-datacenter validation of grid encoded data storage systems
US10108819B1 (en) 2015-07-01 2018-10-23 Amazon Technologies, Inc. Cross-datacenter extension of grid encoded data storage systems
US10089176B1 (en) 2015-07-01 2018-10-02 Amazon Technologies, Inc. Incremental updates of grid encoded data storage systems
US10394762B1 (en) 2015-07-01 2019-08-27 Amazon Technologies, Inc. Determining data redundancy in grid encoded data storage systems
US9998539B1 (en) 2015-07-01 2018-06-12 Amazon Technologies, Inc. Non-parity in grid encoded data storage systems
US9959167B1 (en) 2015-07-01 2018-05-01 Amazon Technologies, Inc. Rebundling grid encoded data storage systems
US10162704B1 (en) 2015-07-01 2018-12-25 Amazon Technologies, Inc. Grid encoded data storage systems for efficient data repair
US9690622B1 (en) 2015-08-24 2017-06-27 Amazon Technologies, Inc. Stateless instance backed mobile devices
US10911404B1 (en) 2015-08-24 2021-02-02 Amazon Technologies, Inc. Attribute based authorization
US9928141B1 (en) 2015-09-21 2018-03-27 Amazon Technologies, Inc. Exploiting variable media size in grid encoded data storage systems
US11386060B1 (en) 2015-09-23 2022-07-12 Amazon Technologies, Inc. Techniques for verifiably processing data in distributed computing systems
US9940474B1 (en) 2015-09-29 2018-04-10 Amazon Technologies, Inc. Techniques and systems for data segregation in data storage systems
US10476773B2 (en) 2015-10-21 2019-11-12 Microsoft Technology Licensing, Llc Substituting window endpoints using a health monitor
US10782990B1 (en) 2015-11-24 2020-09-22 Amazon Technologies, Inc. Container telemetry
US10394789B1 (en) 2015-12-07 2019-08-27 Amazon Technologies, Inc. Techniques and systems for scalable request handling in data processing systems
US10642813B1 (en) 2015-12-14 2020-05-05 Amazon Technologies, Inc. Techniques and systems for storage and processing of operational data
US10248793B1 (en) 2015-12-16 2019-04-02 Amazon Technologies, Inc. Techniques and systems for durable encryption and deletion in data storage systems
US10127105B1 (en) 2015-12-17 2018-11-13 Amazon Technologies, Inc. Techniques for extending grids in data storage systems
US10235402B1 (en) 2015-12-17 2019-03-19 Amazon Technologies, Inc. Techniques for combining grid-encoded data storage systems
US10102065B1 (en) 2015-12-17 2018-10-16 Amazon Technologies, Inc. Localized failure mode decorrelation in redundancy encoded data storage systems
US10324790B1 (en) 2015-12-17 2019-06-18 Amazon Technologies, Inc. Flexible data storage device mapping for data storage systems
US10180912B1 (en) 2015-12-17 2019-01-15 Amazon Technologies, Inc. Techniques and systems for data segregation in redundancy coded data storage systems
US10261782B2 (en) 2015-12-18 2019-04-16 Amazon Technologies, Inc. Software container registry service
US10032032B2 (en) 2015-12-18 2018-07-24 Amazon Technologies, Inc. Software container registry inspection
US10002247B2 (en) 2015-12-18 2018-06-19 Amazon Technologies, Inc. Software container registry container image deployment
US10592336B1 (en) 2016-03-24 2020-03-17 Amazon Technologies, Inc. Layered indexing for asynchronous retrieval of redundancy coded data
US10366062B1 (en) 2016-03-28 2019-07-30 Amazon Technologies, Inc. Cycled clustering for redundancy coded data storage systems
US10061668B1 (en) 2016-03-28 2018-08-28 Amazon Technologies, Inc. Local storage clustering for redundancy coded data storage system
US10678664B1 (en) 2016-03-28 2020-06-09 Amazon Technologies, Inc. Hybridized storage operation for redundancy coded data storage systems
US10228924B2 (en) * 2016-04-19 2019-03-12 International Business Machines Corporation Application deployment and monitoring in a cloud environment to satisfy integrity and geo-fencing constraints
US10069869B2 (en) 2016-05-17 2018-09-04 Amazon Technologies, Inc. Versatile autoscaling
US10791174B2 (en) * 2016-07-28 2020-09-29 Intel Corporation Mechanism for efficient discovery of storage resources in a rack scale architecture system
US11137980B1 (en) 2016-09-27 2021-10-05 Amazon Technologies, Inc. Monotonic time-based data storage
US11281624B1 (en) 2016-09-28 2022-03-22 Amazon Technologies, Inc. Client-based batching of data payload
US10437790B1 (en) 2016-09-28 2019-10-08 Amazon Technologies, Inc. Contextual optimization for data storage systems
US10496327B1 (en) 2016-09-28 2019-12-03 Amazon Technologies, Inc. Command parallelization for data storage systems
US10810157B1 (en) 2016-09-28 2020-10-20 Amazon Technologies, Inc. Command aggregation for data storage operations
US10657097B1 (en) 2016-09-28 2020-05-19 Amazon Technologies, Inc. Data payload aggregation for data storage systems
US11204895B1 (en) 2016-09-28 2021-12-21 Amazon Technologies, Inc. Data payload clustering for data storage systems
US10614239B2 (en) 2016-09-30 2020-04-07 Amazon Technologies, Inc. Immutable cryptographically secured ledger-backed databases
US10412022B1 (en) 2016-10-19 2019-09-10 Amazon Technologies, Inc. On-premises scaling using a versatile scaling service and an application programming interface management service
CN106412094A (en) * 2016-11-02 2017-02-15 深圳前海生生科技有限公司 A method for organizing and managing scattered resources in a public cloud mode
US10296764B1 (en) 2016-11-18 2019-05-21 Amazon Technologies, Inc. Verifiable cryptographically secured ledgers for human resource systems
US10409642B1 (en) 2016-11-22 2019-09-10 Amazon Technologies, Inc. Customer resource monitoring for versatile scaling service scaling policy recommendations
US11269888B1 (en) * 2016-11-28 2022-03-08 Amazon Technologies, Inc. Archival data storage for structured data
KR101714412B1 (en) * 2016-12-28 2017-03-09 주식회사 티맥스클라우드 Method and apparatus for organizing database system in cloud environment
US10585712B2 (en) 2017-05-31 2020-03-10 International Business Machines Corporation Optimizing a workflow of a storlet architecture
US11140455B1 (en) 2017-06-09 2021-10-05 Amazon Technologies, Inc. Video encoder network sandboxing
US10579488B2 (en) * 2017-07-31 2020-03-03 Vmare, Inc. Auto-calculation of recovery plans for disaster recovery solutions
US10827025B2 (en) * 2017-10-18 2020-11-03 Hewlett Packard Enterprise Development Lp Allocations of arbitrary workloads among hyperconverged nodes
US10908940B1 (en) * 2018-02-26 2021-02-02 Amazon Technologies, Inc. Dynamically managed virtual server system
US11144340B2 (en) * 2018-10-04 2021-10-12 Cisco Technology, Inc. Placement of container workloads triggered by network traffic for efficient computing at network edge devices
US11150931B2 (en) * 2018-10-30 2021-10-19 Hewlett Packard Enterprise Development Lp Virtual workload migrations
CN110543363A (en) * 2019-08-05 2019-12-06 慧镕电子系统工程股份有限公司 Virtual machine management method in cloud computing environment
US11669365B1 (en) 2019-08-26 2023-06-06 Amazon Technologies, Inc. Task pool for managed compute instances
CN111832887A (en) * 2020-05-27 2020-10-27 福建亿能达信息技术股份有限公司 Anesthesia doctor workload evaluation system, equipment and medium
US11256493B1 (en) 2020-11-03 2022-02-22 Bank Of America Corporation Container image optimizer
CN113312142B (en) * 2021-02-26 2023-12-26 阿里巴巴集团控股有限公司 Virtualized processing system, method, device and equipment
US11716378B2 (en) * 2021-09-28 2023-08-01 Red Hat, Inc. Optimized network device queue management for hybrid cloud networking workloads

Citations (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060136913A1 (en) 2004-12-09 2006-06-22 International Business Machines Corporation Method, system and computer program product for an automatic resource management of a virtual machine
US20080244600A1 (en) 2007-03-30 2008-10-02 Platform Computing Corporation Method and system for modeling and analyzing computing resource requirements of software applications in a shared and distributed computing environment
US20080295096A1 (en) 2007-05-21 2008-11-27 International Business Machines Corporation DYNAMIC PLACEMENT OF VIRTUAL MACHINES FOR MANAGING VIOLATIONS OF SERVICE LEVEL AGREEMENTS (SLAs)
US20090070771A1 (en) 2007-08-31 2009-03-12 Tom Silangan Yuyitung Method and system for evaluating virtualized environments
US20090228589A1 (en) 2008-03-04 2009-09-10 International Business Machines Corporation Server and storage-aware method for selecting virtual machine migration targets
US20100125473A1 (en) 2008-11-19 2010-05-20 Accenture Global Services Gmbh Cloud computing assessment tool
US20100332658A1 (en) 2009-06-29 2010-12-30 Red Hat Israel, Ltd. Selecting a host from a host cluster to run a virtual machine
US20110016214A1 (en) 2009-07-15 2011-01-20 Cluster Resources, Inc. System and method of brokering cloud computing resources
US20110029672A1 (en) 2009-08-03 2011-02-03 Oracle International Corporation Selection of a suitable node to host a virtual machine in an environment containing a large number of nodes
US20110072293A1 (en) 2009-09-24 2011-03-24 Richard James Mazzaferri Systems and Methods for Attributing An Amount of Power Consumption To A Workload
US20110106516A1 (en) 2009-10-30 2011-05-05 International Business Machines Corporation Automated derivation, design and execution of industry-specific information environment
US20110126207A1 (en) 2009-11-25 2011-05-26 Novell, Inc. System and method for providing annotated service blueprints in an intelligent workload management system
US20110131335A1 (en) 2009-05-08 2011-06-02 Cloudkick, Inc. Methods and systems for cloud computing management
US20110138034A1 (en) 2009-12-03 2011-06-09 International Business Machines Corporation Metering resource usage in a cloud computing environment
US20110161470A1 (en) 2008-07-03 2011-06-30 International Business Machines Corporation Method, System and Computer Program Product for Server Selection, Application Placement and Consolidation Planning of Information Technology Systems
US20110213719A1 (en) * 2010-02-26 2011-09-01 James Michael Ferris Methods and systems for converting standard software licenses for use in cloud computing environments
US20110214005A1 (en) 2010-03-01 2011-09-01 International Business Machines Corporation Optimized placement of virtual machines in a network environment
US20110231899A1 (en) 2009-06-19 2011-09-22 ServiceMesh Corporation System and method for a cloud computing abstraction layer
US20110239010A1 (en) 2010-03-25 2011-09-29 Microsoft Corporation Managing power provisioning in distributed computing
US20110246992A1 (en) 2010-04-01 2011-10-06 International Business Machines Corporation Administration Of Virtual Machine Affinity In A Cloud Computing Environment
US20110264748A1 (en) 2010-04-26 2011-10-27 BitTitan Inc. On-demand mailbox synchronization and migration system
US20110270968A1 (en) 2010-04-30 2011-11-03 Salsburg Michael A Decision support system for moving computing workloads to public clouds
WO2011159842A2 (en) 2010-06-15 2011-12-22 Nimbula, Inc. Virtual computing infrastructure
US20110314466A1 (en) 2010-06-17 2011-12-22 International Business Machines Corporation Creating instances of cloud computing environments
US20110320606A1 (en) 2010-06-29 2011-12-29 International Business Machines Corporation Allocating Computer Resources in a Cloud Environment
US20120023372A1 (en) 2010-07-20 2012-01-26 National Taiwan University Of Science And Technology Estimation method to evaluate a system reliability of a cloud computing network
US20120042055A1 (en) 2010-08-16 2012-02-16 International Business Machines Corporation End-to-end provisioning of storage clouds
US20120066487A1 (en) 2010-09-09 2012-03-15 Novell, Inc. System and method for providing load balancer visibility in an intelligent workload management system
US8140817B2 (en) 2009-02-24 2012-03-20 International Business Machines Corporation Dynamic logical partition management for NUMA machines and clusters
US20120079089A1 (en) 2010-09-29 2012-03-29 National Taiwan University Of Science And Technology Accurate method to evaluate a system reliability of a cloud computing network
US20120079493A1 (en) 2010-09-24 2012-03-29 International Business Machines Corporation Use of constraint-based linear programming to optimize hardware system usage
US20120110260A1 (en) 2010-10-29 2012-05-03 International Business Machines Corporation Automated storage provisioning within a clustered computing environment
US20120117242A1 (en) 2010-11-05 2012-05-10 Hitachi, Ltd. Service linkage system and information processing system
US20120131594A1 (en) 2010-11-24 2012-05-24 Morgan Christopher Edwin Systems and methods for generating dynamically configurable subscription parameters for temporary migration of predictive user workloads in cloud network
US20120130936A1 (en) 2010-11-23 2012-05-24 Novell, Inc. System and method for determining fuzzy cause and effect relationships in an intelligent workload management system
US20120131591A1 (en) 2010-08-24 2012-05-24 Jay Moorthi Method and apparatus for clearing cloud compute demand
US20120137002A1 (en) 2010-11-30 2012-05-31 James Michael Ferris Systems and methods for brokering optimized resource supply costs in host cloud-based network using predictive workloads
US20120173728A1 (en) 2011-01-03 2012-07-05 Gregory Matthew Haskins Policy and identity based workload provisioning
US20120185913A1 (en) 2008-06-19 2012-07-19 Servicemesh, Inc. System and method for a cloud computing abstraction layer with security zone facilities
US20120204169A1 (en) * 2011-02-08 2012-08-09 International Business Machines Corporation Hybrid cloud integrator
US20120272249A1 (en) * 2011-02-25 2012-10-25 International Business Machines Corporation Data Processing Environment Event Correlation
US20120278094A1 (en) 2010-10-12 2012-11-01 Rabit Solutions, LLC Methods and systems for health care record, workflow, and billing management using mobile devices
US20120284408A1 (en) 2011-05-04 2012-11-08 International Business Machines Corporation Workload-aware placement in private heterogeneous clouds
US20120290862A1 (en) 2011-05-13 2012-11-15 International Business Machines Corporation Optimizing energy consumption utilized for workload processing in a networked computing environment
US20120290725A1 (en) 2011-05-09 2012-11-15 Oracle International Corporation Dynamic Cost Model Based Resource Scheduling In Distributed Compute Farms
US20120304179A1 (en) 2011-05-24 2012-11-29 International Business Machines Corporation Workload-to-cloud migration analysis based on cloud aspects
US20120311154A1 (en) 2011-05-31 2012-12-06 Morgan Christopher Edwin Systems and methods for triggering workload movement based on policy stack having multiple selectable inputs
US20130007845A1 (en) * 2011-06-30 2013-01-03 International Business Machines Corporation Authentication and authorization methods for cloud computing security platform
US20130019013A1 (en) 2011-07-12 2013-01-17 Bank Of America Corporation Dynamic Provisioning of Service Requests
US20130024494A1 (en) 2011-06-13 2013-01-24 Steven Guarrieri Methods and systems for platform optimized design
US20130024920A1 (en) 2011-07-21 2013-01-24 International Business Machines Corporation Virtual computer and service
US20130031546A1 (en) 2011-07-28 2013-01-31 International Business Machines Corporation Methods and systems for on-boarding applications to a cloud
US20130042003A1 (en) 2011-08-08 2013-02-14 International Business Machines Corporation Smart cloud workload balancer
US20130042004A1 (en) 2011-08-08 2013-02-14 International Business Machines Corporation Dynamically acquiring computing resources in a networked computing environment
US20130055253A1 (en) 2011-08-30 2013-02-28 Microsoft Corporation Cloud-based build service
US20130061220A1 (en) 2011-09-06 2013-03-07 Xerox Corporation Method for on-demand inter-cloud load provisioning for transient bursts of computing needs
US20130080642A1 (en) * 2011-02-25 2013-03-28 International Business Machines Corporation Data Processing Environment Integration Control
US20130080617A1 (en) 2011-09-22 2013-03-28 Sap Ag Dynamic network load forecasting
US20130111260A1 (en) 2011-10-27 2013-05-02 Sungard Availability Services Lp Dynamic resource allocation in recover to cloud sandbox
US20130111467A1 (en) 2011-10-27 2013-05-02 Cisco Technology, Inc. Dynamic Server Farms
US20130111033A1 (en) 2011-10-31 2013-05-02 Yun Mao Systems, methods, and articles of manufacture to provide cloud resource orchestration
US20130132456A1 (en) 2011-11-17 2013-05-23 Microsoft Corporation Decoupling cluster data from cloud depolyment
US20130179941A1 (en) * 2012-01-06 2013-07-11 International Business Machines Corporation Identifying guests in web meetings
US20130185433A1 (en) 2012-01-13 2013-07-18 Accenture Global Services Limited Performance interference model for managing consolidated workloads in qos-aware clouds
US20130185667A1 (en) 2012-01-18 2013-07-18 International Business Machines Corporation Open resilience framework for simplified and coordinated orchestration of multiple availability managers
US8639791B2 (en) 2010-05-20 2014-01-28 Novell, Inc. Techniques for evaluating and managing cloud networks
US8656019B2 (en) 2009-12-17 2014-02-18 International Business Machines Corporation Data processing workload administration in a cloud computing environment

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101840346B (en) * 2009-12-30 2013-08-21 北京世纪互联宽带数据中心有限公司 Method and system for deploying cloud host computer
US9176773B2 (en) * 2011-06-29 2015-11-03 Microsoft Technology Licensing, Llc Virtual machine migration tool

Patent Citations (76)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060136913A1 (en) 2004-12-09 2006-06-22 International Business Machines Corporation Method, system and computer program product for an automatic resource management of a virtual machine
US20080244600A1 (en) 2007-03-30 2008-10-02 Platform Computing Corporation Method and system for modeling and analyzing computing resource requirements of software applications in a shared and distributed computing environment
US20140047119A1 (en) 2007-03-30 2014-02-13 International Business Machines Corporation Method and system for modeling and analyzing computing resource requirements of software applications in a shared and distributed computing environment
US20080295096A1 (en) 2007-05-21 2008-11-27 International Business Machines Corporation DYNAMIC PLACEMENT OF VIRTUAL MACHINES FOR MANAGING VIOLATIONS OF SERVICE LEVEL AGREEMENTS (SLAs)
US20090070771A1 (en) 2007-08-31 2009-03-12 Tom Silangan Yuyitung Method and system for evaluating virtualized environments
US20090228589A1 (en) 2008-03-04 2009-09-10 International Business Machines Corporation Server and storage-aware method for selecting virtual machine migration targets
US20120185913A1 (en) 2008-06-19 2012-07-19 Servicemesh, Inc. System and method for a cloud computing abstraction layer with security zone facilities
US20110161470A1 (en) 2008-07-03 2011-06-30 International Business Machines Corporation Method, System and Computer Program Product for Server Selection, Application Placement and Consolidation Planning of Information Technology Systems
US20100125473A1 (en) 2008-11-19 2010-05-20 Accenture Global Services Gmbh Cloud computing assessment tool
US20110276686A1 (en) 2008-11-19 2011-11-10 Accenture Global Services Limited Cloud computing assessment tool
US8140817B2 (en) 2009-02-24 2012-03-20 International Business Machines Corporation Dynamic logical partition management for NUMA machines and clusters
US20110131335A1 (en) 2009-05-08 2011-06-02 Cloudkick, Inc. Methods and systems for cloud computing management
US20110231899A1 (en) 2009-06-19 2011-09-22 ServiceMesh Corporation System and method for a cloud computing abstraction layer
US20100332658A1 (en) 2009-06-29 2010-12-30 Red Hat Israel, Ltd. Selecting a host from a host cluster to run a virtual machine
US20110016214A1 (en) 2009-07-15 2011-01-20 Cluster Resources, Inc. System and method of brokering cloud computing resources
US20110029672A1 (en) 2009-08-03 2011-02-03 Oracle International Corporation Selection of a suitable node to host a virtual machine in an environment containing a large number of nodes
US20110072293A1 (en) 2009-09-24 2011-03-24 Richard James Mazzaferri Systems and Methods for Attributing An Amount of Power Consumption To A Workload
US20110106516A1 (en) 2009-10-30 2011-05-05 International Business Machines Corporation Automated derivation, design and execution of industry-specific information environment
US20110126275A1 (en) 2009-11-25 2011-05-26 Novell, Inc. System and method for discovery enrichment in an intelligent workload management system
US20110126197A1 (en) 2009-11-25 2011-05-26 Novell, Inc. System and method for controlling cloud and virtualized data centers in an intelligent workload management system
US20110126207A1 (en) 2009-11-25 2011-05-26 Novell, Inc. System and method for providing annotated service blueprints in an intelligent workload management system
US20110138034A1 (en) 2009-12-03 2011-06-09 International Business Machines Corporation Metering resource usage in a cloud computing environment
US8656019B2 (en) 2009-12-17 2014-02-18 International Business Machines Corporation Data processing workload administration in a cloud computing environment
US20110213719A1 (en) * 2010-02-26 2011-09-01 James Michael Ferris Methods and systems for converting standard software licenses for use in cloud computing environments
US20110214005A1 (en) 2010-03-01 2011-09-01 International Business Machines Corporation Optimized placement of virtual machines in a network environment
US8627123B2 (en) 2010-03-25 2014-01-07 Microsoft Corporation Managing power provisioning in distributed computing
US20110239010A1 (en) 2010-03-25 2011-09-29 Microsoft Corporation Managing power provisioning in distributed computing
US20110246992A1 (en) 2010-04-01 2011-10-06 International Business Machines Corporation Administration Of Virtual Machine Affinity In A Cloud Computing Environment
US20110264748A1 (en) 2010-04-26 2011-10-27 BitTitan Inc. On-demand mailbox synchronization and migration system
US20110270968A1 (en) 2010-04-30 2011-11-03 Salsburg Michael A Decision support system for moving computing workloads to public clouds
US8639791B2 (en) 2010-05-20 2014-01-28 Novell, Inc. Techniques for evaluating and managing cloud networks
WO2011159842A2 (en) 2010-06-15 2011-12-22 Nimbula, Inc. Virtual computing infrastructure
US20110314466A1 (en) 2010-06-17 2011-12-22 International Business Machines Corporation Creating instances of cloud computing environments
US20110320606A1 (en) 2010-06-29 2011-12-29 International Business Machines Corporation Allocating Computer Resources in a Cloud Environment
US20120023372A1 (en) 2010-07-20 2012-01-26 National Taiwan University Of Science And Technology Estimation method to evaluate a system reliability of a cloud computing network
US20120042055A1 (en) 2010-08-16 2012-02-16 International Business Machines Corporation End-to-end provisioning of storage clouds
US8478845B2 (en) 2010-08-16 2013-07-02 International Business Machines Corporation End-to-end provisioning of storage clouds
US20120131591A1 (en) 2010-08-24 2012-05-24 Jay Moorthi Method and apparatus for clearing cloud compute demand
US20120066487A1 (en) 2010-09-09 2012-03-15 Novell, Inc. System and method for providing load balancer visibility in an intelligent workload management system
US20120079493A1 (en) 2010-09-24 2012-03-29 International Business Machines Corporation Use of constraint-based linear programming to optimize hardware system usage
US20120079089A1 (en) 2010-09-29 2012-03-29 National Taiwan University Of Science And Technology Accurate method to evaluate a system reliability of a cloud computing network
US20120278094A1 (en) 2010-10-12 2012-11-01 Rabit Solutions, LLC Methods and systems for health care record, workflow, and billing management using mobile devices
US20120110260A1 (en) 2010-10-29 2012-05-03 International Business Machines Corporation Automated storage provisioning within a clustered computing environment
US8489812B2 (en) 2010-10-29 2013-07-16 International Business Machines Corporation Automated storage provisioning within a clustered computing environment
US20120117242A1 (en) 2010-11-05 2012-05-10 Hitachi, Ltd. Service linkage system and information processing system
US20120130936A1 (en) 2010-11-23 2012-05-24 Novell, Inc. System and method for determining fuzzy cause and effect relationships in an intelligent workload management system
US8620851B2 (en) * 2010-11-23 2013-12-31 Novell, Inc. System and method for determining fuzzy cause and effect relationships in an intelligent workload management system
US20120131594A1 (en) 2010-11-24 2012-05-24 Morgan Christopher Edwin Systems and methods for generating dynamically configurable subscription parameters for temporary migration of predictive user workloads in cloud network
US20120137002A1 (en) 2010-11-30 2012-05-31 James Michael Ferris Systems and methods for brokering optimized resource supply costs in host cloud-based network using predictive workloads
US20120173728A1 (en) 2011-01-03 2012-07-05 Gregory Matthew Haskins Policy and identity based workload provisioning
US20120204169A1 (en) * 2011-02-08 2012-08-09 International Business Machines Corporation Hybrid cloud integrator
US20130080642A1 (en) * 2011-02-25 2013-03-28 International Business Machines Corporation Data Processing Environment Integration Control
US20120272249A1 (en) * 2011-02-25 2012-10-25 International Business Machines Corporation Data Processing Environment Event Correlation
US20120284408A1 (en) 2011-05-04 2012-11-08 International Business Machines Corporation Workload-aware placement in private heterogeneous clouds
US20120290725A1 (en) 2011-05-09 2012-11-15 Oracle International Corporation Dynamic Cost Model Based Resource Scheduling In Distributed Compute Farms
US20120290862A1 (en) 2011-05-13 2012-11-15 International Business Machines Corporation Optimizing energy consumption utilized for workload processing in a networked computing environment
US8612785B2 (en) 2011-05-13 2013-12-17 International Business Machines Corporation Optimizing energy consumption utilized for workload processing in a networked computing environment
US20120304179A1 (en) 2011-05-24 2012-11-29 International Business Machines Corporation Workload-to-cloud migration analysis based on cloud aspects
US20120311154A1 (en) 2011-05-31 2012-12-06 Morgan Christopher Edwin Systems and methods for triggering workload movement based on policy stack having multiple selectable inputs
US20130024494A1 (en) 2011-06-13 2013-01-24 Steven Guarrieri Methods and systems for platform optimized design
US20130007845A1 (en) * 2011-06-30 2013-01-03 International Business Machines Corporation Authentication and authorization methods for cloud computing security platform
US20130019013A1 (en) 2011-07-12 2013-01-17 Bank Of America Corporation Dynamic Provisioning of Service Requests
US20130024920A1 (en) 2011-07-21 2013-01-24 International Business Machines Corporation Virtual computer and service
US20130031546A1 (en) 2011-07-28 2013-01-31 International Business Machines Corporation Methods and systems for on-boarding applications to a cloud
US20130042004A1 (en) 2011-08-08 2013-02-14 International Business Machines Corporation Dynamically acquiring computing resources in a networked computing environment
US20130042003A1 (en) 2011-08-08 2013-02-14 International Business Machines Corporation Smart cloud workload balancer
US20130055253A1 (en) 2011-08-30 2013-02-28 Microsoft Corporation Cloud-based build service
US20130061220A1 (en) 2011-09-06 2013-03-07 Xerox Corporation Method for on-demand inter-cloud load provisioning for transient bursts of computing needs
US20130080617A1 (en) 2011-09-22 2013-03-28 Sap Ag Dynamic network load forecasting
US20130111467A1 (en) 2011-10-27 2013-05-02 Cisco Technology, Inc. Dynamic Server Farms
US20130111260A1 (en) 2011-10-27 2013-05-02 Sungard Availability Services Lp Dynamic resource allocation in recover to cloud sandbox
US20130111033A1 (en) 2011-10-31 2013-05-02 Yun Mao Systems, methods, and articles of manufacture to provide cloud resource orchestration
US20130132456A1 (en) 2011-11-17 2013-05-23 Microsoft Corporation Decoupling cluster data from cloud depolyment
US20130179941A1 (en) * 2012-01-06 2013-07-11 International Business Machines Corporation Identifying guests in web meetings
US20130185433A1 (en) 2012-01-13 2013-07-18 Accenture Global Services Limited Performance interference model for managing consolidated workloads in qos-aware clouds
US20130185667A1 (en) 2012-01-18 2013-07-18 International Business Machines Corporation Open resilience framework for simplified and coordinated orchestration of multiple availability managers

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
International Search Report for International Application No. PCT/CA2013/050041 dated May 6, 2013, pp. 1-3.
Mell et al., "The NIST Definition of Cloud Computing," Special Publication 800-145, Sep. 2011.
Office Action for U.S. Appl. No. 13/356,427 dated Mar. 19, 2014, pp. 1-21.
Vecchiola et al., "Aneka: A Software Platform for .NET-based Cloud Computing," http://www.buyya.conn/gridbus/reports/AnekaCloudPlatform2009.pdf, 2009, pp. 1-30.

Cited By (96)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11436210B2 (en) 2008-09-05 2022-09-06 Commvault Systems, Inc. Classification of virtualization data
US12001295B2 (en) 2010-06-04 2024-06-04 Commvault Systems, Inc. Heterogeneous indexing and load balancing of backup and indexing resources
US11449394B2 (en) 2010-06-04 2022-09-20 Commvault Systems, Inc. Failover systems and methods for performing backup operations, including heterogeneous indexing and load balancing of backup and indexing resources
US10554508B2 (en) 2012-10-26 2020-02-04 International Business Machines Corporation Updating a topology graph representing a distributed computing system by monitoring predefined parameters with respect to predetermined performance threshold values and using predetermined rules to select a combination of application, storage and database server nodes to meet at least one service level objective (SLO)
US9455881B2 (en) * 2012-10-26 2016-09-27 International Business Machines Corporation Method for determining system topology graph changes in a distributed computing system
US20140122706A1 (en) * 2012-10-26 2014-05-01 International Business Machines Corporation Method for determining system topology graph changes in a distributed computing system
US9740702B2 (en) 2012-12-21 2017-08-22 Commvault Systems, Inc. Systems and methods to identify unprotected virtual machines
US9311121B2 (en) 2012-12-21 2016-04-12 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US10733143B2 (en) 2012-12-21 2020-08-04 Commvault Systems, Inc. Systems and methods to identify unprotected virtual machines
US10824464B2 (en) 2012-12-21 2020-11-03 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US11544221B2 (en) 2012-12-21 2023-01-03 Commvault Systems, Inc. Systems and methods to identify unprotected virtual machines
US9684535B2 (en) 2012-12-21 2017-06-20 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US10684883B2 (en) 2012-12-21 2020-06-16 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US11468005B2 (en) 2012-12-21 2022-10-11 Commvault Systems, Inc. Systems and methods to identify unprotected virtual machines
US9286086B2 (en) 2012-12-21 2016-03-15 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US11099886B2 (en) 2012-12-21 2021-08-24 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US9965316B2 (en) 2012-12-21 2018-05-08 Commvault Systems, Inc. Archiving virtual machines in a data storage system
US20140196039A1 (en) * 2013-01-08 2014-07-10 Commvault Systems, Inc. Virtual machine categorization system and method
US10896053B2 (en) 2013-01-08 2021-01-19 Commvault Systems, Inc. Virtual machine load balancing
US11922197B2 (en) 2013-01-08 2024-03-05 Commvault Systems, Inc. Virtual server agent load balancing
US9977687B2 (en) 2013-01-08 2018-05-22 Commvault Systems, Inc. Virtual server agent load balancing
US10474483B2 (en) 2013-01-08 2019-11-12 Commvault Systems, Inc. Virtual server agent load balancing
US9703584B2 (en) 2013-01-08 2017-07-11 Commvault Systems, Inc. Virtual server agent load balancing
US11734035B2 (en) 2013-01-08 2023-08-22 Commvault Systems, Inc. Virtual machine load balancing
US9495404B2 (en) 2013-01-11 2016-11-15 Commvault Systems, Inc. Systems and methods to process block-level backup for selective file restoration for virtual machines
US10108652B2 (en) 2013-01-11 2018-10-23 Commvault Systems, Inc. Systems and methods to process block-level backup for selective file restoration for virtual machines
US9489244B2 (en) 2013-01-14 2016-11-08 Commvault Systems, Inc. Seamless virtual machine recall in a data storage system
US9766989B2 (en) 2013-01-14 2017-09-19 Commvault Systems, Inc. Creation of virtual machine placeholders in a data storage system
US9652283B2 (en) 2013-01-14 2017-05-16 Commvault Systems, Inc. Creation of virtual machine placeholders in a data storage system
US9286110B2 (en) 2013-01-14 2016-03-15 Commvault Systems, Inc. Seamless virtual machine recall in a data storage system
US9939981B2 (en) 2013-09-12 2018-04-10 Commvault Systems, Inc. File manager integration with virtualization in an information management system with an enhanced storage manager, including user control and storage management of virtual machines
US11010011B2 (en) 2013-09-12 2021-05-18 Commvault Systems, Inc. File manager integration with virtualization in an information management system with an enhanced storage manager, including user control and storage management of virtual machines
US11321189B2 (en) 2014-04-02 2022-05-03 Commvault Systems, Inc. Information management by a media agent in the absence of communications with a storage manager
US11625439B2 (en) 2014-07-16 2023-04-11 Commvault Systems, Inc. Volume or virtual machine level backup and generating placeholders for virtual machine files
US10650057B2 (en) 2014-07-16 2020-05-12 Commvault Systems, Inc. Volume or virtual machine level backup and generating placeholders for virtual machine files
US9996534B2 (en) 2014-09-22 2018-06-12 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US10452303B2 (en) 2014-09-22 2019-10-22 Commvault Systems, Inc. Efficient live-mount of a backed up virtual machine in a storage management system
US10437505B2 (en) 2014-09-22 2019-10-08 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US10048889B2 (en) 2014-09-22 2018-08-14 Commvault Systems, Inc. Efficient live-mount of a backed up virtual machine in a storage management system
US9928001B2 (en) 2014-09-22 2018-03-27 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US10572468B2 (en) 2014-09-22 2020-02-25 Commvault Systems, Inc. Restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US9710465B2 (en) 2014-09-22 2017-07-18 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US9436555B2 (en) 2014-09-22 2016-09-06 Commvault Systems, Inc. Efficient live-mount of a backed up virtual machine in a storage management system
US9417968B2 (en) 2014-09-22 2016-08-16 Commvault Systems, Inc. Efficiently restoring execution of a backed up virtual machine based on coordination with virtual-machine-file-relocation operations
US10776209B2 (en) 2014-11-10 2020-09-15 Commvault Systems, Inc. Cross-platform virtual machine backup and replication
US12061798B2 (en) 2014-11-20 2024-08-13 Commvault Systems, Inc. Virtual machine change block tracking
US11422709B2 (en) 2014-11-20 2022-08-23 Commvault Systems, Inc. Virtual machine change block tracking
US10509573B2 (en) 2014-11-20 2019-12-17 Commvault Systems, Inc. Virtual machine change block tracking
US9996287B2 (en) 2014-11-20 2018-06-12 Commvault Systems, Inc. Virtual machine change block tracking
US9983936B2 (en) 2014-11-20 2018-05-29 Commvault Systems, Inc. Virtual machine change block tracking
US9823977B2 (en) 2014-11-20 2017-11-21 Commvault Systems, Inc. Virtual machine change block tracking
US10592350B2 (en) 2016-03-09 2020-03-17 Commvault Systems, Inc. Virtual server cloud file system for virtual machine restore to cloud operations
US10565067B2 (en) 2016-03-09 2020-02-18 Commvault Systems, Inc. Virtual server cloud file system for virtual machine backup from cloud operations
US12038814B2 (en) 2016-03-09 2024-07-16 Commvault Systems, Inc. Virtual server cloud file system for backing up cloud-based virtual machine data
US10747630B2 (en) 2016-09-30 2020-08-18 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, including operations by a master monitor node
US10417102B2 (en) 2016-09-30 2019-09-17 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, including virtual machine distribution logic
US10896104B2 (en) 2016-09-30 2021-01-19 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, using ping monitoring of target virtual machines
US11429499B2 (en) 2016-09-30 2022-08-30 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, including operations by a master monitor node
US10474548B2 (en) 2016-09-30 2019-11-12 Commvault Systems, Inc. Heartbeat monitoring of virtual machines for initiating failover operations in a data storage management system, using ping monitoring of target virtual machines
US10210022B2 (en) 2016-10-14 2019-02-19 International Business Machines Corporation Feedback mechanism for controlling dispatching work tasks in a multi-tier storage environment
US10296390B2 (en) 2016-10-14 2019-05-21 International Business Machines Corporation Feedback mechanism for controlling dispatching work tasks in a multi-tier storage environment
US11934859B2 (en) 2016-10-25 2024-03-19 Commvault Systems, Inc. Targeted snapshot based on virtual machine location
US10824459B2 (en) 2016-10-25 2020-11-03 Commvault Systems, Inc. Targeted snapshot based on virtual machine location
US11416280B2 (en) 2016-10-25 2022-08-16 Commvault Systems, Inc. Targeted snapshot based on virtual machine location
US10162528B2 (en) 2016-10-25 2018-12-25 Commvault Systems, Inc. Targeted snapshot based on virtual machine location
US10152251B2 (en) 2016-10-25 2018-12-11 Commvault Systems, Inc. Targeted backup of virtual machine
US11436202B2 (en) 2016-11-21 2022-09-06 Commvault Systems, Inc. Cross-platform virtual machine data and memory backup and replication
US10678758B2 (en) 2016-11-21 2020-06-09 Commvault Systems, Inc. Cross-platform virtual machine data and memory backup and replication
US10423455B2 (en) 2017-02-03 2019-09-24 Microsoft Technology Licensing, Llc Method for deploying virtual machines in cloud computing systems based on predicted lifetime
US11526410B2 (en) 2017-03-24 2022-12-13 Commvault Systems, Inc. Time-based virtual machine reversion
US10877851B2 (en) 2017-03-24 2020-12-29 Commvault Systems, Inc. Virtual machine recovery point selection
US10474542B2 (en) 2017-03-24 2019-11-12 Commvault Systems, Inc. Time-based virtual machine reversion
US10896100B2 (en) 2017-03-24 2021-01-19 Commvault Systems, Inc. Buffered virtual machine replication
US10983875B2 (en) 2017-03-24 2021-04-20 Commvault Systems, Inc. Time-based virtual machine reversion
US12032455B2 (en) 2017-03-24 2024-07-09 Commvault Systems, Inc. Time-based virtual machine reversion
US11669414B2 (en) 2017-03-29 2023-06-06 Commvault Systems, Inc. External dynamic virtual machine synchronization
US11249864B2 (en) 2017-03-29 2022-02-15 Commvault Systems, Inc. External dynamic virtual machine synchronization
US10387073B2 (en) 2017-03-29 2019-08-20 Commvault Systems, Inc. External dynamic virtual machine synchronization
US10877928B2 (en) 2018-03-07 2020-12-29 Commvault Systems, Inc. Using utilities injected into cloud-based virtual machines for speeding up virtual machine backup operations
US11520736B2 (en) 2018-03-07 2022-12-06 Commvault Systems, Inc. Using utilities injected into cloud-based virtual machines for speeding up virtual machine backup operations
US11550680B2 (en) 2018-12-06 2023-01-10 Commvault Systems, Inc. Assigning backup resources in a data storage management system based on failover of partnered data storage resources
US10996974B2 (en) 2019-01-30 2021-05-04 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data, including management of cache storage for virtual machine data
US10768971B2 (en) 2019-01-30 2020-09-08 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data
US11947990B2 (en) 2019-01-30 2024-04-02 Commvault Systems, Inc. Cross-hypervisor live-mount of backed up virtual machine data
US11467863B2 (en) 2019-01-30 2022-10-11 Commvault Systems, Inc. Cross-hypervisor live mount of backed up virtual machine data
US11467753B2 (en) 2020-02-14 2022-10-11 Commvault Systems, Inc. On-demand restore of virtual machine data
US11714568B2 (en) 2020-02-14 2023-08-01 Commvault Systems, Inc. On-demand restore of virtual machine data
US11442768B2 (en) 2020-03-12 2022-09-13 Commvault Systems, Inc. Cross-hypervisor live recovery of virtual machines
US11663099B2 (en) 2020-03-26 2023-05-30 Commvault Systems, Inc. Snapshot-based disaster recovery orchestration of virtual machine failover and failback operations
US12086624B2 (en) 2020-05-15 2024-09-10 Commvault Systems, Inc. Live recovery of virtual machines in a public cloud computing environment based on temporary live mount
US11748143B2 (en) 2020-05-15 2023-09-05 Commvault Systems, Inc. Live mount of virtual machines in a public cloud computing environment
US11500669B2 (en) 2020-05-15 2022-11-15 Commvault Systems, Inc. Live recovery of virtual machines in a public cloud computing environment
US11656951B2 (en) 2020-10-28 2023-05-23 Commvault Systems, Inc. Data loss vulnerability detection
US20230168943A1 (en) * 2021-11-29 2023-06-01 Red Hat, Inc. Aggregating host machines into a single cloud node for workloads requiring excessive resources
US11755375B2 (en) * 2021-11-29 2023-09-12 Red Hat, Inc. Aggregating host machines into a single cloud node for workloads requiring excessive resources
US12124338B2 (en) 2023-04-10 2024-10-22 Commvault Systems, Inc. Data loss vulnerability detection

Also Published As

Publication number Publication date
CN104067260A (en) 2014-09-24
WO2013110188A1 (en) 2013-08-01
US8930542B2 (en) 2015-01-06
CN104067260B (en) 2016-09-28
US20130191527A1 (en) 2013-07-25
US20130227131A1 (en) 2013-08-29

Similar Documents

Publication Publication Date Title
US8930543B2 (en) Dynamically building a set of compute nodes to host the user's workload
US9021097B2 (en) Unified cloud computing infrastructure to manage and deploy physical and virtual environments
US9286130B2 (en) Optimizing virtual machine deployment time by temporarily allocating more processing resources during the initial deployment time of the virtual machine
US8949848B2 (en) Reducing usage of resource utilized by a virtual machine whose resource utilization is adversely affecting neighboring virtual machines
US9588797B2 (en) Returning terminated virtual machines to a pool of available virtual machines to be reused thereby optimizing cloud resource usage and workload deployment time
US8972990B2 (en) Providing a seamless transition for resizing virtual machines from a development environment to a production environment
US9300726B2 (en) Implementing a private network isolated from a user network for virtual machine deployment and migration and for monitoring and managing the cloud environment
US9071587B2 (en) Optimizing deployment of virtual machines by pre-generating a virtual machine's first-boot metadata by cloud controller
US9015325B2 (en) Performing maintenance operations on cloud computing node without requiring to stop all virtual machines in the node
US9170850B2 (en) Minimizing workload migrations during cloud maintenance operations
US9705744B2 (en) Updating hardware and software components of cloud computing environment at optimal times
US9665388B2 (en) Selecting optimal hypervisor platforms that satisfy application workload requirements
US9009359B2 (en) Emulating multiple universal serial bus (USB) keys so as to efficiently configure different types of hardware
US9282147B2 (en) Determining location of hardware components in a cloud computing environment based on hardware components self-locating other hardware components
US9529624B2 (en) Determining a relative priority for a job using context and environmental considerations
US9043648B2 (en) Intelligently responding to hardware failures so as to optimize system performance
US9053446B2 (en) Dynamically quantifying the demand of each software component of each software stack deployed in the cloud environment

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ASHOK, ROHITH K.;BRABSON, ROY F.;HOCKETT, HUGH E.;AND OTHERS;SIGNING DATES FROM 20130319 TO 20130325;REEL/FRAME:030173/0260

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8