US20220050711A1 - Systems and methods to orchestrate infrastructure installation of a hybrid system - Google Patents
Systems and methods to orchestrate infrastructure installation of a hybrid system Download PDFInfo
- Publication number
- US20220050711A1 US20220050711A1 US17/514,897 US202117514897A US2022050711A1 US 20220050711 A1 US20220050711 A1 US 20220050711A1 US 202117514897 A US202117514897 A US 202117514897A US 2022050711 A1 US2022050711 A1 US 2022050711A1
- Authority
- US
- United States
- Prior art keywords
- task
- dependency
- error
- server
- processor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000009434 installation Methods 0.000 title claims abstract description 68
- 238000000034 method Methods 0.000 title claims abstract description 47
- 238000003860 storage Methods 0.000 claims description 38
- 238000011084 recovery Methods 0.000 claims description 17
- 238000013479 data entry Methods 0.000 claims description 16
- 230000001419 dependent effect Effects 0.000 claims description 8
- 230000001960 triggered effect Effects 0.000 claims description 8
- 239000003795 chemical substances by application Substances 0.000 description 102
- 238000007726 management method Methods 0.000 description 97
- 238000004891 communication Methods 0.000 description 22
- 230000008569 process Effects 0.000 description 15
- 238000010200 validation analysis Methods 0.000 description 11
- 238000013515 script Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 238000004519 manufacturing process Methods 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000013500 data storage Methods 0.000 description 4
- 238000011161 development Methods 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 238000012546 transfer Methods 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 238000013475 authorization Methods 0.000 description 3
- 230000008859 change Effects 0.000 description 3
- 238000011156 evaluation Methods 0.000 description 3
- 238000002955 isolation Methods 0.000 description 3
- 230000006855 networking Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000012795 verification Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000003139 buffering effect Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 239000008241 heterogeneous mixture Substances 0.000 description 1
- 230000003116 impacting effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000000644 propagated effect Effects 0.000 description 1
- 230000003362 replicative effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 238000013341 scale-up Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Program initiating; Program switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/4881—Scheduling strategies for dispatcher, e.g. round robin, multi-level priority queues
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/34—Signalling channels for network management communication
- H04L41/344—Out-of-band transfers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/61—Installation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/60—Software deployment
- G06F8/65—Updates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/455—Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
- G06F9/45533—Hypervisors; Virtual machine monitors
- G06F9/45558—Hypervisor-specific management and integration aspects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/50—Allocation of resources, e.g. of the central processing unit [CPU]
- G06F9/5005—Allocation of resources, e.g. of the central processing unit [CPU] to service a request
- G06F9/5027—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
- G06F9/5055—Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals considering software capabilities, i.e. software resources associated or available to the machine
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/34—Network arrangements or protocols for supporting network services or applications involving the movement of software or configuration parameters
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/455—Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
- G06F9/45533—Hypervisors; Virtual machine monitors
- G06F9/45558—Hypervisor-specific management and integration aspects
- G06F2009/4557—Distribution of virtual machine instances; Migration and load balancing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/455—Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
- G06F9/45533—Hypervisors; Virtual machine monitors
- G06F9/45558—Hypervisor-specific management and integration aspects
- G06F2009/45595—Network integration; Enabling network access in virtual machine instances
Definitions
- the present disclosure relates generally to cloud computing and, more particularly, to methods and apparatus to orchestrate infrastructure installation of a hybrid system in a distributed environment such as a cloud computing environment.
- Virtualizing computer systems provides benefits such as an ability to execute multiple computer systems on a single hardware computer, replicating computer systems, moving computer systems among multiple hardware computers, and so forth.
- IaaS infrastructure-as-a-Service
- cloud computing platform generally describes a suite of technologies provided as an integrated solution to allow for elastic creation of a virtualized, networked, and pooled computing platform (sometimes referred to as a “cloud computing platform”).
- Enterprises may use IaaS as a business-internal organizational cloud computing platform (sometimes referred to as a “private cloud”) that gives an application developer access to infrastructure resources, such as virtualized servers, storage, and networking resources.
- infrastructure resources such as virtualized servers, storage, and networking resources.
- Cloud computing environments may include many processing units (e.g., servers). Other components of a cloud computing environment include storage devices, networking devices (e.g., switches), etc. Current cloud computing environment configuration relies on much manual user input and configuration to install, configure, and deploy the components of the cloud computing environment.
- FIG. 1 depicts an example system constructed in accordance with the teachings of this disclosure for managing a cloud computing platform.
- FIG. 2 illustrates an example generation of a multi-machine blueprint by the example blueprint manager of FIG. 1 .
- FIG. 3 illustrates an example installation of deployed virtual machines and associated servers acting as hosts for deployment of component servers for a customer.
- FIG. 4 illustrates an example implementation of a virtual appliance.
- FIG. 5 illustrates a block diagram of an example implementation of a management agent.
- FIG. 6 illustrates an example data flow diagram showing an exchange of information between an appliance and a component server for installation.
- FIG. 7 illustrates an example schedule of components for installation in an example system.
- FIGS. 8-9 depict a flowcharts representative of computer readable instructions that may be executed to implement example infrastructure installation.
- FIG. 10 is a block diagram of an example processing platform capable of executing the example machine-readable instructions of FIGS. 8-9 .
- Cloud computing is based on the deployment of many physical resources across a network, virtualizing the physical resources into virtual resources, and provisioning the virtual resources to perform cloud computing services and applications.
- Example systems for virtualizing computer systems are described in U.S. patent application Ser. No. 11/903,374, entitled “METHOD AND SYSTEM FOR MANAGING VIRTUAL AND REAL MACHINES,” filed Sep. 21, 2007, and granted as U.S. Pat. No. 8,171,485, U.S. Provisional Patent Application No. 60/919,965, entitled “METHOD AND SYSTEM FOR MANAGING VIRTUAL AND REAL MACHINES,” filed Mar. 26, 2007, and U.S. Provisional Patent Application No. 61/736,422, entitled “METHODS AND APPARATUS FOR VIRTUALIZED COMPUTING,” filed Dec. 12, 2012, all three of which are hereby incorporated herein by reference in their entirety.
- Cloud computing platforms may provide many powerful capabilities for performing computing operations. However, taking advantage of these computing capabilities manually may be complex and/or require significant training and/or expertise. Prior techniques to providing cloud computing platforms and services often require customers to understand details and configurations of hardware and software resources to establish and configure the cloud computing platform. Methods and apparatus disclosed herein facilitate the management of virtual machine resources in cloud computing platforms.
- a virtual machine is a software computer that, like a physical computer, runs an operating system and applications.
- An operating system installed on a virtual machine is referred to as a guest operating system.
- VMs virtual machines
- Virtual machines can be used as desktop or workstation environments, as testing environments, to consolidate server applications, etc.
- Virtual machines can run on hosts or clusters. The same host can run a plurality of VMs, for example.
- methods and apparatus disclosed herein provide for automation of management tasks such as provisioning multiple virtual machines for a multiple-machine computing system (e.g., a group of servers that inter-operate), linking provisioned virtual machines and tasks to desired systems to execute those virtual machines or tasks, and/or reclaiming cloud computing resources that are no longer in use.
- the improvements to cloud management systems e.g., the vCloud Automation Center (vCAC) from VMware®, the vRealize Automation Cloud Automation Software from VMware®), interfaces, portals, etc. disclosed herein may be utilized individually and/or in any combination. For example, all or a subset of the described improvements may be utilized.
- availability refers to the level of redundancy required to provide continuous operation expected for the workload domain.
- performance refers to the computer processing unit (CPU) operating speeds (e.g., CPU gigahertz (GHz)), memory (e.g., gigabytes (GB) of random access memory (RAM)), mass storage (e.g., GB hard drive disk (HDD), GB solid state drive (SSD)), and power capabilities of a workload domain.
- capacity refers to the aggregate number of resources (e.g., aggregate storage, aggregate CPU, etc.) across all servers associated with a cluster and/or a workload domain.
- the number of resources (e.g., capacity) for a workload domain is determined based on the redundancy, the CPU operating speed, the memory, the storage, the security, and/or the power requirements selected by a user. For example, more resources are required for a workload domain as the user-selected requirements increase (e.g., higher redundancy, CPU speed, memory, storage, security, and/or power options require more resources than lower redundancy, CPU speed, memory, storage, security, and/or power options).
- Virtualization environments Many different types exist. Three example types of virtualization environment are: full virtualization, paravirtualization, and operating system virtualization.
- Full virtualization is a virtualization environment in which hardware resources are managed by a hypervisor to provide virtual hardware resources to a virtual machine.
- the virtual machines do not have direct access to the underlying hardware resources.
- a host operating system with embedded hypervisor e.g., VMware ESXi®
- Virtual machines including virtual hardware resources are then deployed on the hypervisor.
- a guest operating system is installed in the virtual machine.
- the hypervisor manages the association between the hardware resources of the server hardware and the virtual resources allocated to the virtual machines (e.g., associating physical random access memory (RAM) with virtual RAM).
- RAM physical random access memory
- full virtualization typically, the virtual machine and the guest operating system have no visibility and/or direct access to the hardware resources of the underlying server. Additionally, in full virtualization, a full guest operating system is typically installed in the virtual machine while a host operating system is installed on the server hardware.
- Example full virtualization environments include VMware ESX®, Microsoft Hyper-V®, and Kernel Based Virtual Machine (KVM).
- Paravirtualization is a virtualization environment in which hardware resources are managed by a hypervisor to provide virtual hardware resources to a virtual machine and guest operating systems are also allowed direct access to some or all of the underlying hardware resources of the server (e.g., without accessing an intermediate virtual hardware resource).
- a host operating system e.g., a Linux-based operating system
- a hypervisor e.g., the Xen® hypervisor
- Virtual machines including virtual hardware resources are then deployed on the hypervisor.
- the hypervisor manages the association between the hardware resources of the server hardware and the virtual resources allocated to the virtual machines (e.g., associating physical random access memory (RAM) with virtual RAM).
- the guest operating system installed in the virtual machine is configured also to have direct access to some or all of the hardware resources of the server.
- the guest operating system may be precompiled with special drivers that allow the guest operating system to access the hardware resources without passing through a virtual hardware layer.
- a guest operating system may be precompiled with drivers that allow the guest operating system to access a sound card installed in the server hardware. Directly accessing the hardware (e.g., without accessing the virtual hardware resources of the virtual machine) may be more efficient, may allow for performance of operations that are not supported by the virtual machine and/or the hypervisor, etc.
- Operating system virtualization is also referred to herein as container virtualization.
- operating system virtualization refers to a system in which processes are isolated in an operating system.
- a host operating system is installed on the server hardware.
- the host operating system may be installed in a virtual machine of a full virtualization environment or a paravirtualization environment.
- the host operating system of an operating system virtualization system is configured (e.g., utilizing a customized kernel) to provide isolation and resource management for processes that execute within the host operating system (e.g., applications that execute on the host operating system).
- the isolation of the processes is known as a container.
- Several containers may share a host operating system.
- operating system virtualization provides isolation and resource management capabilities without the resource overhead utilized by a full virtualization environment or a paravirtualization environment.
- the host operating system may be installed in a virtual machine of a full virtualization environment or a paravirtualization environment.
- Example operating system virtualization environments include Linux Containers LXC and LXD, DockerTM, OpenVZTM, etc.
- a data center may include multiple different virtualization environments.
- a data center may include hardware resources that are managed by a full virtualization environment, a paravirtualization environment, and an operating system virtualization environment.
- a workload may be deployed to any of the virtualization environments.
- FIG. 1 depicts an example system 100 constructed in accordance with the teachings of this disclosure for managing a cloud computing platform.
- the example system 100 includes an application director 106 and a cloud manager 138 to manage a cloud computing platform provider 110 as described in more detail below.
- the example system 100 facilitates management of the cloud provider 110 and does not include the cloud provider 110 .
- the system 100 could be included in the cloud provider 110 .
- the cloud computing platform provider 110 provisions virtual computing resources (e.g., virtual machines, or “VMs,” 114 ) that may be accessed by users of the cloud computing platform 110 (e.g., users associated with an administrator 116 and/or a developer 118 ) and/or other programs, software, device. etc.
- virtual computing resources e.g., virtual machines, or “VMs,” 114
- users of the cloud computing platform 110 e.g., users associated with an administrator 116 and/or a developer 118
- An example application 102 of FIG. 1 includes multiple VMs 114 .
- the example VMs 114 of FIG. 1 provide different functions within the application 102 (e.g., services, portions of the application 102 , etc.).
- One or more of the VMs 114 of the illustrated example are customized by an administrator 116 and/or a developer 118 of the application 102 relative to a stock or out-of-the-box (e.g., commonly available purchased copy) version of the services and/or application components.
- the services executing on the example VMs 114 may have dependencies on other ones of the VMs 114 .
- the example cloud computing platform provider 110 may provide multiple deployment environments 112 , for example, for development, testing, staging, and/or production of applications.
- the administrator 116 , the developer 118 , other programs, and/or other devices may access services from the cloud computing platform provider 110 , for example, via REST (Representational State Transfer) APIs (Application Programming Interface) and/or via any other client-server communication protocol.
- REST Representational State Transfer
- Example implementations of a REST API for cloud computing services include a vCloud Administrator CenterTM (vCAC) and/or vRealize AutomationTM (vRA) API and a vCloud DirectorTM API available from VMware, Inc.
- the example cloud computing platform provider 110 provisions virtual computing resources (e.g., the VMs 114 ) to provide the deployment environments 112 in which the administrator 116 and/or the developer 118 can deploy multi-tier application(s).
- virtual computing resources e.g., the VMs 114
- One particular example implementation of a deployment environment that may be used to implement the deployment environments 112 of FIG. 1 is vCloud DataCenter cloud computing services available from VMware, Inc.
- a lighter-weight virtualization is employed by using containers in place of the VMs 114 in the development environment 112 .
- Example containers 114 a are software constructs that run on top of a host operating system without the need for a hypervisor or a separate guest operating system. Unlike virtual machines, the containers 114 a do not instantiate their own operating systems. Like virtual machines, the containers 114 a are logically separate from one another. Numerous containers can run on a single computer, processor system and/or in the same development environment 112 . Also like virtual machines, the containers 114 a can execute instances of applications or programs (e.g., an example application 102 a ) separate from application/program instances executed by the other containers in the same development environment 112 .
- applications or programs e.g., an example application 102 a
- the example application director 106 of FIG. 1 which may be running in one or more VMs, orchestrates deployment of multi-tier applications onto one of the example deployment environments 112 .
- the example application director 106 includes a topology generator 120 , a deployment plan generator 122 , and a deployment director 124 .
- the example topology generator 120 generates a basic blueprint 126 that specifies a logical topology of an application to be deployed.
- the example basic blueprint 126 generally captures the structure of an application as a collection of application components executing on virtual computing resources.
- the basic blueprint 126 generated by the example topology generator 120 for an online store application may specify a web application (e.g., in the form of a Java web application archive or “WAR” file including dynamic web pages, static web pages, Java servlets, Java classes, and/or other property, configuration and/or resources files that make up a Java web application) executing on an application server (e.g., Apache Tomcat application server) that uses a database (e.g., MongoDB) as a data store.
- a web application e.g., in the form of a Java web application archive or “WAR” file including dynamic web pages, static web pages, Java servlets, Java classes, and/or other property, configuration and/or resources files that make up a Java web application
- the term “application” generally refers to a logical deployment unit, including one or more application packages and their dependent middleware and/or operating systems. Applications may be distributed across multiple VMs. Thus, in the example described above, the term “application” refers to the entire online store application, including application server and database components, rather than just the web application itself. In some instances, the application may include the underlying hardware and/or virtual computing hardware utilized to implement the components.
- the example basic blueprint 126 of FIG. 1 may be assembled from items (e.g., templates) from a catalog 130 , which is a listing of available virtual computing resources (e.g., VMs, networking, storage, etc.) that may be provisioned from the cloud computing platform provider 110 and available application components (e.g., software services, scripts, code components, application-specific packages) that may be installed on the provisioned virtual computing resources.
- the example catalog 130 may be pre-populated and/or customized by an administrator 116 (e.g., IT (Information Technology) or system administrator) that enters in specifications, configurations, properties, and/or other details about items in the catalog 130 .
- IT Information Technology
- the example blueprints 126 may define one or more dependencies between application components to indicate an installation order of the application components during deployment. For example, since a load balancer usually cannot be configured until a web application is up and running, the developer 118 may specify a dependency from an Apache service to an application code package.
- the example deployment plan generator 122 of the example application director 106 of FIG. 1 generates a deployment plan 128 based on the basic blueprint 126 that includes deployment settings for the basic blueprint 126 (e.g., virtual computing resources' cluster size, CPU, memory, networks, etc.) and an execution plan of tasks having a specified order in which virtual computing resources are provisioned and application components are installed, configured, and started.
- the example deployment plan 128 of FIG. 1 provides an IT administrator with a process-oriented view of the basic blueprint 126 that indicates discrete actions to be performed to deploy the application.
- Different deployment plans 128 may be generated from a single basic blueprint 126 to test prototypes (e.g., new application versions), to scale up and/or scale down deployments, and/or to deploy the application to different deployment environments 112 (e.g., testing, staging, production).
- the deployment plan 128 is separated and distributed as local deployment plans having a series of tasks to be executed by the VMs 114 provisioned from the deployment environment 112 .
- Each VM 114 coordinates execution of each task with a centralized deployment module (e.g., the deployment director 124 ) to ensure that tasks are executed in an order that complies with dependencies specified in the application blueprint 126 .
- the example deployment director 124 of FIG. 1 executes the deployment plan 128 by communicating with the cloud computing platform provider 110 via a cloud interface 132 to provision and configure the VMs 114 in the deployment environment 112 .
- the example cloud interface 132 of FIG. 1 provides a communication abstraction layer by which the application director 106 may communicate with a heterogeneous mixture of cloud provider 110 and deployment environments 112 .
- the deployment director 124 provides each VM 114 with a series of tasks specific to the receiving VM 114 (herein referred to as a “local deployment plan”). Tasks are executed by the VMs 114 to install, configure, and/or start one or more application components.
- a task may be a script that, when executed by a VM 114 , causes the VM 114 to retrieve and install particular software packages from a central package repository 134 .
- the example deployment director 124 coordinates with the VMs 114 to execute the tasks in an order that observes installation dependencies between VMs 114 according to the deployment plan 128 .
- the application director 106 may be utilized to monitor and/or modify (e.g., scale) the deployment.
- the example cloud manager 138 of FIG. 1 interacts with the components of the system 100 (e.g., the application director 106 and the cloud provider 110 ) to facilitate the management of the resources of the cloud provider 110 .
- the example cloud manager 138 includes a blueprint manager 140 to facilitate the creation and management of multi-machine blueprints and a resource manager 144 to reclaim unused cloud resources.
- the cloud manager 138 may additionally include other components for managing a cloud environment.
- the example blueprint manager 140 of the illustrated example manages the creation of multi-machine blueprints that define the attributes of multiple virtual machines as a single group that can be provisioned, deployed, managed, etc. as a single unit.
- a multi-machine blueprint may include definitions for multiple basic blueprints that make up a service (e.g., an e-commerce provider that includes web servers, application servers, and database servers).
- a basic blueprint is a definition of policies (e.g., hardware policies, security policies, network policies, etc.) for a single machine (e.g., a single virtual machine such as a web server virtual machine and/or container). Accordingly, the blueprint manager 140 facilitates more efficient management of multiple virtual machines and/or containers than manually managing (e.g., deploying) basic blueprints individually. Example management of multi-machine blueprints is described in further detail in conjunction with FIG. 2 .
- the example blueprint manager 140 of FIG. 1 additionally annotates basic blueprints and/or multi-machine blueprints to control how workflows associated with the basic blueprints and/or multi-machine blueprints are executed.
- a workflow is a series of actions and decisions to be executed in a virtual computing platform.
- the example system 100 includes first and second distributed execution manager(s) (DEM(s)) 146 A and 146 B to execute workflows.
- the first DEM 146 A includes a first set of characteristics and is physically located at a first location 148 A.
- the second DEM 146 B includes a second set of characteristics and is physically located at a second location 148 B.
- the location and characteristics of a DEM may make that DEM more suitable for performing certain workflows.
- a DEM may include hardware particularly suited for performance of certain tasks (e.g., high-end calculations), may be located in a desired area (e.g., for compliance with local laws that require certain operations to be physically performed within a country's boundaries), may specify a location or distance to other DEMS for selecting a nearby DEM (e.g., for reducing data transmission latency), etc.
- the example blueprint manager 140 annotates basic blueprints and/or multi-machine blueprints with capabilities that can be performed by a DEM that is labeled with the same or similar capabilities.
- the resource manager 144 of the illustrated example facilitates recovery of cloud computing resources of the cloud provider 110 that are no longer being activity utilized.
- Automated reclamation may include identification, verification and/or reclamation of unused, underutilized, etc. resources to improve the efficiency of the running cloud infrastructure.
- FIG. 2 illustrates an example implementation of the blueprint 126 as a multi-machine blueprint generated by the example blueprint manager 140 of FIG. 1 .
- three example basic blueprints (a web server blueprint 202 , an application server blueprint 204 , and a database (DB) server blueprint 206 ) have been created (e.g., by the topology generator 120 ).
- the web server blueprint 202 , the application server blueprint 204 , and the database server blueprint 206 may define the components of an e-commerce online store.
- the example blueprint manager 140 provides a user interface for a user of the blueprint manager 140 (e.g., the administrator 116 , the developer 118 , etc.) to specify blueprints (e.g., basic blueprints and/or multi-machine blueprints) to be assigned to an instance of a multi-machine blueprint 208 .
- the user interface may include a list of previously generated basic blueprints (e.g., the web server blueprint 202 , the application server blueprint 204 , the database server blueprint 206 , etc.) to allow selection of desired blueprints.
- the blueprint manager 140 combines the selected blueprints into the definition of the multi-machine blueprint 208 and stores information about the blueprints in a multi-machine blueprint record defining the multi-machine blueprint 208 .
- the blueprint manager 140 may additionally include a user interface to specify other characteristics corresponding to the multi-machine blueprint 208 .
- a creator of the multi-machine blueprint 208 may specify a minimum number and a maximum number of each blueprint component of the multi-machine blueprint 208 that may be provisioned during provisioning of the multi-machine blueprint 208 .
- any number of virtual machines e.g., the virtual machines associated with the blueprints in the multi-machine blueprint 208
- the multiple virtual machines corresponding to the multi-machine blueprint 208 may be provisioned based on an instruction to provision the multi-machine blueprint 208 , may be power cycled by an instruction, may be shut down by an instruction, may be booted by an instruction, etc.
- an instruction to provision the multi-machine blueprint 208 may result in the provisioning of a multi-machine service formed from one or more VMs 114 that includes virtualized web server(s) 210 A, virtualized application server(s) 210 B, and virtualized database server(s) 210 C.
- the number of virtual machines and/or containers provisioned for each blueprint may be specified during the provisioning of the multi-machine blueprint 208 (e.g., subject to the limits specified during creation or management of the multi-machine blueprint 208 ).
- the multi-machine blueprint 208 maintains the reference to the basic blueprints 202 , 204 , 206 . Accordingly, changes made to the blueprints (e.g., by a manager of the blueprints different than the manager of the multi-machine blueprint 208 ) may be incorporated into future provisioning of the multi-machine blueprint 208 . Accordingly, an administrator maintaining the source blueprints (e.g., an administrator charged with managing the web server blueprint 202 ) may change or update the source blueprint and the changes may be automatically propagated to the machines provisioned from the multi-machine blueprint 208 .
- an administrator maintaining the source blueprints e.g., an administrator charged with managing the web server blueprint 202
- the changes may be automatically propagated to the machines provisioned from the multi-machine blueprint 208 .
- the updated disk image is utilized when deploying the multi-machine blueprint.
- the blueprints may specify that the machines 210 A, 210 B, 210 C of the multi-machine service 210 provisioned from the multi-machine blueprint 208 operate in different environments.
- some components may be physical machines, some may be on-premise virtual machines, and some may be virtual machines at a cloud service.
- multi-machine blueprints may be generated to provide one or more varied or customized services. For example, if virtual machines deployed in the various States of the United States require different settings, a multi-machine blueprint could be generated for each state.
- the multi-machine blueprints could reference the same build profile and/or disk image, but may include different settings specific to each state.
- the deployment workflow may include an operation to set a locality setting of an operating system to identify a particular state in which a resource is physically located.
- a single disk image may be utilized for multiple multi-machine blueprints reducing the amount of storage space for storing disk images compared with storing a disk image for each customized setting.
- FIG. 3 illustrates an example installation of deployed appliances or virtual appliances (vAs) (e.g., VMs 114 and/or containers 114 a ) and associated virtualized servers acting as hosts for deployment of component servers (e.g., Web server, application server, database server, etc.) for a customer.
- the vAs can be deployed as an automation tool, for example, used to deliver VMs and associated applications for on-premise automation and/or handling of external cloud resources (e.g., Microsoft AzureTM, Amazon Web ServicesTM, etc.).
- an installation 300 includes a load balancer (LB) 310 to assign tasks and/or manage access among a plurality of vAs 320 , 322 , 324 .
- LB load balancer
- Each vA 320 - 324 is a deployed VM 114 and/or container 114 a .
- the vA 320 communicates with a plurality of component or host servers 330 , 332 , 334 , 336 which store components for execution by users (e.g., Web server 210 A with Web components, App server 210 B with application components, DB server 210 C with database components, etc.).
- FIG. 1 load balancer
- component servers 334 , 336 can stem from component server 330 rather than or in addition to directly from the virtual appliance 320 , although the vA 320 can still communicate with such servers 334 , 336 .
- the LB 310 enables the multiple vAs 320 - 324 and multiple servers 330 - 336 to appear as one device to a user. Access to functionality can then be distributed among appliances 320 - 324 by the LB 310 and among servers 330 - 336 by the respective appliance 320 , for example.
- the LB 310 can use least response time, round-robin, and/or other method to balance traffic to vAs 320 - 324 and servers 330 - 336 , for example.
- each vA 320 , 322 , 324 includes a management endpoint 340 , 342 , 344 .
- Each component server 330 , 332 , 334 , 336 includes a management agent 350 , 352 , 354 , 356 .
- the management agents 350 - 356 can communicate with their respective endpoint 340 to facilitate transfer of data, execution of tasks, etc., for example.
- management endpoints 340 , 342 , 344 share a data store, and any management agent 350 - 356 can connect to any management endpoint 340 , 342 , 344 to retrieve a task and/or associated data from the data store.
- management endpoints 340 , 342 , 344 are interconnected via the data store.
- a management endpoint 340 - 342 doubles as an agent 350 - 356 , allowing the endpoint 340 - 344 on a vA 320 - 324 to automate a task on another vA 320 - 324 .
- each agent 350 - 356 maintains a pool of available endpoints 340 - 344 .
- agent 350 - 356 can automatically switch to a different endpoint 340 - 344 , from which the agent 350 - 356 can retrieve execution tasks.
- Agents 350 - 356 connected via different endpoints 340 - 344 can initiate task execution from a plurality of vAs 320 - 324 , even if not currently connected to that vA 320 - 324 because, due to the connection between endpoints 340 - 344 , for example.
- the management agents 350 - 356 synchronize component servers 330 - 336 with the vA 320 - 324 and facilitate host access and associated services (e.g., hostd, ntpd, sfcbd, slpd, wsman, vobd, etc.).
- the management agents 350 - 356 can communicate with their respective endpoint 340 to facilitate transfer of data, execution of tasks, etc., for example.
- the relationship between management endpoint 340 , 342 , 344 and associated management agents 350 , 352 , 354 , 356 can be used to deploy and install software on multiple component machines 330 , 332 , 334 , 336 .
- component servers 330 - 336 can be installed and/or managed even when the vA 320 - 324 and/or its endpoint 340 - 344 are physically restricted from accessing the server 330 - 336 and/or its agent 350 - 356 .
- the agent 350 - 356 polls the endpoint 340 - 344 for work items, so an inbound connection to the component server 330 - 336 can be absent, for example.
- a graphical user interface associated with a front end of the load balancer 310 guides a customer through one or more questions to determine system requirements for the installation 300 .
- the agents 350 - 356 communicate with the endpoint 340 without customer involvement.
- a manager selects an option (e.g., clicks a button, etc.) via the graphical user interface to install a VM 114 and/or container 114 a that is managed through the installation 300 .
- the virtual appliance (vA) 320 is accessing different servers 330 - 336 depending upon what functionality is to be executed.
- agents 350 - 356 are deployed in a same data center as the endpoint 340 to which the agents 350 - 356 are associated.
- the deployment can include a plurality of agent servers 330 - 336 distributed worldwide, and the deployment can be scalable to accommodate additional server(s) with agent(s) to increase throughput and concurrency, for example.
- FIG. 4 illustrates an example implementation of the vA 320 .
- the vA 320 includes a service provisioner 410 , an orchestrator 420 , an event broker 430 , an authentication provider 440 , an internal reverse proxy 450 , and a database 460 .
- the components 410 , 420 , 430 , 440 , 450 , 460 of the vA 320 may be implemented by one or more of the VMs 114 .
- the example service provisioner 410 provides services to provision interfaces (e.g., Web interface, application interface, etc.) for the vA 320 .
- provision interfaces e.g., Web interface, application interface, etc.
- the example orchestrator (e.g., vCO) 420 is an embedded or internal orchestrator that can leverage a provisioning manager, such as the application director 106 and/or cloud manager 138 , to provision VM services but is embedded in the vA 320 .
- a provisioning manager such as the application director 106 and/or cloud manager 138
- the vCO 420 can be used to invoke a blueprint to provision a manager for services.
- Example services can include catalog services, identity services, component registry services, event broker services, IaaS, XaaS, etc.
- Catalog services provide a user interface via which a user can request provisioning of different preset environments (e.g., a VM including an operating system and software and some customization, etc.), for example.
- Identity services facilitate authentication and authorization of users and assigned roles, for example.
- the component registry maintains information corresponding to installed and deployed services (e.g., uniform resource locators for services installed in a VM/vA, etc.), for example.
- the event broker provides a messaging broker for event-based communication, for example.
- the IaaS provisions one or more VMs and/or containers for a customer via the vA 320 .
- the XaaS can extend the provisioning to also request, approve, provision, operate, and decommission any type of catalog items (e.g., storage, applications, accounts, and anything else that the catalog provides as a service).
- the example event broker 430 provides a mechanism to handle tasks which are transferred between services with the orchestrator 420 .
- the example authentication provider 440 e.g., VMware HorizonTM services, etc.
- the components of the vA 320 access each other through REST API calls behind the internal reverse proxy 450 (e.g., a high availability (HA) proxy HAProxy) which provides a high availability load balancer and proxy for Transmission Control Protocol (TCP)- and Hypertext Transfer Protocol (HTTP)-based application requests.
- the proxy 450 forwards communication traffic from within the vA 320 and/or between vAs 320 , 322 , 324 of FIG. 3 to the appropriate component(s) of the vA 320 .
- services access the local host/proxy 450 on a particular port, and the call is masked by the proxy 450 and forwarded to the particular component of the vA 320 . Since the call is masked by the proxy 450 , components can be adjusted within the vA 320 without impacting outside users.
- a cloud computing (e.g., vCACTM, vRATM, etc.) deployment includes one or more vAs 320 - 324 and one or more component servers 330 - 336 (e.g., Microsoft WindowsTM machines, etc.) on which are installed components (e.g., software such as Web services, application services, database services, etc.) that form the IaaS portion of the product.
- vCACTM vCACTM, vRATM, etc.
- component servers 330 - 336 e.g., Microsoft WindowsTM machines, etc.
- installed components e.g., software such as Web services, application services, database services, etc.
- a plurality of component servers 330 - 336 form the installed product, and having to install the IaaS components manually on all of the component servers 330 - 336 is a time-consuming process, involving, among other things, multiple context switches and many opportunities for user misconfiguration of the deployed system.
- manual installation involves installing components on an appliance, downloading an installer, and then visit each server to install the components manually using the installer.
- the installation may not function.
- data entry is required for each manual installation, and mis-typing of the manual data entry can invalidate the entire installation. Further, such a mistake may not be realized until the erroneous installation is deployed, resulting in lost time, money, errors, and inoperable systems. Simplification and automation of this process reduces the time needed and errors involved in setting up a new instance of the cloud computing system.
- installation can be executed on each node from a centralized location via the management agent 350 - 356 installed on each component server 330 - 336 .
- the agent 350 - 356 is installed and registered with the vA 320 . After registration, communication with the vA 320 is authenticated using a client certificate and signed token(s). The vA's 320 root credentials are not persisted on the server 330 - 336 .
- Each instance of the management agent 350 - 356 has a node identifier (ID), which uniquely identifies the agent 330 - 336 in a cluster of machines 330 - 336 forming the cloud deployment.
- ID node identifier
- the agent 330 - 336 After registration, the agent 330 - 336 starts polling the vA 320 in a configurable time interval to obtain commands to be executed. The commands are executed by the corresponding server 330 - 336 , and a result is reported back to the vA 320 by the agent 350 - 356 and can be used for further processing, for example.
- installation of a hybrid system including a plurality of appliances 320 - 324 and component servers 330 - 336 having a plurality of roles can be orchestrated via the management agents 350 - 356 .
- the example installation 300 can be coordinated without manual user action throughout phases of the installation.
- FIG. 5 illustrates a block diagram of an example implementation of the management agent 350 (and/or 352 , 354 , 356 ).
- the management agent 350 includes a communication interface 510 through which the agent 350 can communicate with the endpoint 340 (and/or 342 , 344 ) of the vA 320 (and/or 322 , 324 ).
- the communication interface 510 is a hardware and/or software interface allowing the agent 350 to exchange data, commands, etc., with the endpoint 340 and/or other communication node, for example.
- the example agent 350 also includes an agent control processor 520 .
- the agent control processor 520 executes instructions to control the agent 350 for command and/or other application execution, communication, storage, etc.
- the instructions can be transmitted to the agent control processor 520 via the communication interface 510 and/or via a data storage 530 , for example.
- the example data storage 530 includes a configuration file 540 and a machine identifier 550 .
- the example configuration file 540 can include information such as credentials to authenticate and/or validate the agent 350 to the vA 320 , etc. Credentials can include a certificate (e.g., with a public key and private key for authentication, etc.), a unique identifier, etc.
- the example agent control processor 520 can process instructions, generate communications, etc.
- the example data storage 530 can also include instructions (e.g., computer program code, etc.) to be executed by the agent control processor 520 .
- FIG. 6 illustrates an example data flow diagram showing an exchange of information 600 between the vA 320 , 322 and the component server 330 to install system 300 components including the vA 320 , 322 and component server 330 including management agent 350 .
- a trigger 601 e.g., a user via a vA management webpage, an automated script, etc.
- an installation wizard is also triggered 601 to be run by the vA 320 , for example.
- a management agent 350 (and/or 352 - 356 ) is installed on each component server 330 (and/or 332 - 336 ).
- each server 330 (and/or 332 - 336 ) acknowledges the installation of the management agent 350 (and/or 352 - 356 ).
- the acknowledgement can include an authentication of management agent 350 credentials by the vA 320 , for example.
- Authentication and/or authorization can include an exchange and verification of a certificate, identifier, etc., associated with the management agent 350 and/or its associated server 330 by the vA 320 , for example.
- a cloud-based installation may include one or more vAs 320 - 324 and one or more servers 330 - 336 (e.g., “WindowsTM machines”, etc.) on which a plurality of components (e.g., five, six, seven, ten, etc.) are installed (e.g., applications, database, management, etc.) to form an IaaS in a distributed, high availability environment.
- the management agents 350 - 356 communicate with the management endpoint(s) 340 - 344 to receive commands, execute commands, install software, upgrade an installation at the server 330 - 336 , etc.
- Each management agent 350 - 356 has a node identifier (ID) that uniquely identifies the agent 350 - 356 in a cluster of machines 330 - 336 forming the system 300 .
- ID node identifier
- an address and root credentials of the primary vA 320 are entered so that the agent 350 - 356 can register itself in the vA 320 .
- communication with the vA 320 is authenticated using a client certificate and signed token(s).
- the certificate is used for communication between the agent 350 - 356 and the endpoint 340 , the root credentials of the vA 320 are not persisted on the machines 330 - 336 .
- other appliances are identified to the vA 320 .
- the primary vA 320 is informed of other appliances (e.g., vA 322 , 324 , etc.) to be included in the configuration.
- a role is provided for each server 330 (and/or 332 - 336 ).
- the server 330 is assigned a role as a database server.
- the server 330 is assigned a role as a Web server.
- the server 330 may be assigned a role as an application server, for example.
- the server 330 may be assigned a role as a WindowsTM server, for example.
- Each role is associated with one or more rules that guide and/or establish criteria for the associated role.
- Each rule can be associated with one or more prerequisites for a server 330 - 336 to execute the rule and perform the role.
- rules may specify that there are at least two servers 330 - 336 for each role to provide redundancy and increased availability if one server 330 of a given role is busy or otherwise unavailable, for example.
- the vA 320 evaluates or checks applicable rule prerequisite(s) for a given role to ensure the associated server 330 can perform the role.
- the vA 320 attempts to fix or otherwise remedy any error(s) identified in its evaluation of the server 330 .
- prerequisites can include a determination of whether the load balancer 310 , vA 320 - 324 , and/or component server 330 - 336 , etc., is/are reachable, and an associated fix can include an address look-up and resolution.
- Another prerequisite can include registration of the server 330 - 336 and/or other IaaS node with the vA 320 - 324 , and an associated fix can include registration of the server 330 - 336 and/or other IaaS node with the vA 320 - 324 , such as described above, for example.
- Another fix of a server 330 - 336 error can include downloading and installing a service pack on the server 330 - 336 , for example.
- Another prerequisite can include database (e.g., object-relational database such as Postgres, etc.) access, and an associated fix can include access and log in to the database, for example.
- the server 330 reports back to the vA 320 with an indication of what cannot be automatically fixed. For example, the server 330 generates a report of what prerequisite errors were fixed and what prerequisite errors were identified and not fixed, and sends 616 the report to the vA 320 .
- errors that were identified but not fixed in the prerequisite checks are handled based on data entry at the vA 320 .
- the user via the trigger 601 , the user can provide data entry and/or one or more automated scripts can be executed to provide answer(s), setting(s), address(es), password(s), credential(s), etc., to satisfy missing prerequisite information for the server 330 at the vA 320 .
- the data entry 618 can be guided by a wizard for component address(es), username(s), identity of service(s) to be executed, identification of port(s) (e.g., default or custom), database credentials/address, etc.
- a user provides data entry which is recorded and converted into a script for subsequent automated execution, for example.
- the vA 320 asks to configure the load balancer (LB) 310 with an external device/system 603 .
- the LB 310 can be configured with information regarding connected components 320 - 324 , 330 - 336 , etc., and their capabilities and ability to distribute workload for a plurality of tasks.
- rules and/or other prerequisites are sent to server(s) 330 - 336 along with other instructions to configure a particular customer environment.
- install manager service etc.
- a snapshot of the installation is obtained.
- the snapshot forms a recovery point.
- the snapshot stores the configuration for later use, restoration, error recovery, etc.
- the installation is executed.
- the installation is executed at the vA 320 (e.g., the primary appliance).
- the installation occurs at the server 330 (e.g., via the management agent 350 ).
- the installation occurs at the vA 322 (e.g., a secondary appliance).
- a sequence of commands is built and targeted for an environment including server(s) 330 .
- the commands are triggered for orchestration of services via the management agent(s) 350 .
- the central or primary vA 320 does not have access to individual nodes but instead can be accessed by the management agent(s) 350 - 356 of the respective server(s) 330 - 336 , which acts to execute installation instructions from the vA 320 .
- the vA 320 awaits acknowledgement from the agent(s) 350 - 356 .
- FIG. 7 illustrates an example schedule of components 700 for installation in an example system.
- the example vA 320 includes information for installation 632 - 638 of multiple components after prerequisite checks, validation, and snapshot generation 602 - 630 , as described above with respect to FIG. 6 .
- vA 320 referred to in the example as the primary or main appliance vA1
- the vA 320 also interacts with a plurality of secondary appliances 322 - 324 , referred to in the example as vA 2 and vA 3 .
- the vA1 320 has a queue of commands 702 (e.g., install commands, joinder commands, etc.) established via its management endpoint 340 .
- the commands 702 e.g., install vA1, install identity vA1, install web component on server1 (e.g., primary WindowsTM server), install manager service component on server4, install DEM on server3, etc., then join appliance to cluster vA2, join appliance to cluster vA3, etc., then install web on server2, install manager service on server3, etc., install management agent on server5, etc.
- the commands 702 e.g., install vA1, install identity vA1, install web component on server1 (e.g., primary WindowsTM server), install manager service component on server4, install DEM on server3, etc., then join appliance to cluster vA2, join appliance to cluster vA3, etc., then install web on server2, install manager service on server3, etc., install management agent on server5, etc.
- Some or all of the commands 702 have certain dependency 704 on other command(s) 702 . That is, one command 702 may depend on nothing, but another command 702 may depend on a prior command 702 . Thus, while some commands can be executed in parallel and/or otherwise asynchronously, other commands must be executed serially and/or otherwise synchronously when their dependency 704 has been satisfied, for example.
- the vA1 320 builds and then deploys the installation including copies of involved components.
- Each component server 330 - 336 polls the vA1 320 for work.
- the endpoint 340 of the vA 320 examines the queue 702 to determine a next task. If the next task has no dependency 704 , the endpoint 350 transmits the task to the server 330 - 336 via its agent 350 - 356 for execution. If the task in the queue 702 has a dependency 704 , then the endpoint 340 determines whether the dependent task has been completed, started, or not yet started, for example. If the dependent task has not yet started, for example, then the endpoint 340 will not pass the associated task from the queue 702 to the server 330 - 336 .
- Each task in the queue 702 is also associated with and/or otherwise intended for a server 706 .
- server1 queries the vA1 to determine whether the queue 702 includes any task for server1.
- the endpoint 340 of the appliance vA1 determines that yes, task 3 in the queue 702 is for server1 (e.g., installation of web server on server1, etc.).
- task 3 depends on the execution of task 2 by the vA1
- task 2 depends on the execution of task 1 by the vA1.
- the endpoint vA1 does not release task 3 to the agent of server1 until the vA1 has executed tasks 1 and 2.
- Task 4 however, has no dependency 704 , so the endpoint can release task 4 to the agent of server4 when the agent polls for its task.
- FIG. 7 once task execution reaches task 7, all remaining tasks depend on task 5, which has then been completed.
- tasks 7-12 can be dispensed and/or otherwise released to be executed in parallel by their respective component.
- the vA 320 publishes the commands 702 to available servers 330 - 336 (e.g., via communication between the endpoint 340 and agents 350 - 356 ) and is aware of the sequence 702 and dependency 704 of the commands.
- the servers 330 - 336 accept and execute the commands without knowledge of the sequence 702 and dependency 704 .
- the server's agent 350 - 356 facilitates a download of an installer from the vA 320 , which the server 330 - 336 extracts and executes to install the component.
- the vA 320 is aware of the servers 330 - 336 available to execute tasks but does not instruct the servers 330 - 36 to execute tasks. Rather, the servers 330 - 336 poll the queue of instructions 702 , and the vA 320 checks whether dependencies 704 have been met to release the task to be performed by the server 330 - 336 and/or other component.
- example implementations of the example cloud computing system 100 and virtual machine installation 300 are illustrated in FIGS. 1-7 , one or more of the elements, processes and/or devices illustrated in FIGS. 1-7 may be combined, divided, re-arranged, omitted, eliminated and/or implemented in any other way.
- 1-7 can be implemented by one or more analog or digital circuit(s), logic circuits, programmable processor(s), application specific integrated circuit(s) (ASIC(s)), programmable logic device(s) (PLD(s)) and/or field programmable logic device(s) (FPLD(s)).
- ASIC application specific integrated circuit
- PLD programmable logic device
- FPLD field programmable logic device
- the example systems 100 and/or 300 of FIGS. 1-7 may include one or more elements, processes and/or devices in addition to, or instead of, those illustrated in FIGS. 1-7 , and/or may include more than one of any or all of the illustrated elements, processes and devices.
- FIGS. 8-9 Flowcharts representative of example machine readable instructions that may be executed to deploy and manage the example application director 106 , example cloud provider 110 , example cloud manager 138 , example distributed execution managers 146 A, 146 B, example multi-machine service 210 , example load balancer 310 , example virtual appliances 320 - 324 , example component servers 330 - 336 , example management endpoints 340 - 344 , example management agents 350 - 356 , and/or, more generally, the example systems 100 and/or 300 of FIGS. 1-7 are shown in FIGS. 8-9 .
- the machine readable instructions implement programs for execution by a processor such as the processor 1012 shown in the example processor platform 1000 discussed below in connection with FIG. 10 .
- the programs may be embodied in software stored on a tangible computer readable storage medium such as a CD-ROM, a floppy disk, a hard drive, a digital versatile disk (DVD), a Blu-ray disk, or a memory associated with the processor 1012 , but the entire program and/or parts thereof could alternatively be executed by a device other than the processor 1012 and/or embodied in firmware or dedicated hardware.
- a tangible computer readable storage medium such as a CD-ROM, a floppy disk, a hard drive, a digital versatile disk (DVD), a Blu-ray disk, or a memory associated with the processor 1012 , but the entire program and/or parts thereof could alternatively be executed by a device other than the processor 1012 and/or embodied in firmware or dedicated hardware.
- FIGS. 8-9 many other methods of deploying, managing, and updating workload domains in accordance with the teachings of this disclosure may alternatively be used. For example, the order of execution of the blocks may be changed, and/or some of the
- FIGS. 8-9 may be implemented using coded instructions (e.g., computer and/or machine readable instructions) stored on a tangible computer readable storage medium such as a hard disk drive, a flash memory, a read-only memory (ROM), a compact disk (CD), a digital versatile disk (DVD), a cache, a random-access memory (RAM) and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information).
- a tangible computer readable storage medium such as a hard disk drive, a flash memory, a read-only memory (ROM), a compact disk (CD), a digital versatile disk (DVD), a cache, a random-access memory (RAM) and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information).
- tangible computer readable storage medium and “tangible machine readable storage medium” are used interchangeably.
- the example processes of FIGS. 8-9 may be implemented using coded instructions (e.g., computer and/or machine readable instructions) stored on a non-transitory computer and/or machine readable medium such as a hard disk drive, a flash memory, a read-only memory, a compact disk, a digital versatile disk, a cache, a random-access memory and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information).
- coded instructions e.g., computer and/or machine readable instructions
- a non-transitory computer and/or machine readable medium such as a hard disk drive, a flash memory, a read-only memory, a compact disk, a digital versatile disk, a cache, a random-access memory and/or any other storage device or storage disk in which information is stored
- non-transitory computer readable medium is expressly defined to include any type of computer readable storage device and/or storage disk and to exclude propagating signals and to exclude transmission media.
- phrase “at least” when used as the transition term in a preamble of a claim, it is open-ended in the same manner as the term “comprising” is open ended.
- Comprising and all other variants of “comprise” are expressly defined to be open-ended terms. Including and all other variants of “include” are also defined to be open-ended terms. In contrast, the term consisting and/or other forms of consist are defined to be close-ended terms.
- FIG. 8 depicts a flowchart representative of computer readable instructions that may be executed to implement the example infrastructure installation 300 .
- An example program 800 is illustrated in FIG. 8 .
- the first, primary, or main vA 320 is deployed (e.g., triggered by a user, an automated script, an event, etc.).
- the installation wizard is also triggered to be run by the vA 320 , for example.
- a cloud-based installation may include one or more vAs 320 - 324 and one or more servers 330 - 336 (e.g., “WindowsTM machines”, etc.) on which a plurality of components (e.g., five, six, seven, ten, etc.) are installed (e.g., applications, database, management, etc.) to form an IaaS in a distributed, high availability environment.
- the management agents 350 - 356 communicate with the management endpoint(s) 340 - 344 to receive commands, execute commands, install software, upgrade an installation at the server 330 - 336 , etc.
- the system 300 can also include one or more devices such as a load balancer 310 , etc.
- the management agent 350 - 356 is installed on each component server 330 - 336 .
- Each server 330 - 336 acknowledges the installation of the management agent 350 - 356 .
- the acknowledgement can include an authentication of management agent 350 credentials by the vA 320 , for example.
- Authentication and/or authorization can include an exchange and verification of a certificate, identifier, etc., associated with the management agent 350 and/or its associated server 330 by the vA 320 , for example.
- each management agent 350 - 356 has a node identifier (ID) that uniquely identifies the agent 350 - 356 in a cluster of machines 330 - 336 forming the system 300 .
- ID node identifier
- an address and root credentials of the primary vA 320 are entered so that the agent 350 - 356 can register itself in the vA 320 .
- communication with the vA 320 is authenticated using a self-signed certificate.
- the self-signed certificate is used for communication between the agent 350 - 356 and the endpoint 340 , the root credentials of the vA 320 are not persisted on the machines 330 - 336 .
- a role is provided for each server 330 (and/or 332 - 336 ).
- the server 330 is assigned a role as a database server.
- the server 330 is assigned a role as a Web server.
- the server 330 may be assigned a role as an application server, for example.
- the server 330 may be assigned a role as a WindowsTM server, for example.
- Each role is associated with one or more rules that guide and/or establish criteria for the associated role.
- Each rule can be associated with one or more prerequisites for a server 330 - 336 to execute the rule and perform the role.
- rules may specify that there are at least two servers 330 - 336 for each role to provide redundancy and increased availability if one server 330 of a given role is busy or otherwise unavailable, for example.
- the vA 320 evaluates or checks applicable prerequisite(s) for a given role to ensure the associated server 330 can perform the role.
- prerequisites can include a) a determination of whether the load balancer 310 , vA 320 - 324 , and/or component server 330 - 336 , etc., is/are reachable; b) registration of the server 330 - 336 and/or other IaaS node with the vA 320 - 324 ; c) presence of a minimum software and/or firmware version; d) database (e.g., object-relational database such as Postgres, etc.) access; etc.
- object-relational database such as Postgres, etc.
- an error is identified in the prerequisite evaluation. For example, a prerequisite may not be satisfied and may need to be addressed before installation can continue. If no error is identified, then control advances to block 820 to evaluate whether prerequisites remain to be evaluated. If an error is identified, then, at block 814 , the error is evaluated to determine whether or not the error is automatically fixable by the vA 320 and/or other connected component. If the error is fixable, then, at block 816 , the vA 320 automatically fixes the error. If the error is not automatically fixable, then, at block 818 , the error is logged (e.g., included in a report, etc.).
- prerequisite(s) are evaluated for the role to determine whether or not additional prerequisite(s) remain to be evaluated. If prerequisite(s) remain to be evaluated, the control reverts to block 810 to evaluate the prerequisite(s) for the role(s).
- the error log is reviewed to identify errors logged during prerequisite analysis. If no error has been logged, then, at block 828 , a recovery point (e.g., a snapshot) is created for the system 300 configuration. However, if one or more errors are identified, then, at block 824 , error resolution is attempted. For example, data entry can be provided via a user and/or one or more automated scripts to provide answer(s), setting(s), address(es), password(s), credential(s), etc., to satisfy missing prerequisite information for the server 330 at the vA 320 .
- a recovery point e.g., a snapshot
- error resolution is attempted. For example, data entry can be provided via a user and/or one or more automated scripts to provide answer(s), setting(s), address(es), password(s), credential(s), etc., to satisfy missing prerequisite information for the server 330 at the vA 320 .
- the data entry 618 can be guided by a wizard for component address(es), username(s), identity of service(s) to be executed, identification of port(s) (e.g., default or custom), database credentials/address, etc.
- a user provides data entry which is recorded and converted into a script for subsequent automated execution, for example.
- repeated data entry at block 824 can be used to adjust network settings and/or other change in data to then repeat validation at block 826 , for example. If validation is successful, then the installation, configuration, and validation process is complete after the process has been conducted for each logged error (block 822 ).
- recovery point is created.
- the recovery point can be formed from a snapshot taken of the installation, for example.
- the snapshot stores the configuration for later use, restoration, error recovery, etc.
- the installation is executed.
- the installation is executed at the vA 320 (e.g., the primary appliance).
- the installation occurs at the server 330 (e.g., via the management agent 350 ).
- the installation also occurs at the vA 322 (e.g., a secondary appliance).
- a sequence of commands is built and targeted for particular server(s) 330 .
- the commands are triggered for orchestration of services via the management agent(s) 350 .
- the central or primary vA 320 does not have access to individual nodes but instead accesses the management agent(s) 350 - 356 of the respective server(s) 330 - 336 , which acts to execute installation instructions from the vA 320 .
- the vA 320 awaits acknowledgement from the agent(s) 350 - 356 .
- FIG. 9 illustrates an example implementation of executing the installation at block 830 of the example flow diagram of FIG. 8 .
- the server 330 - 336 polls the vA 320 for a task to be executed (e.g., via communication between the management agent 350 - 356 and the management endpoint 340 , etc.).
- the vA 320 reviews its queue of tasks 702 to be executed (e.g., install commands, joinder commands, etc.) established via the management endpoint 340 to identify a task 702 for the requesting server 330 - 336 .
- an identified task is evaluated to identify a dependency. That is, some or all of the tasks 702 have a dependency 704 on other task(s) 702 . That is, one task 702 may depend on nothing, but another task 702 may depend on a prior task 702 . Thus, while some tasks 702 can be executed in parallel and/or otherwise asynchronously, other tasks 702 are to be executed serially and/or otherwise synchronously when their dependency 704 has been satisfied, for example.
- the dependency 704 is evaluated to determine whether or not the dependency has been met. If the dependency 704 has been satisfied, then, at block 910 , the task 702 is deployed by the vA 320 for execution by the server 330 - 336 . However, if the dependency 704 has not been met, then, at block 914 , the vA 320 does not release the task 702 to the server 330 - 336 until the dependency 704 has been satisfied.
- the server 330 - 336 queries the vA 320 to determine whether the queue 702 includes any task for the server 330 - 336 .
- the endpoint 340 of the vA 320 determines that a task in the queue 702 is for the server 330 - 336 .
- the task 702 depends on the execution of another task.
- the endpoint 340 of the vA 320 does not release the task 702 to the agent 350 - 356 of the server 330 - 36 until the vA 320 has executed the dependent task.
- the endpoint 340 can release the task to the agent 350 - 356 of the server 330 - 336 when the agent 350 - 356 polls the endpoint 340 for its task 702 .
- FIG. 10 is a block diagram of an example processor platform 1000 capable of executing the instructions of FIGS. 8-9 to implement the example systems, operation, and management of FIGS. 1-7 .
- the processor platform 1000 of the illustrated example includes a processor 1012 .
- the processor 1012 of the illustrated example is hardware.
- the processor 1012 can be implemented by one or more integrated circuits, logic circuits, microprocessors or controllers from any desired family or manufacturer.
- the processor 1012 of the illustrated example includes a local memory 1013 (e.g., a cache), and executes instructions to implement the example systems 100 , 300 or portions thereof, such as the vA 320 - 324 , component server 330 - 336 , management endpoint 340 - 344 , and management agent 350 - 356 .
- the processor 1012 of the illustrated example is in communication with a main memory including a volatile memory 1014 and a non-volatile memory 1016 via a bus 1018 .
- the volatile memory 1014 may be implemented by Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM) and/or any other type of random access memory device.
- the non-volatile memory 1016 may be implemented by flash memory and/or any other desired type of memory device. Access to the main memory 1014 , 1016 is controlled by a memory controller.
- the processor platform 1000 of the illustrated example also includes an interface circuit 1020 .
- the interface circuit 1020 may be implemented by any type of interface standard, such as an Ethernet interface, a universal serial bus (USB), and/or a PCI express interface.
- one or more input devices 1022 are connected to the interface circuit 1020 .
- the input device(s) 1022 permit(s) a user to enter data and commands into the processor 1012 .
- the input device(s) can be implemented by, for example, an audio sensor, a microphone, a keyboard, a button, a mouse, a touchscreen, a track-pad, a trackball, isopoint and/or a voice recognition system.
- One or more output devices 1024 are also connected to the interface circuit 1020 of the illustrated example.
- the output devices 1024 can be implemented, for example, by display devices (e.g., a light emitting diode (LED), an organic light emitting diode (OLED), a liquid crystal display, a cathode ray tube display (CRT), a touchscreen, a tactile output device, a printer and/or speakers).
- the interface circuit 1020 of the illustrated example thus, typically includes a graphics driver card, a graphics driver chip or a graphics driver processor.
- the interface circuit 1020 of the illustrated example also includes a communication device such as a transmitter, a receiver, a transceiver, a modem and/or network interface card to facilitate exchange of data with external machines (e.g., computing devices of any kind) via a network 1026 (e.g., an Ethernet connection, a digital subscriber line (DSL), a telephone line, coaxial cable, a cellular telephone system, etc.).
- a communication device such as a transmitter, a receiver, a transceiver, a modem and/or network interface card to facilitate exchange of data with external machines (e.g., computing devices of any kind) via a network 1026 (e.g., an Ethernet connection, a digital subscriber line (DSL), a telephone line, coaxial cable, a cellular telephone system, etc.).
- DSL digital subscriber line
- the processor platform 1000 of the illustrated example also includes one or more mass storage devices 1028 for storing software and/or data.
- mass storage devices 1028 include flash devices, floppy disk drives, hard drive disks, optical compact disk (CD) drives, optical Blu-ray disk drives, RAID systems, and optical digital versatile disk (DVD) drives.
- Coded instructions 1032 representative of the example machine readable instructions of FIGS. 8-9 may be stored in the mass storage device 1028 , in the volatile memory 1014 , in the non-volatile memory 1016 , and/or on a removable tangible computer readable storage medium such as a CD or DVD.
- the processor 1012 can be used to implement the virtual appliance 320 (and vAs 322 - 324 ) and the component server 330 (and servers 332 - 336 ) and their components including the service provisioner 410 , orchestrator 420 , event broker 430 , authentication provider 440 , proxy 450 , management endpoint 340 , management agent 350 , communication interface 510 , agent control processor 520 , data storage 530 , etc.
- the above disclosed methods, apparatus and articles of manufacture facilitate installation of a virtual appliance and associated component servers as an IaaS in a distributed environment such as a cloud computing environment and management of agents in the distributed environment. Examples disclosed herein facilitate self-evaluation and installation of servers and agents without further user intervention or cloud oversight.
- each node can be facilitated from a centralized location via a management endpoint and management agents running on each component server.
- the management agents are registered with the virtual appliance, and further communication between the agent and the appliance is authenticated using a self-signed certificate.
- the appliance's root credentials are not persisted on the individual component servers.
- Each instance of the management agent has a node identifier, which uniquely identifies the node in the cluster of machines forming the infrastructure deployment.
- the management agent starts polling the virtual appliance in configurable time for commands to be executed. The commands are executed by the server(s), and the result(s) is/are reported back to the virtual appliance and can be used for further processing.
- the example apparatus includes a first virtual appliance including a management endpoint, the first virtual appliance to organize tasks to be executed to install a computing infrastructure.
- the example apparatus includes a first component server to execute tasks, the component server including a management agent to communicate with the management endpoint to receive a task to be executed to install the computing infrastructure, the first virtual appliance to associate a role with the first component server and to determine whether the first component server satisfies a prerequisite associated with the role, the first virtual appliance to facilitate addressing an error when the first component server is determined not to satisfy the prerequisite.
- the first virtual appliance of the example apparatus is to facilitate addressing the error by at least one of a) automatically fixing the error such that the first component server satisfies the prerequisite or b) facilitating data entry to resolve the error.
- the first virtual appliance of the example apparatus is to validate the first component server to confirm resolution of the error.
- the first virtual appliance of the example apparatus is to capture a snapshot as a recovery point for installation of the computing infrastructure.
- the first virtual appliance of the example apparatus is to facilitate installation of a service on the first component server by deploying the task to the first component server.
- the management agent of the first component server of the example apparatus is to poll the management endpoint of the first virtual appliance to request the task.
- the first virtual appliance of the example apparatus is to evaluate a dependency associated with the task and evaluate whether an identified dependency is satisfied or unsatisfied, and wherein the first virtual appliance is to deploy the task to the first component server when no unsatisfied dependency is found and is to hold the task when an unsatisfied dependency is found.
- the example method includes deploying, by executing an instruction with a processor, a first virtual appliance, the first virtual appliance including a management endpoint, the first virtual appliance to organize tasks to be executed to install a computing infrastructure.
- the example method includes installing, by executing an instruction with the processor, a first component server to execute tasks, the component server including a management agent to communicate with the management endpoint.
- the example method includes associating, by executing an instruction with the processor, the first virtual appliance with a role.
- the example method includes determining, by executing an instruction with the processor, whether the first component server satisfies a prerequisite associated with the role.
- the example method includes facilitating, by executing an instruction with the processor, addressing of an error by the first virtual appliance when the first component server is determined not to satisfy the prerequisite.
- facilitating addressing of the error includes facilitating addressing of the error by at least one of a) automatically fixing the error such that the first component server satisfies the prerequisite or b) facilitating data entry to resolve the error.
- the method further includes validating, by executing an instruction with the processor, the first component server to confirm resolution of the error.
- the method further includes capturing, by executing an instruction with the processor, a snapshot as a recovery point for installation of the computing infrastructure.
- the method further includes facilitating, by executing an instruction with the processor, installation of a service on the first component server by deploying the task to the first component server.
- the method further includes polling, by executing an instruction with the processor, of the management endpoint of the first virtual appliance by the management agent of the first component server to request the task.
- the method further includes evaluating, by executing an instruction with the processor, a dependency associated with the task; evaluating, by executing an instruction with the processor, whether an identified dependency is satisfied or unsatisfied; deploying, by executing an instruction with the processor, the task to the first component server when no unsatisfied dependency is found; and holding, by executing an instruction with the processor, the task when an unsatisfied dependency is found.
- Certain examples provide a computer readable storage medium including instructions that, when executed, cause a machine to at least: deploy a first virtual appliance, the first virtual appliance including a management endpoint, the first virtual appliance to organize tasks to be executed to install a computing infrastructure; install a first component server to execute tasks, the component server including a management agent to communicate with the management endpoint; associate the first virtual appliance with a role; determine whether the first component server satisfies a prerequisite associated with the role; and facilitate addressing of an error by the first virtual appliance when the first component server is determined not to satisfy the prerequisite.
- the instructions when executed, cause the machine to facilitate addressing of the error including facilitating addressing of the error by at least one of a) automatically fixing the error such that the first component server satisfies the prerequisite or b) facilitating data entry to resolve the error.
- the instructions when executed, further cause the machine to validate the first component server to confirm resolution of the error.
- the instructions when executed, further cause the machine to facilitate installation of a service on the first component server by deploying the task to the first component server.
- the instructions when executed, further cause the machine to facilitate polling of the management endpoint of the first virtual appliance by the management agent of the first component server to request the task.
- the instructions when executed, further cause the machine to: evaluate a dependency associated with the task; evaluate whether an identified dependency is satisfied or unsatisfied; deploy the task to the first component server when no unsatisfied dependency is found; and hold the task when an unsatisfied dependency is found.
Landscapes
- Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Computer Security & Cryptography (AREA)
- Stored Programmes (AREA)
- Automation & Control Theory (AREA)
Abstract
Description
- This patent claims priority as a continuation to U.S. Non-Provisional application Ser. No. 15/370,477 entitled “SYSTEMS AND METHODS TO ORCHESTRATE INFRASTRUCTURE INSTALLATION OF A HYBRID SYSTEM”, which was filed on Dec. 6, 2016, and is hereby incorporated herein by reference in its entirety.
- The present disclosure relates generally to cloud computing and, more particularly, to methods and apparatus to orchestrate infrastructure installation of a hybrid system in a distributed environment such as a cloud computing environment.
- Virtualizing computer systems provides benefits such as an ability to execute multiple computer systems on a single hardware computer, replicating computer systems, moving computer systems among multiple hardware computers, and so forth.
- “Infrastructure-as-a-Service” (also commonly referred to as “IaaS”) generally describes a suite of technologies provided as an integrated solution to allow for elastic creation of a virtualized, networked, and pooled computing platform (sometimes referred to as a “cloud computing platform”). Enterprises may use IaaS as a business-internal organizational cloud computing platform (sometimes referred to as a “private cloud”) that gives an application developer access to infrastructure resources, such as virtualized servers, storage, and networking resources. By providing ready access to the hardware resources required to run an application, the cloud computing platform enables developers to build, deploy, and manage the lifecycle of a web application (or any other type of networked application) at a greater scale and at a faster pace than ever before.
- Cloud computing environments may include many processing units (e.g., servers). Other components of a cloud computing environment include storage devices, networking devices (e.g., switches), etc. Current cloud computing environment configuration relies on much manual user input and configuration to install, configure, and deploy the components of the cloud computing environment.
-
FIG. 1 depicts an example system constructed in accordance with the teachings of this disclosure for managing a cloud computing platform. -
FIG. 2 illustrates an example generation of a multi-machine blueprint by the example blueprint manager ofFIG. 1 . -
FIG. 3 illustrates an example installation of deployed virtual machines and associated servers acting as hosts for deployment of component servers for a customer. -
FIG. 4 illustrates an example implementation of a virtual appliance. -
FIG. 5 illustrates a block diagram of an example implementation of a management agent. -
FIG. 6 illustrates an example data flow diagram showing an exchange of information between an appliance and a component server for installation. -
FIG. 7 illustrates an example schedule of components for installation in an example system. -
FIGS. 8-9 depict a flowcharts representative of computer readable instructions that may be executed to implement example infrastructure installation. -
FIG. 10 is a block diagram of an example processing platform capable of executing the example machine-readable instructions ofFIGS. 8-9 . - Cloud computing is based on the deployment of many physical resources across a network, virtualizing the physical resources into virtual resources, and provisioning the virtual resources to perform cloud computing services and applications. Example systems for virtualizing computer systems are described in U.S. patent application Ser. No. 11/903,374, entitled “METHOD AND SYSTEM FOR MANAGING VIRTUAL AND REAL MACHINES,” filed Sep. 21, 2007, and granted as U.S. Pat. No. 8,171,485, U.S. Provisional Patent Application No. 60/919,965, entitled “METHOD AND SYSTEM FOR MANAGING VIRTUAL AND REAL MACHINES,” filed Mar. 26, 2007, and U.S. Provisional Patent Application No. 61/736,422, entitled “METHODS AND APPARATUS FOR VIRTUALIZED COMPUTING,” filed Dec. 12, 2012, all three of which are hereby incorporated herein by reference in their entirety.
- Cloud computing platforms may provide many powerful capabilities for performing computing operations. However, taking advantage of these computing capabilities manually may be complex and/or require significant training and/or expertise. Prior techniques to providing cloud computing platforms and services often require customers to understand details and configurations of hardware and software resources to establish and configure the cloud computing platform. Methods and apparatus disclosed herein facilitate the management of virtual machine resources in cloud computing platforms.
- A virtual machine is a software computer that, like a physical computer, runs an operating system and applications. An operating system installed on a virtual machine is referred to as a guest operating system. Because each virtual machine is an isolated computing environment, virtual machines (VMs) can be used as desktop or workstation environments, as testing environments, to consolidate server applications, etc. Virtual machines can run on hosts or clusters. The same host can run a plurality of VMs, for example.
- As disclosed in detail herein, methods and apparatus disclosed herein provide for automation of management tasks such as provisioning multiple virtual machines for a multiple-machine computing system (e.g., a group of servers that inter-operate), linking provisioned virtual machines and tasks to desired systems to execute those virtual machines or tasks, and/or reclaiming cloud computing resources that are no longer in use. The improvements to cloud management systems (e.g., the vCloud Automation Center (vCAC) from VMware®, the vRealize Automation Cloud Automation Software from VMware®), interfaces, portals, etc. disclosed herein may be utilized individually and/or in any combination. For example, all or a subset of the described improvements may be utilized.
- As used herein, availability refers to the level of redundancy required to provide continuous operation expected for the workload domain. As used herein, performance refers to the computer processing unit (CPU) operating speeds (e.g., CPU gigahertz (GHz)), memory (e.g., gigabytes (GB) of random access memory (RAM)), mass storage (e.g., GB hard drive disk (HDD), GB solid state drive (SSD)), and power capabilities of a workload domain. As used herein, capacity refers to the aggregate number of resources (e.g., aggregate storage, aggregate CPU, etc.) across all servers associated with a cluster and/or a workload domain. In examples disclosed herein, the number of resources (e.g., capacity) for a workload domain is determined based on the redundancy, the CPU operating speed, the memory, the storage, the security, and/or the power requirements selected by a user. For example, more resources are required for a workload domain as the user-selected requirements increase (e.g., higher redundancy, CPU speed, memory, storage, security, and/or power options require more resources than lower redundancy, CPU speed, memory, storage, security, and/or power options).
- Example Virtualization Environments
- Many different types of virtualization environments exist. Three example types of virtualization environment are: full virtualization, paravirtualization, and operating system virtualization.
- Full virtualization, as used herein, is a virtualization environment in which hardware resources are managed by a hypervisor to provide virtual hardware resources to a virtual machine. In a full virtualization environment, the virtual machines do not have direct access to the underlying hardware resources. In a typical full virtualization environment, a host operating system with embedded hypervisor (e.g., VMware ESXi®) is installed on the server hardware. Virtual machines including virtual hardware resources are then deployed on the hypervisor. A guest operating system is installed in the virtual machine. The hypervisor manages the association between the hardware resources of the server hardware and the virtual resources allocated to the virtual machines (e.g., associating physical random access memory (RAM) with virtual RAM). Typically, in full virtualization, the virtual machine and the guest operating system have no visibility and/or direct access to the hardware resources of the underlying server. Additionally, in full virtualization, a full guest operating system is typically installed in the virtual machine while a host operating system is installed on the server hardware. Example full virtualization environments include VMware ESX®, Microsoft Hyper-V®, and Kernel Based Virtual Machine (KVM).
- Paravirtualization, as used herein, is a virtualization environment in which hardware resources are managed by a hypervisor to provide virtual hardware resources to a virtual machine and guest operating systems are also allowed direct access to some or all of the underlying hardware resources of the server (e.g., without accessing an intermediate virtual hardware resource). In a typical paravirtualization system, a host operating system (e.g., a Linux-based operating system) is installed on the server hardware. A hypervisor (e.g., the Xen® hypervisor) executes on the host operating system. Virtual machines including virtual hardware resources are then deployed on the hypervisor. The hypervisor manages the association between the hardware resources of the server hardware and the virtual resources allocated to the virtual machines (e.g., associating physical random access memory (RAM) with virtual RAM). In paravirtualization, the guest operating system installed in the virtual machine is configured also to have direct access to some or all of the hardware resources of the server. For example, the guest operating system may be precompiled with special drivers that allow the guest operating system to access the hardware resources without passing through a virtual hardware layer. For example, a guest operating system may be precompiled with drivers that allow the guest operating system to access a sound card installed in the server hardware. Directly accessing the hardware (e.g., without accessing the virtual hardware resources of the virtual machine) may be more efficient, may allow for performance of operations that are not supported by the virtual machine and/or the hypervisor, etc.
- Operating system virtualization is also referred to herein as container virtualization. As used herein, operating system virtualization refers to a system in which processes are isolated in an operating system. In a typical operating system virtualization system, a host operating system is installed on the server hardware. Alternatively, the host operating system may be installed in a virtual machine of a full virtualization environment or a paravirtualization environment. The host operating system of an operating system virtualization system is configured (e.g., utilizing a customized kernel) to provide isolation and resource management for processes that execute within the host operating system (e.g., applications that execute on the host operating system). The isolation of the processes is known as a container. Several containers may share a host operating system. Thus, a process executing within a container is isolated the process from other processes executing on the host operating system. Thus, operating system virtualization provides isolation and resource management capabilities without the resource overhead utilized by a full virtualization environment or a paravirtualization environment. Alternatively, the host operating system may be installed in a virtual machine of a full virtualization environment or a paravirtualization environment. Example operating system virtualization environments include Linux Containers LXC and LXD, Docker™, OpenVZ™, etc.
- In some instances, a data center (or pool of linked data centers) may include multiple different virtualization environments. For example, a data center may include hardware resources that are managed by a full virtualization environment, a paravirtualization environment, and an operating system virtualization environment. In such a data center, a workload may be deployed to any of the virtualization environments.
-
FIG. 1 depicts anexample system 100 constructed in accordance with the teachings of this disclosure for managing a cloud computing platform. Theexample system 100 includes anapplication director 106 and acloud manager 138 to manage a cloudcomputing platform provider 110 as described in more detail below. As described herein, theexample system 100 facilitates management of thecloud provider 110 and does not include thecloud provider 110. Alternatively, thesystem 100 could be included in thecloud provider 110. - The cloud
computing platform provider 110 provisions virtual computing resources (e.g., virtual machines, or “VMs,” 114) that may be accessed by users of the cloud computing platform 110 (e.g., users associated with anadministrator 116 and/or a developer 118) and/or other programs, software, device. etc. - An
example application 102 ofFIG. 1 includesmultiple VMs 114. Theexample VMs 114 ofFIG. 1 provide different functions within the application 102 (e.g., services, portions of theapplication 102, etc.). One or more of theVMs 114 of the illustrated example are customized by anadministrator 116 and/or adeveloper 118 of theapplication 102 relative to a stock or out-of-the-box (e.g., commonly available purchased copy) version of the services and/or application components. Additionally, the services executing on theexample VMs 114 may have dependencies on other ones of theVMs 114. - As illustrated in
FIG. 1 , the example cloudcomputing platform provider 110 may providemultiple deployment environments 112, for example, for development, testing, staging, and/or production of applications. Theadministrator 116, thedeveloper 118, other programs, and/or other devices may access services from the cloudcomputing platform provider 110, for example, via REST (Representational State Transfer) APIs (Application Programming Interface) and/or via any other client-server communication protocol. Example implementations of a REST API for cloud computing services include a vCloud Administrator Center™ (vCAC) and/or vRealize Automation™ (vRA) API and a vCloud Director™ API available from VMware, Inc. The example cloudcomputing platform provider 110 provisions virtual computing resources (e.g., the VMs 114) to provide thedeployment environments 112 in which theadministrator 116 and/or thedeveloper 118 can deploy multi-tier application(s). One particular example implementation of a deployment environment that may be used to implement thedeployment environments 112 ofFIG. 1 is vCloud DataCenter cloud computing services available from VMware, Inc. - In some examples disclosed herein, a lighter-weight virtualization is employed by using containers in place of the
VMs 114 in thedevelopment environment 112.Example containers 114 a are software constructs that run on top of a host operating system without the need for a hypervisor or a separate guest operating system. Unlike virtual machines, thecontainers 114 a do not instantiate their own operating systems. Like virtual machines, thecontainers 114 a are logically separate from one another. Numerous containers can run on a single computer, processor system and/or in thesame development environment 112. Also like virtual machines, thecontainers 114 a can execute instances of applications or programs (e.g., anexample application 102 a) separate from application/program instances executed by the other containers in thesame development environment 112. - The
example application director 106 ofFIG. 1 , which may be running in one or more VMs, orchestrates deployment of multi-tier applications onto one of theexample deployment environments 112. As illustrated inFIG. 1 , theexample application director 106 includes atopology generator 120, adeployment plan generator 122, and adeployment director 124. - The
example topology generator 120 generates abasic blueprint 126 that specifies a logical topology of an application to be deployed. The examplebasic blueprint 126 generally captures the structure of an application as a collection of application components executing on virtual computing resources. For example, thebasic blueprint 126 generated by theexample topology generator 120 for an online store application may specify a web application (e.g., in the form of a Java web application archive or “WAR” file including dynamic web pages, static web pages, Java servlets, Java classes, and/or other property, configuration and/or resources files that make up a Java web application) executing on an application server (e.g., Apache Tomcat application server) that uses a database (e.g., MongoDB) as a data store. As used herein, the term “application” generally refers to a logical deployment unit, including one or more application packages and their dependent middleware and/or operating systems. Applications may be distributed across multiple VMs. Thus, in the example described above, the term “application” refers to the entire online store application, including application server and database components, rather than just the web application itself. In some instances, the application may include the underlying hardware and/or virtual computing hardware utilized to implement the components. - The example
basic blueprint 126 ofFIG. 1 may be assembled from items (e.g., templates) from acatalog 130, which is a listing of available virtual computing resources (e.g., VMs, networking, storage, etc.) that may be provisioned from the cloudcomputing platform provider 110 and available application components (e.g., software services, scripts, code components, application-specific packages) that may be installed on the provisioned virtual computing resources. Theexample catalog 130 may be pre-populated and/or customized by an administrator 116 (e.g., IT (Information Technology) or system administrator) that enters in specifications, configurations, properties, and/or other details about items in thecatalog 130. Based on the application, theexample blueprints 126 may define one or more dependencies between application components to indicate an installation order of the application components during deployment. For example, since a load balancer usually cannot be configured until a web application is up and running, thedeveloper 118 may specify a dependency from an Apache service to an application code package. - The example
deployment plan generator 122 of theexample application director 106 ofFIG. 1 generates a deployment plan 128 based on thebasic blueprint 126 that includes deployment settings for the basic blueprint 126 (e.g., virtual computing resources' cluster size, CPU, memory, networks, etc.) and an execution plan of tasks having a specified order in which virtual computing resources are provisioned and application components are installed, configured, and started. The example deployment plan 128 ofFIG. 1 provides an IT administrator with a process-oriented view of thebasic blueprint 126 that indicates discrete actions to be performed to deploy the application. Different deployment plans 128 may be generated from a singlebasic blueprint 126 to test prototypes (e.g., new application versions), to scale up and/or scale down deployments, and/or to deploy the application to different deployment environments 112 (e.g., testing, staging, production). The deployment plan 128 is separated and distributed as local deployment plans having a series of tasks to be executed by theVMs 114 provisioned from thedeployment environment 112. EachVM 114 coordinates execution of each task with a centralized deployment module (e.g., the deployment director 124) to ensure that tasks are executed in an order that complies with dependencies specified in theapplication blueprint 126. - The
example deployment director 124 ofFIG. 1 executes the deployment plan 128 by communicating with the cloudcomputing platform provider 110 via acloud interface 132 to provision and configure theVMs 114 in thedeployment environment 112. Theexample cloud interface 132 ofFIG. 1 provides a communication abstraction layer by which theapplication director 106 may communicate with a heterogeneous mixture ofcloud provider 110 anddeployment environments 112. Thedeployment director 124 provides eachVM 114 with a series of tasks specific to the receiving VM 114 (herein referred to as a “local deployment plan”). Tasks are executed by theVMs 114 to install, configure, and/or start one or more application components. For example, a task may be a script that, when executed by aVM 114, causes theVM 114 to retrieve and install particular software packages from acentral package repository 134. Theexample deployment director 124 coordinates with theVMs 114 to execute the tasks in an order that observes installation dependencies betweenVMs 114 according to the deployment plan 128. After the application has been deployed, theapplication director 106 may be utilized to monitor and/or modify (e.g., scale) the deployment. - The
example cloud manager 138 ofFIG. 1 interacts with the components of the system 100 (e.g., theapplication director 106 and the cloud provider 110) to facilitate the management of the resources of thecloud provider 110. Theexample cloud manager 138 includes ablueprint manager 140 to facilitate the creation and management of multi-machine blueprints and aresource manager 144 to reclaim unused cloud resources. Thecloud manager 138 may additionally include other components for managing a cloud environment. - The
example blueprint manager 140 of the illustrated example manages the creation of multi-machine blueprints that define the attributes of multiple virtual machines as a single group that can be provisioned, deployed, managed, etc. as a single unit. For example, a multi-machine blueprint may include definitions for multiple basic blueprints that make up a service (e.g., an e-commerce provider that includes web servers, application servers, and database servers). A basic blueprint is a definition of policies (e.g., hardware policies, security policies, network policies, etc.) for a single machine (e.g., a single virtual machine such as a web server virtual machine and/or container). Accordingly, theblueprint manager 140 facilitates more efficient management of multiple virtual machines and/or containers than manually managing (e.g., deploying) basic blueprints individually. Example management of multi-machine blueprints is described in further detail in conjunction withFIG. 2 . - The
example blueprint manager 140 ofFIG. 1 additionally annotates basic blueprints and/or multi-machine blueprints to control how workflows associated with the basic blueprints and/or multi-machine blueprints are executed. As used herein, a workflow is a series of actions and decisions to be executed in a virtual computing platform. Theexample system 100 includes first and second distributed execution manager(s) (DEM(s)) 146A and 146B to execute workflows. According to the illustrated example, thefirst DEM 146A includes a first set of characteristics and is physically located at afirst location 148A. Thesecond DEM 146B includes a second set of characteristics and is physically located at asecond location 148B. The location and characteristics of a DEM may make that DEM more suitable for performing certain workflows. For example, a DEM may include hardware particularly suited for performance of certain tasks (e.g., high-end calculations), may be located in a desired area (e.g., for compliance with local laws that require certain operations to be physically performed within a country's boundaries), may specify a location or distance to other DEMS for selecting a nearby DEM (e.g., for reducing data transmission latency), etc. Thus, theexample blueprint manager 140 annotates basic blueprints and/or multi-machine blueprints with capabilities that can be performed by a DEM that is labeled with the same or similar capabilities. - The
resource manager 144 of the illustrated example facilitates recovery of cloud computing resources of thecloud provider 110 that are no longer being activity utilized. Automated reclamation may include identification, verification and/or reclamation of unused, underutilized, etc. resources to improve the efficiency of the running cloud infrastructure. -
FIG. 2 illustrates an example implementation of theblueprint 126 as a multi-machine blueprint generated by theexample blueprint manager 140 ofFIG. 1 . In the illustrated example ofFIG. 2 , three example basic blueprints (aweb server blueprint 202, an application server blueprint 204, and a database (DB) server blueprint 206) have been created (e.g., by the topology generator 120). For example, theweb server blueprint 202, the application server blueprint 204, and thedatabase server blueprint 206 may define the components of an e-commerce online store. - The
example blueprint manager 140 provides a user interface for a user of the blueprint manager 140 (e.g., theadministrator 116, thedeveloper 118, etc.) to specify blueprints (e.g., basic blueprints and/or multi-machine blueprints) to be assigned to an instance of amulti-machine blueprint 208. For example, the user interface may include a list of previously generated basic blueprints (e.g., theweb server blueprint 202, the application server blueprint 204, thedatabase server blueprint 206, etc.) to allow selection of desired blueprints. Theblueprint manager 140 combines the selected blueprints into the definition of themulti-machine blueprint 208 and stores information about the blueprints in a multi-machine blueprint record defining themulti-machine blueprint 208. Theblueprint manager 140 may additionally include a user interface to specify other characteristics corresponding to themulti-machine blueprint 208. For example, a creator of themulti-machine blueprint 208 may specify a minimum number and a maximum number of each blueprint component of themulti-machine blueprint 208 that may be provisioned during provisioning of themulti-machine blueprint 208. - Accordingly, any number of virtual machines (e.g., the virtual machines associated with the blueprints in the multi-machine blueprint 208) and/or containers may be managed collectively. For example, the multiple virtual machines corresponding to the
multi-machine blueprint 208 may be provisioned based on an instruction to provision themulti-machine blueprint 208, may be power cycled by an instruction, may be shut down by an instruction, may be booted by an instruction, etc. As illustrated inFIG. 2 , an instruction to provision themulti-machine blueprint 208 may result in the provisioning of a multi-machine service formed from one ormore VMs 114 that includes virtualized web server(s) 210A, virtualized application server(s) 210B, and virtualized database server(s) 210C. The number of virtual machines and/or containers provisioned for each blueprint may be specified during the provisioning of the multi-machine blueprint 208 (e.g., subject to the limits specified during creation or management of the multi-machine blueprint 208). - The
multi-machine blueprint 208 maintains the reference to thebasic blueprints multi-machine blueprint 208. Accordingly, an administrator maintaining the source blueprints (e.g., an administrator charged with managing the web server blueprint 202) may change or update the source blueprint and the changes may be automatically propagated to the machines provisioned from themulti-machine blueprint 208. For example, if an operating system update is applied to a disk image referenced by the web server blueprint 202 (e.g., a disk image embodying the primary disk of the web server blueprint 202), the updated disk image is utilized when deploying the multi-machine blueprint. Additionally, the blueprints may specify that themachines multi-machine service 210 provisioned from themulti-machine blueprint 208 operate in different environments. For example, some components may be physical machines, some may be on-premise virtual machines, and some may be virtual machines at a cloud service. - Several multi-machine blueprints may be generated to provide one or more varied or customized services. For example, if virtual machines deployed in the various States of the United States require different settings, a multi-machine blueprint could be generated for each state. The multi-machine blueprints could reference the same build profile and/or disk image, but may include different settings specific to each state. For example, the deployment workflow may include an operation to set a locality setting of an operating system to identify a particular state in which a resource is physically located. Thus, a single disk image may be utilized for multiple multi-machine blueprints reducing the amount of storage space for storing disk images compared with storing a disk image for each customized setting.
-
FIG. 3 illustrates an example installation of deployed appliances or virtual appliances (vAs) (e.g.,VMs 114 and/orcontainers 114 a) and associated virtualized servers acting as hosts for deployment of component servers (e.g., Web server, application server, database server, etc.) for a customer. The vAs can be deployed as an automation tool, for example, used to deliver VMs and associated applications for on-premise automation and/or handling of external cloud resources (e.g., Microsoft Azure™, Amazon Web Services™, etc.). - As shown in the example of
FIG. 3 , aninstallation 300 includes a load balancer (LB) 310 to assign tasks and/or manage access among a plurality of vAs 320, 322, 324. Each vA 320-324 is a deployedVM 114 and/orcontainer 114 a. In this example, thevA 320 communicates with a plurality of component orhost servers Web server 210A with Web components,App server 210B with application components,DB server 210C with database components, etc.). As shown in the example ofFIG. 3 ,component servers component server 330 rather than or in addition to directly from thevirtual appliance 320, although thevA 320 can still communicate withsuch servers LB 310 enables the multiple vAs 320-324 and multiple servers 330-336 to appear as one device to a user. Access to functionality can then be distributed among appliances 320-324 by theLB 310 and among servers 330-336 by therespective appliance 320, for example. TheLB 310 can use least response time, round-robin, and/or other method to balance traffic to vAs 320-324 and servers 330-336, for example. - In the
example installation 300, eachvA management endpoint component server management agent respective endpoint 340 to facilitate transfer of data, execution of tasks, etc., for example. - In certain examples,
management endpoints management endpoint management endpoints - In certain examples, the management agents 350-356 synchronize component servers 330-336 with the vA 320-324 and facilitate host access and associated services (e.g., hostd, ntpd, sfcbd, slpd, wsman, vobd, etc.). The management agents 350-356 can communicate with their
respective endpoint 340 to facilitate transfer of data, execution of tasks, etc., for example. The relationship betweenmanagement endpoint management agents multiple component machines - In certain examples, a graphical user interface associated with a front end of the
load balancer 310 guides a customer through one or more questions to determine system requirements for theinstallation 300. Once the customer has completed the questionnaire and provided firewall access to install the agents 350-356, the agents 350-356 communicate with theendpoint 340 without customer involvement. Thus, for example, if a new employee needs a Microsoft Windows® machine, a manager selects an option (e.g., clicks a button, etc.) via the graphical user interface to install aVM 114 and/orcontainer 114 a that is managed through theinstallation 300. To the user, he or she is working on a single machine, but behind the scenes, the virtual appliance (vA) 320 is accessing different servers 330-336 depending upon what functionality is to be executed. - In certain examples agents 350-356 are deployed in a same data center as the
endpoint 340 to which the agents 350-356 are associated. The deployment can include a plurality of agent servers 330-336 distributed worldwide, and the deployment can be scalable to accommodate additional server(s) with agent(s) to increase throughput and concurrency, for example. -
FIG. 4 illustrates an example implementation of thevA 320. In the example ofFIG. 4 , thevA 320 includes aservice provisioner 410, anorchestrator 420, anevent broker 430, anauthentication provider 440, aninternal reverse proxy 450, and adatabase 460. Thecomponents vA 320 may be implemented by one or more of theVMs 114. Theexample service provisioner 410 provides services to provision interfaces (e.g., Web interface, application interface, etc.) for thevA 320. The example orchestrator (e.g., vCO) 420 is an embedded or internal orchestrator that can leverage a provisioning manager, such as theapplication director 106 and/orcloud manager 138, to provision VM services but is embedded in thevA 320. For example, thevCO 420 can be used to invoke a blueprint to provision a manager for services. - Example services can include catalog services, identity services, component registry services, event broker services, IaaS, XaaS, etc. Catalog services provide a user interface via which a user can request provisioning of different preset environments (e.g., a VM including an operating system and software and some customization, etc.), for example. Identity services facilitate authentication and authorization of users and assigned roles, for example. The component registry maintains information corresponding to installed and deployed services (e.g., uniform resource locators for services installed in a VM/vA, etc.), for example. The event broker provides a messaging broker for event-based communication, for example. The IaaS provisions one or more VMs and/or containers for a customer via the
vA 320. The XaaS can extend the provisioning to also request, approve, provision, operate, and decommission any type of catalog items (e.g., storage, applications, accounts, and anything else that the catalog provides as a service). - The
example event broker 430 provides a mechanism to handle tasks which are transferred between services with theorchestrator 420. The example authentication provider 440 (e.g., VMware Horizon™ services, etc.) authenticates access to services and data, for example. - The components of the
vA 320 access each other through REST API calls behind the internal reverse proxy 450 (e.g., a high availability (HA) proxy HAProxy) which provides a high availability load balancer and proxy for Transmission Control Protocol (TCP)- and Hypertext Transfer Protocol (HTTP)-based application requests. In this example, theproxy 450 forwards communication traffic from within thevA 320 and/or between vAs 320, 322, 324 ofFIG. 3 to the appropriate component(s) of thevA 320. In certain examples, services access the local host/proxy 450 on a particular port, and the call is masked by theproxy 450 and forwarded to the particular component of thevA 320. Since the call is masked by theproxy 450, components can be adjusted within thevA 320 without impacting outside users. - Example Infrastructure Installation
- In certain examples, a cloud computing (e.g., vCAC™, vRA™, etc.) deployment includes one or more vAs 320-324 and one or more component servers 330-336 (e.g., Microsoft Windows™ machines, etc.) on which are installed components (e.g., software such as Web services, application services, database services, etc.) that form the IaaS portion of the product. In a distributed and/or high availability deployment, a plurality of component servers 330-336 form the installed product, and having to install the IaaS components manually on all of the component servers 330-336 is a time-consuming process, involving, among other things, multiple context switches and many opportunities for user misconfiguration of the deployed system. For example, manual installation involves installing components on an appliance, downloading an installer, and then visit each server to install the components manually using the installer. However, if a component is deployed out of order, the installation may not function. Additionally, data entry is required for each manual installation, and mis-typing of the manual data entry can invalidate the entire installation. Further, such a mistake may not be realized until the erroneous installation is deployed, resulting in lost time, money, errors, and inoperable systems. Simplification and automation of this process reduces the time needed and errors involved in setting up a new instance of the cloud computing system.
- In certain examples, rather than requiring customers to manually install an IaaS component on each server 330-336, installation can be executed on each node from a centralized location via the management agent 350-356 installed on each component server 330-336. The agent 350-356 is installed and registered with the
vA 320. After registration, communication with thevA 320 is authenticated using a client certificate and signed token(s). The vA's 320 root credentials are not persisted on the server 330-336. Each instance of the management agent 350-356 has a node identifier (ID), which uniquely identifies the agent 330-336 in a cluster of machines 330-336 forming the cloud deployment. After registration, the agent 330-336 starts polling thevA 320 in a configurable time interval to obtain commands to be executed. The commands are executed by the corresponding server 330-336, and a result is reported back to thevA 320 by the agent 350-356 and can be used for further processing, for example. - In certain examples, installation of a hybrid system including a plurality of appliances 320-324 and component servers 330-336 having a plurality of roles can be orchestrated via the management agents 350-356. Using the management agents 350-356 in communication with the management endpoints 340-344 at their respective vAs 320-324, the
example installation 300 can be coordinated without manual user action throughout phases of the installation. -
FIG. 5 illustrates a block diagram of an example implementation of the management agent 350 (and/or 352, 354, 356). As shown in the example ofFIG. 5 , themanagement agent 350 includes acommunication interface 510 through which theagent 350 can communicate with the endpoint 340 (and/or 342, 344) of the vA 320 (and/or 322, 324). Thecommunication interface 510 is a hardware and/or software interface allowing theagent 350 to exchange data, commands, etc., with theendpoint 340 and/or other communication node, for example. - The
example agent 350 also includes anagent control processor 520. Theagent control processor 520 executes instructions to control theagent 350 for command and/or other application execution, communication, storage, etc. The instructions can be transmitted to theagent control processor 520 via thecommunication interface 510 and/or via adata storage 530, for example. - The
example data storage 530 includes a configuration file 540 and amachine identifier 550. The example configuration file 540 can include information such as credentials to authenticate and/or validate theagent 350 to thevA 320, etc. Credentials can include a certificate (e.g., with a public key and private key for authentication, etc.), a unique identifier, etc. The exampleagent control processor 520 can process instructions, generate communications, etc. Theexample data storage 530 can also include instructions (e.g., computer program code, etc.) to be executed by theagent control processor 520. -
FIG. 6 illustrates an example data flow diagram showing an exchange ofinformation 600 between thevA component server 330 to installsystem 300 components including the vA320, 322 andcomponent server 330 includingmanagement agent 350. At 602, a trigger 601 (e.g., a user via a vA management webpage, an automated script, etc.) initiates deployment of thevA vA 320, for example. - At 606, a management agent 350 (and/or 352-356) is installed on each component server 330 (and/or 332-336). At 608, each server 330 (and/or 332-336) acknowledges the installation of the management agent 350 (and/or 352-356). The acknowledgement can include an authentication of
management agent 350 credentials by thevA 320, for example. Authentication and/or authorization can include an exchange and verification of a certificate, identifier, etc., associated with themanagement agent 350 and/or its associatedserver 330 by thevA 320, for example. - For example, a cloud-based installation may include one or more vAs 320-324 and one or more servers 330-336 (e.g., “Windows™ machines”, etc.) on which a plurality of components (e.g., five, six, seven, ten, etc.) are installed (e.g., applications, database, management, etc.) to form an IaaS in a distributed, high availability environment. The management agents 350-356 communicate with the management endpoint(s) 340-344 to receive commands, execute commands, install software, upgrade an installation at the server 330-336, etc.
- Each management agent 350-356 has a node identifier (ID) that uniquely identifies the agent 350-356 in a cluster of machines 330-336 forming the
system 300. When installing the agent 350-356, an address and root credentials of theprimary vA 320 are entered so that the agent 350-356 can register itself in thevA 320. After the registration, communication with thevA 320 is authenticated using a client certificate and signed token(s). In certain examples, since the certificate is used for communication between the agent 350-356 and theendpoint 340, the root credentials of thevA 320 are not persisted on the machines 330-336. - At 610, other appliances (e.g.,
other vAs 324, etc.) are identified to thevA 320. For example, theprimary vA 320 is informed of other appliances (e.g.,vA server 330 is assigned a role as a database server. Alternatively, theserver 330 is assigned a role as a Web server. Theserver 330 may be assigned a role as an application server, for example. Theserver 330 may be assigned a role as a Windows™ server, for example. - Each role is associated with one or more rules that guide and/or establish criteria for the associated role. Each rule can be associated with one or more prerequisites for a server 330-336 to execute the rule and perform the role. In a high availability (HA) environment, rules may specify that there are at least two servers 330-336 for each role to provide redundancy and increased availability if one
server 330 of a given role is busy or otherwise unavailable, for example. - At 614, the
vA 320 evaluates or checks applicable rule prerequisite(s) for a given role to ensure the associatedserver 330 can perform the role. ThevA 320 attempts to fix or otherwise remedy any error(s) identified in its evaluation of theserver 330. For example, prerequisites can include a determination of whether theload balancer 310, vA 320-324, and/or component server 330-336, etc., is/are reachable, and an associated fix can include an address look-up and resolution. Another prerequisite can include registration of the server 330-336 and/or other IaaS node with the vA 320-324, and an associated fix can include registration of the server 330-336 and/or other IaaS node with the vA 320-324, such as described above, for example. Another fix of a server 330-336 error can include downloading and installing a service pack on the server 330-336, for example. Another prerequisite can include database (e.g., object-relational database such as Postgres, etc.) access, and an associated fix can include access and log in to the database, for example. - At 616, the
server 330 reports back to thevA 320 with an indication of what cannot be automatically fixed. For example, theserver 330 generates a report of what prerequisite errors were fixed and what prerequisite errors were identified and not fixed, and sends 616 the report to thevA 320. - At 618, errors that were identified but not fixed in the prerequisite checks are handled based on data entry at the
vA 320. For example, via thetrigger 601, the user can provide data entry and/or one or more automated scripts can be executed to provide answer(s), setting(s), address(es), password(s), credential(s), etc., to satisfy missing prerequisite information for theserver 330 at thevA 320. In certain examples, thedata entry 618 can be guided by a wizard for component address(es), username(s), identity of service(s) to be executed, identification of port(s) (e.g., default or custom), database credentials/address, etc. In certain examples, a user provides data entry which is recorded and converted into a script for subsequent automated execution, for example. - At 620, once data entry is complete, the
vA 320 asks to configure the load balancer (LB) 310 with an external device/system 603. Thus, theLB 310 can be configured with information regarding connected components 320-324, 330-336, etc., and their capabilities and ability to distribute workload for a plurality of tasks. Using data and the management agent(s) 350-356, rules and/or other prerequisites are sent to server(s) 330-336 along with other instructions to configure a particular customer environment. - At 622, the
vA 320 validates the server(s) 330-336, and, at 624, thevA 320 validates itself. For example, at 622, thevA 320 sends one or more commands to theserver 330 based on the role of the server 330 (e.g., install web service (validate=true), install manager service, etc.) to validate the installation and/or other configuration of theserver 330. Theserver 330 returns an indication of whether or not the validation is okay (e.g., has been completed successfully, etc.). At 626, if the validation failed, repeateddata entry 618 can be used to adjust network settings and/or other change in data to then repeat validation at 622, for example. At 628, if validation is successful, then the installation, configuration, and validation process is complete. - At 630, a snapshot of the installation is obtained. The snapshot forms a recovery point. The snapshot stores the configuration for later use, restoration, error recovery, etc. At 632, the installation is executed. First, at 632, the installation is executed at the vA 320 (e.g., the primary appliance). Then, at 634, the installation occurs at the server 330 (e.g., via the management agent 350). At 636, the installation occurs at the vA 322 (e.g., a secondary appliance). Thus, a sequence of commands is built and targeted for an environment including server(s) 330. The commands are triggered for orchestration of services via the management agent(s) 350. The central or
primary vA 320 does not have access to individual nodes but instead can be accessed by the management agent(s) 350-356 of the respective server(s) 330-336, which acts to execute installation instructions from thevA 320. ThevA 320 awaits acknowledgement from the agent(s) 350-356. -
FIG. 7 illustrates an example schedule ofcomponents 700 for installation in an example system. Theexample vA 320 includes information for installation 632-638 of multiple components after prerequisite checks, validation, and snapshot generation 602-630, as described above with respect toFIG. 6 . As illustrated in the example ofFIG. 7 ,vA 320, referred to in the example as the primary or main appliance vA1, interacts with a plurality of servers 330-336, referred to in the example as server1, server2, server3, server4, and server5. ThevA 320 also interacts with a plurality of secondary appliances 322-324, referred to in the example as vA2 and vA3. - As illustrated in the example of
FIG. 7 , thevA1 320 has a queue of commands 702 (e.g., install commands, joinder commands, etc.) established via itsmanagement endpoint 340. The commands 702 (e.g., install vA1, install identity vA1, install web component on server1 (e.g., primary Windows™ server), install manager service component on server4, install DEM on server3, etc., then join appliance to cluster vA2, join appliance to cluster vA3, etc., then install web on server2, install manager service on server3, etc., install management agent on server5, etc.) are executed by the vA1, alone or in conjunction with one or more servers 330-336 and/or external device(s), for example. - Some or all of the
commands 702 havecertain dependency 704 on other command(s) 702. That is, onecommand 702 may depend on nothing, but anothercommand 702 may depend on aprior command 702. Thus, while some commands can be executed in parallel and/or otherwise asynchronously, other commands must be executed serially and/or otherwise synchronously when theirdependency 704 has been satisfied, for example. - The
vA1 320 builds and then deploys the installation including copies of involved components. Each component server 330-336 polls thevA1 320 for work. Theendpoint 340 of thevA 320 examines thequeue 702 to determine a next task. If the next task has nodependency 704, theendpoint 350 transmits the task to the server 330-336 via its agent 350-356 for execution. If the task in thequeue 702 has adependency 704, then theendpoint 340 determines whether the dependent task has been completed, started, or not yet started, for example. If the dependent task has not yet started, for example, then theendpoint 340 will not pass the associated task from thequeue 702 to the server 330-336. - Each task in the
queue 702 is also associated with and/or otherwise intended for aserver 706. For example, server1 queries the vA1 to determine whether thequeue 702 includes any task for server1. Theendpoint 340 of the appliance vA1 determines that yes,task 3 in thequeue 702 is for server1 (e.g., installation of web server on server1, etc.). However,task 3 depends on the execution oftask 2 by the vA1, andtask 2 depends on the execution oftask 1 by the vA1. The endpoint vA1 does not releasetask 3 to the agent of server1 until the vA1 has executedtasks Task 4 however, has nodependency 704, so the endpoint can releasetask 4 to the agent of server4 when the agent polls for its task. Additionally, in the example ofFIG. 7 , once task execution reachestask 7, all remaining tasks depend ontask 5, which has then been completed. Thus, tasks 7-12 can be dispensed and/or otherwise released to be executed in parallel by their respective component. - In certain examples, the
vA 320 publishes thecommands 702 to available servers 330-336 (e.g., via communication between theendpoint 340 and agents 350-356) and is aware of thesequence 702 anddependency 704 of the commands. The servers 330-336, however, accept and execute the commands without knowledge of thesequence 702 anddependency 704. For example, when a server 330-336 is instructed to install a component, the server's agent 350-356 facilitates a download of an installer from thevA 320, which the server 330-336 extracts and executes to install the component. Thus, the vA320 is aware of the servers 330-336 available to execute tasks but does not instruct the servers 330-36 to execute tasks. Rather, the servers 330-336 poll the queue ofinstructions 702, and thevA 320 checks whetherdependencies 704 have been met to release the task to be performed by the server 330-336 and/or other component. - While example implementations of the example
cloud computing system 100 andvirtual machine installation 300 are illustrated inFIGS. 1-7 , one or more of the elements, processes and/or devices illustrated inFIGS. 1-7 may be combined, divided, re-arranged, omitted, eliminated and/or implemented in any other way. Further, theexample application director 106,example cloud provider 110,example cloud manager 138, example distributedexecution managers example multi-machine service 210,example load balancer 310, example virtual appliances 320-324, example component servers 330-336, example management endpoints 340-344, example management agents 350-356, and/or, more generally, theexample systems 100 and/or 300 ofFIGS. 1-7 can be implemented by hardware, software, firmware and/or any combination of hardware, software and/or firmware. Thus, for example, any of theexample application director 106,example cloud provider 110,example cloud manager 138, example distributedexecution managers example multi-machine service 210,example load balancer 310, example virtual appliances 320-324, example component servers 330-336, example management endpoints 340-344, example management agents 350-356, and/or, more generally, theexample systems 100 and/or 300 ofFIGS. 1-7 can be implemented by one or more analog or digital circuit(s), logic circuits, programmable processor(s), application specific integrated circuit(s) (ASIC(s)), programmable logic device(s) (PLD(s)) and/or field programmable logic device(s) (FPLD(s)). When reading any of the apparatus or system claims of this patent to cover a purely software and/or firmware implementation, at least one of theexample application director 106,example cloud provider 110,example cloud manager 138, example distributedexecution managers example multi-machine service 210,example load balancer 310, example virtual appliances 320-324, example component servers 330-336, example management endpoints 340-344, example management agents 350-356, and/or, more generally, theexample systems 100 and/or 300 ofFIGS. 1-7 is/are hereby expressly defined to include a tangible computer readable storage device or storage disk such as a memory, a digital versatile disk (DVD), a compact disk (CD), a Blu-ray disk, etc. storing the software and/or firmware. Further still, theexample application director 106,example cloud provider 110,example cloud manager 138, example distributedexecution managers example multi-machine service 210,example load balancer 310, example virtual appliances 320-324, example component servers 330-336, example management endpoints 340-344, example management agents 350-356, and/or, more generally, theexample systems 100 and/or 300 ofFIGS. 1-7 may include one or more elements, processes and/or devices in addition to, or instead of, those illustrated inFIGS. 1-7 , and/or may include more than one of any or all of the illustrated elements, processes and devices. - Flowcharts representative of example machine readable instructions that may be executed to deploy and manage the
example application director 106,example cloud provider 110,example cloud manager 138, example distributedexecution managers example multi-machine service 210,example load balancer 310, example virtual appliances 320-324, example component servers 330-336, example management endpoints 340-344, example management agents 350-356, and/or, more generally, theexample systems 100 and/or 300 ofFIGS. 1-7 are shown inFIGS. 8-9 . In these examples, the machine readable instructions implement programs for execution by a processor such as theprocessor 1012 shown in theexample processor platform 1000 discussed below in connection withFIG. 10 . The programs may be embodied in software stored on a tangible computer readable storage medium such as a CD-ROM, a floppy disk, a hard drive, a digital versatile disk (DVD), a Blu-ray disk, or a memory associated with theprocessor 1012, but the entire program and/or parts thereof could alternatively be executed by a device other than theprocessor 1012 and/or embodied in firmware or dedicated hardware. Further, although the example programs are described with reference to the flowcharts illustrated inFIGS. 8-9 , many other methods of deploying, managing, and updating workload domains in accordance with the teachings of this disclosure may alternatively be used. For example, the order of execution of the blocks may be changed, and/or some of the blocks described may be changed, eliminated, or combined. - As mentioned above, the example processes of
FIGS. 8-9 may be implemented using coded instructions (e.g., computer and/or machine readable instructions) stored on a tangible computer readable storage medium such as a hard disk drive, a flash memory, a read-only memory (ROM), a compact disk (CD), a digital versatile disk (DVD), a cache, a random-access memory (RAM) and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information). As used herein, the term tangible computer readable storage medium is expressly defined to include any type of computer readable storage device and/or storage disk and to exclude propagating signals and to exclude transmission media. As used herein, “tangible computer readable storage medium” and “tangible machine readable storage medium” are used interchangeably. In some examples, the example processes ofFIGS. 8-9 may be implemented using coded instructions (e.g., computer and/or machine readable instructions) stored on a non-transitory computer and/or machine readable medium such as a hard disk drive, a flash memory, a read-only memory, a compact disk, a digital versatile disk, a cache, a random-access memory and/or any other storage device or storage disk in which information is stored for any duration (e.g., for extended time periods, permanently, for brief instances, for temporarily buffering, and/or for caching of the information). As used herein, the term non-transitory computer readable medium is expressly defined to include any type of computer readable storage device and/or storage disk and to exclude propagating signals and to exclude transmission media. As used herein, when the phrase “at least” is used as the transition term in a preamble of a claim, it is open-ended in the same manner as the term “comprising” is open ended. Comprising and all other variants of “comprise” are expressly defined to be open-ended terms. Including and all other variants of “include” are also defined to be open-ended terms. In contrast, the term consisting and/or other forms of consist are defined to be close-ended terms. -
FIG. 8 depicts a flowchart representative of computer readable instructions that may be executed to implement theexample infrastructure installation 300. Anexample program 800 is illustrated inFIG. 8 . Initially, atblock 802, the first, primary, ormain vA 320 is deployed (e.g., triggered by a user, an automated script, an event, etc.). The installation wizard is also triggered to be run by thevA 320, for example. - At
block 804, components in thesystem 300 are identified. For example, a cloud-based installation may include one or more vAs 320-324 and one or more servers 330-336 (e.g., “Windows™ machines”, etc.) on which a plurality of components (e.g., five, six, seven, ten, etc.) are installed (e.g., applications, database, management, etc.) to form an IaaS in a distributed, high availability environment. The management agents 350-356 communicate with the management endpoint(s) 340-344 to receive commands, execute commands, install software, upgrade an installation at the server 330-336, etc. Thesystem 300 can also include one or more devices such as aload balancer 310, etc. - At
block 806, the management agent 350-356 is installed on each component server 330-336. Each server 330-336 acknowledges the installation of the management agent 350-356. The acknowledgement can include an authentication ofmanagement agent 350 credentials by thevA 320, for example. Authentication and/or authorization can include an exchange and verification of a certificate, identifier, etc., associated with themanagement agent 350 and/or its associatedserver 330 by thevA 320, for example. - In certain examples, each management agent 350-356 has a node identifier (ID) that uniquely identifies the agent 350-356 in a cluster of machines 330-336 forming the
system 300. When installing the agent 350-356, an address and root credentials of theprimary vA 320 are entered so that the agent 350-356 can register itself in thevA 320. After the registration, communication with thevA 320 is authenticated using a self-signed certificate. In certain examples, since the self-signed certificate is used for communication between the agent 350-356 and theendpoint 340, the root credentials of thevA 320 are not persisted on the machines 330-336. - At
block 808, a role is provided for each server 330 (and/or 332-336). For example, theserver 330 is assigned a role as a database server. Alternatively, theserver 330 is assigned a role as a Web server. Theserver 330 may be assigned a role as an application server, for example. Theserver 330 may be assigned a role as a Windows™ server, for example. - Each role is associated with one or more rules that guide and/or establish criteria for the associated role. Each rule can be associated with one or more prerequisites for a server 330-336 to execute the rule and perform the role. In a high availability (HA) environment, rules may specify that there are at least two servers 330-336 for each role to provide redundancy and increased availability if one
server 330 of a given role is busy or otherwise unavailable, for example. - At
block 810, thevA 320 evaluates or checks applicable prerequisite(s) for a given role to ensure the associatedserver 330 can perform the role. For example, prerequisites can include a) a determination of whether theload balancer 310, vA 320-324, and/or component server 330-336, etc., is/are reachable; b) registration of the server 330-336 and/or other IaaS node with the vA 320-324; c) presence of a minimum software and/or firmware version; d) database (e.g., object-relational database such as Postgres, etc.) access; etc. - At
block 812, an error is identified in the prerequisite evaluation. For example, a prerequisite may not be satisfied and may need to be addressed before installation can continue. If no error is identified, then control advances to block 820 to evaluate whether prerequisites remain to be evaluated. If an error is identified, then, atblock 814, the error is evaluated to determine whether or not the error is automatically fixable by thevA 320 and/or other connected component. If the error is fixable, then, atblock 816, thevA 320 automatically fixes the error. If the error is not automatically fixable, then, atblock 818, the error is logged (e.g., included in a report, etc.). - At
block 820, prerequisite(s) are evaluated for the role to determine whether or not additional prerequisite(s) remain to be evaluated. If prerequisite(s) remain to be evaluated, the control reverts to block 810 to evaluate the prerequisite(s) for the role(s). - If all prerequisite(s) have been evaluated, then, at
block 822, the error log is reviewed to identify errors logged during prerequisite analysis. If no error has been logged, then, atblock 828, a recovery point (e.g., a snapshot) is created for thesystem 300 configuration. However, if one or more errors are identified, then, atblock 824, error resolution is attempted. For example, data entry can be provided via a user and/or one or more automated scripts to provide answer(s), setting(s), address(es), password(s), credential(s), etc., to satisfy missing prerequisite information for theserver 330 at thevA 320. In certain examples, thedata entry 618 can be guided by a wizard for component address(es), username(s), identity of service(s) to be executed, identification of port(s) (e.g., default or custom), database credentials/address, etc. In certain examples, a user provides data entry which is recorded and converted into a script for subsequent automated execution, for example. - At
block 826, thevA 320 validates the server(s) 330-336 and itself. For example, thevA 320 sends one or more commands to theserver 330 based on the role of the server 330 (e.g., install web service (validate=true), install manager service, etc.) to validate the installation and/or other configuration of theserver 330. Theserver 330 returns an indication of whether or not the validation is okay (e.g., has been completed successfully, etc.). Atblock 826, if the validation failed, repeated data entry atblock 824 can be used to adjust network settings and/or other change in data to then repeat validation atblock 826, for example. If validation is successful, then the installation, configuration, and validation process is complete after the process has been conducted for each logged error (block 822). - After all logged errors have been evaluated, at
block 828, recovery point is created. The recovery point can be formed from a snapshot taken of the installation, for example. The snapshot stores the configuration for later use, restoration, error recovery, etc. Atblock 830, the installation is executed. For example, the installation is executed at the vA 320 (e.g., the primary appliance). Then, the installation occurs at the server 330 (e.g., via the management agent 350). The installation also occurs at the vA 322 (e.g., a secondary appliance). Thus, a sequence of commands is built and targeted for particular server(s) 330. The commands are triggered for orchestration of services via the management agent(s) 350. The central orprimary vA 320 does not have access to individual nodes but instead accesses the management agent(s) 350-356 of the respective server(s) 330-336, which acts to execute installation instructions from thevA 320. ThevA 320 awaits acknowledgement from the agent(s) 350-356. -
FIG. 9 illustrates an example implementation of executing the installation atblock 830 of the example flow diagram ofFIG. 8 . Atblock 902, the server 330-336 polls thevA 320 for a task to be executed (e.g., via communication between the management agent 350-356 and themanagement endpoint 340, etc.). Atblock 904, thevA 320 reviews its queue oftasks 702 to be executed (e.g., install commands, joinder commands, etc.) established via themanagement endpoint 340 to identify atask 702 for the requesting server 330-336. - At
block 906, an identified task is evaluated to identify a dependency. That is, some or all of thetasks 702 have adependency 704 on other task(s) 702. That is, onetask 702 may depend on nothing, but anothertask 702 may depend on aprior task 702. Thus, while sometasks 702 can be executed in parallel and/or otherwise asynchronously,other tasks 702 are to be executed serially and/or otherwise synchronously when theirdependency 704 has been satisfied, for example. - At
block 908, presence or absence of adependency 704 for thetask 702 is examined. If nodependency 704 exists, then, atblock 910, thetask 702 is deployed by thevA 320 for execution by the server 330-336. - At
block 912, if a dependency exists, then thedependency 704 is evaluated to determine whether or not the dependency has been met. If thedependency 704 has been satisfied, then, atblock 910, thetask 702 is deployed by thevA 320 for execution by the server 330-336. However, if thedependency 704 has not been met, then, atblock 914, thevA 320 does not release thetask 702 to the server 330-336 until thedependency 704 has been satisfied. - For example, the server 330-336 queries the
vA 320 to determine whether thequeue 702 includes any task for the server 330-336. Theendpoint 340 of thevA 320 determines that a task in thequeue 702 is for the server 330-336. However, thetask 702 depends on the execution of another task. Theendpoint 340 of thevA 320 does not release thetask 702 to the agent 350-356 of the server 330-36 until thevA 320 has executed the dependent task. If no dependencies are identified or the dependent task has already executed, however, theendpoint 340 can release the task to the agent 350-356 of the server 330-336 when the agent 350-356 polls theendpoint 340 for itstask 702. -
FIG. 10 is a block diagram of anexample processor platform 1000 capable of executing the instructions ofFIGS. 8-9 to implement the example systems, operation, and management ofFIGS. 1-7 . Theprocessor platform 1000 of the illustrated example includes aprocessor 1012. Theprocessor 1012 of the illustrated example is hardware. For example, theprocessor 1012 can be implemented by one or more integrated circuits, logic circuits, microprocessors or controllers from any desired family or manufacturer. - The
processor 1012 of the illustrated example includes a local memory 1013 (e.g., a cache), and executes instructions to implement theexample systems processor 1012 of the illustrated example is in communication with a main memory including avolatile memory 1014 and anon-volatile memory 1016 via abus 1018. Thevolatile memory 1014 may be implemented by Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM) and/or any other type of random access memory device. Thenon-volatile memory 1016 may be implemented by flash memory and/or any other desired type of memory device. Access to themain memory - The
processor platform 1000 of the illustrated example also includes aninterface circuit 1020. Theinterface circuit 1020 may be implemented by any type of interface standard, such as an Ethernet interface, a universal serial bus (USB), and/or a PCI express interface. - In the illustrated example, one or
more input devices 1022 are connected to theinterface circuit 1020. The input device(s) 1022 permit(s) a user to enter data and commands into theprocessor 1012. The input device(s) can be implemented by, for example, an audio sensor, a microphone, a keyboard, a button, a mouse, a touchscreen, a track-pad, a trackball, isopoint and/or a voice recognition system. - One or
more output devices 1024 are also connected to theinterface circuit 1020 of the illustrated example. Theoutput devices 1024 can be implemented, for example, by display devices (e.g., a light emitting diode (LED), an organic light emitting diode (OLED), a liquid crystal display, a cathode ray tube display (CRT), a touchscreen, a tactile output device, a printer and/or speakers). Theinterface circuit 1020 of the illustrated example, thus, typically includes a graphics driver card, a graphics driver chip or a graphics driver processor. - The
interface circuit 1020 of the illustrated example also includes a communication device such as a transmitter, a receiver, a transceiver, a modem and/or network interface card to facilitate exchange of data with external machines (e.g., computing devices of any kind) via a network 1026 (e.g., an Ethernet connection, a digital subscriber line (DSL), a telephone line, coaxial cable, a cellular telephone system, etc.). - The
processor platform 1000 of the illustrated example also includes one or moremass storage devices 1028 for storing software and/or data. Examples of suchmass storage devices 1028 include flash devices, floppy disk drives, hard drive disks, optical compact disk (CD) drives, optical Blu-ray disk drives, RAID systems, and optical digital versatile disk (DVD) drives. -
Coded instructions 1032 representative of the example machine readable instructions ofFIGS. 8-9 may be stored in themass storage device 1028, in thevolatile memory 1014, in thenon-volatile memory 1016, and/or on a removable tangible computer readable storage medium such as a CD or DVD. - In certain examples, the
processor 1012 can be used to implement the virtual appliance 320 (and vAs 322-324) and the component server 330 (and servers 332-336) and their components including theservice provisioner 410,orchestrator 420,event broker 430,authentication provider 440,proxy 450,management endpoint 340,management agent 350,communication interface 510,agent control processor 520,data storage 530, etc. - From the foregoing, it will be appreciated that the above disclosed methods, apparatus and articles of manufacture facilitate installation of a virtual appliance and associated component servers as an IaaS in a distributed environment such as a cloud computing environment and management of agents in the distributed environment. Examples disclosed herein facilitate self-evaluation and installation of servers and agents without further user intervention or cloud oversight.
- As described above, rather than requiring customers to manually install the IaaS on each component server, the installation of each node can be facilitated from a centralized location via a management endpoint and management agents running on each component server. The management agents are registered with the virtual appliance, and further communication between the agent and the appliance is authenticated using a self-signed certificate. The appliance's root credentials are not persisted on the individual component servers. Each instance of the management agent has a node identifier, which uniquely identifies the node in the cluster of machines forming the infrastructure deployment. After registration, the management agent starts polling the virtual appliance in configurable time for commands to be executed. The commands are executed by the server(s), and the result(s) is/are reported back to the virtual appliance and can be used for further processing.
- Certain examples provide an apparatus. The example apparatus includes a first virtual appliance including a management endpoint, the first virtual appliance to organize tasks to be executed to install a computing infrastructure. The example apparatus includes a first component server to execute tasks, the component server including a management agent to communicate with the management endpoint to receive a task to be executed to install the computing infrastructure, the first virtual appliance to associate a role with the first component server and to determine whether the first component server satisfies a prerequisite associated with the role, the first virtual appliance to facilitate addressing an error when the first component server is determined not to satisfy the prerequisite.
- In certain examples, the first virtual appliance of the example apparatus is to facilitate addressing the error by at least one of a) automatically fixing the error such that the first component server satisfies the prerequisite or b) facilitating data entry to resolve the error.
- In certain examples, the first virtual appliance of the example apparatus is to validate the first component server to confirm resolution of the error.
- In certain examples, the first virtual appliance of the example apparatus is to capture a snapshot as a recovery point for installation of the computing infrastructure.
- In certain examples, the first virtual appliance of the example apparatus is to facilitate installation of a service on the first component server by deploying the task to the first component server.
- In certain examples, the management agent of the first component server of the example apparatus is to poll the management endpoint of the first virtual appliance to request the task.
- In certain examples, the first virtual appliance of the example apparatus is to evaluate a dependency associated with the task and evaluate whether an identified dependency is satisfied or unsatisfied, and wherein the first virtual appliance is to deploy the task to the first component server when no unsatisfied dependency is found and is to hold the task when an unsatisfied dependency is found.
- Certain examples provide a method. The example method includes deploying, by executing an instruction with a processor, a first virtual appliance, the first virtual appliance including a management endpoint, the first virtual appliance to organize tasks to be executed to install a computing infrastructure. The example method includes installing, by executing an instruction with the processor, a first component server to execute tasks, the component server including a management agent to communicate with the management endpoint. The example method includes associating, by executing an instruction with the processor, the first virtual appliance with a role. The example method includes determining, by executing an instruction with the processor, whether the first component server satisfies a prerequisite associated with the role. The example method includes facilitating, by executing an instruction with the processor, addressing of an error by the first virtual appliance when the first component server is determined not to satisfy the prerequisite.
- In certain examples, facilitating addressing of the error includes facilitating addressing of the error by at least one of a) automatically fixing the error such that the first component server satisfies the prerequisite or b) facilitating data entry to resolve the error.
- In certain examples, the method further includes validating, by executing an instruction with the processor, the first component server to confirm resolution of the error.
- In certain examples, the method further includes capturing, by executing an instruction with the processor, a snapshot as a recovery point for installation of the computing infrastructure.
- In certain examples, the method further includes facilitating, by executing an instruction with the processor, installation of a service on the first component server by deploying the task to the first component server.
- In certain examples, the method further includes polling, by executing an instruction with the processor, of the management endpoint of the first virtual appliance by the management agent of the first component server to request the task.
- In certain examples, the method further includes evaluating, by executing an instruction with the processor, a dependency associated with the task; evaluating, by executing an instruction with the processor, whether an identified dependency is satisfied or unsatisfied; deploying, by executing an instruction with the processor, the task to the first component server when no unsatisfied dependency is found; and holding, by executing an instruction with the processor, the task when an unsatisfied dependency is found.
- Certain examples provide a computer readable storage medium including instructions that, when executed, cause a machine to at least: deploy a first virtual appliance, the first virtual appliance including a management endpoint, the first virtual appliance to organize tasks to be executed to install a computing infrastructure; install a first component server to execute tasks, the component server including a management agent to communicate with the management endpoint; associate the first virtual appliance with a role; determine whether the first component server satisfies a prerequisite associated with the role; and facilitate addressing of an error by the first virtual appliance when the first component server is determined not to satisfy the prerequisite.
- In certain examples, the instructions, when executed, cause the machine to facilitate addressing of the error including facilitating addressing of the error by at least one of a) automatically fixing the error such that the first component server satisfies the prerequisite or b) facilitating data entry to resolve the error.
- In certain examples, the instructions, when executed, further cause the machine to validate the first component server to confirm resolution of the error.
- In certain examples, the instructions, when executed, further cause the machine to facilitate installation of a service on the first component server by deploying the task to the first component server.
- In certain examples, the instructions, when executed, further cause the machine to facilitate polling of the management endpoint of the first virtual appliance by the management agent of the first component server to request the task.
- In certain examples, the instructions, when executed, further cause the machine to: evaluate a dependency associated with the task; evaluate whether an identified dependency is satisfied or unsatisfied; deploy the task to the first component server when no unsatisfied dependency is found; and hold the task when an unsatisfied dependency is found.
- Although certain example methods, apparatus and articles of manufacture have been disclosed herein, the scope of coverage of this patent is not limited thereto. On the contrary, this patent covers all methods, apparatus and articles of manufacture fairly falling within the scope of the claims of this patent.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/514,897 US20220050711A1 (en) | 2016-12-06 | 2021-10-29 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/370,477 US11182203B2 (en) | 2016-12-06 | 2016-12-06 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
US17/514,897 US20220050711A1 (en) | 2016-12-06 | 2021-10-29 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/370,477 Continuation US11182203B2 (en) | 2016-12-06 | 2016-12-06 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
Publications (1)
Publication Number | Publication Date |
---|---|
US20220050711A1 true US20220050711A1 (en) | 2022-02-17 |
Family
ID=62244206
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/370,477 Active 2038-12-02 US11182203B2 (en) | 2016-12-06 | 2016-12-06 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
US17/514,897 Abandoned US20220050711A1 (en) | 2016-12-06 | 2021-10-29 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/370,477 Active 2038-12-02 US11182203B2 (en) | 2016-12-06 | 2016-12-06 | Systems and methods to orchestrate infrastructure installation of a hybrid system |
Country Status (1)
Country | Link |
---|---|
US (2) | US11182203B2 (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10462123B2 (en) | 2016-12-06 | 2019-10-29 | Vmware, Inc. | Systems and methods for cloning an agent in a distributed environment |
US11153297B2 (en) | 2016-12-06 | 2021-10-19 | Vmware, Inc. | Systems and methods to facilitate certificate and trust management across a distributed environment |
US10338981B2 (en) | 2016-12-06 | 2019-07-02 | Vmware, Inc | Systems and methods to facilitate infrastructure installation checks and corrections in a distributed environment |
GB201702450D0 (en) | 2017-02-15 | 2017-03-29 | Blue Prism Ltd | System for optimising distribution of processing an automated process |
US11269613B2 (en) * | 2019-02-28 | 2022-03-08 | Pivotal Software, Inc. | Software platform upgrades |
CN110109684B (en) * | 2019-04-04 | 2023-06-09 | 平安科技(深圳)有限公司 | Block chain link point management proxy service installation method, electronic device and storage medium |
US11403147B2 (en) * | 2019-07-16 | 2022-08-02 | Vmware, Inc. | Methods and apparatus to improve cloud management |
US11076019B2 (en) * | 2019-11-15 | 2021-07-27 | F5 Networks, Inc. | Scheduling services on a platform including configurable resources |
US11307902B1 (en) | 2020-09-30 | 2022-04-19 | Kyndryl, Inc. | Preventing deployment failures of information technology workloads |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130232497A1 (en) * | 2012-03-02 | 2013-09-05 | Vmware, Inc. | Execution of a distributed deployment plan for a multi-tier application in a cloud infrastructure |
US20140310258A1 (en) * | 2013-04-15 | 2014-10-16 | Vmware, Inc. | Fault Tolerant Distributed Query Processing Using Query Operator Motion |
US20150347264A1 (en) * | 2014-05-28 | 2015-12-03 | Vmware, Inc. | Tracking application deployment errors via cloud logs |
Family Cites Families (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8464250B1 (en) | 2004-09-23 | 2013-06-11 | Transcontinental Events, Llc | System and method for on-demand cloning of virtual machines |
JP4544146B2 (en) * | 2005-11-29 | 2010-09-15 | 株式会社日立製作所 | Disaster recovery method |
US9015703B2 (en) | 2006-10-17 | 2015-04-21 | Manageiq, Inc. | Enforcement of compliance policies in managed virtual systems |
US8719624B2 (en) | 2007-12-26 | 2014-05-06 | Nec Corporation | Redundant configuration management system and method |
GB2488268B (en) | 2009-12-10 | 2017-10-04 | Ibm | A method and system for managing configurations of system management agents in a distributed environment |
US8789041B2 (en) | 2009-12-18 | 2014-07-22 | Verizon Patent And Licensing Inc. | Method and system for bulk automated virtual machine deployment |
US9134982B2 (en) | 2010-01-10 | 2015-09-15 | Microsoft Technology Licensing, Llc | Automated configuration and installation of virtualized solutions |
US8639923B2 (en) | 2010-05-28 | 2014-01-28 | Dell Products, Lp | System and method for component authentication of a secure client hosted virtualization in an information handling system |
US20120054486A1 (en) | 2010-08-31 | 2012-03-01 | MindTree Limited | Securing A Virtual Environment And Virtual Machines |
US8407323B2 (en) | 2011-07-12 | 2013-03-26 | At&T Intellectual Property I, L.P. | Network connectivity wizard to support automated creation of customized configurations for virtual private cloud computing networks |
WO2013093209A1 (en) | 2011-12-21 | 2013-06-27 | Ssh Communications Security Oyj | Automated access, key, certificate, and credential management |
US20130247136A1 (en) | 2012-03-14 | 2013-09-19 | International Business Machines Corporation | Automated Validation of Configuration and Compliance in Cloud Servers |
US8739257B1 (en) | 2012-03-28 | 2014-05-27 | Emc Corporation | Managing authentication of virtual clients |
US8843935B2 (en) | 2012-05-03 | 2014-09-23 | Vmware, Inc. | Automatically changing a pre-selected datastore associated with a requested host for a virtual machine deployment based on resource availability during deployment of the virtual machine |
US9729517B2 (en) | 2013-01-22 | 2017-08-08 | Amazon Technologies, Inc. | Secure virtual machine migration |
US10846074B2 (en) | 2013-05-10 | 2020-11-24 | Box, Inc. | Identification and handling of items to be ignored for synchronization with a cloud-based platform by a synchronization client |
US9471474B2 (en) | 2013-08-19 | 2016-10-18 | Microsoft Technology Licensing, Llc | Cloud deployment infrastructure validation engine |
US9092248B1 (en) | 2013-08-21 | 2015-07-28 | Symantec Corporation | Systems and methods for restoring distributed applications within virtual data centers |
US9356962B2 (en) | 2013-09-10 | 2016-05-31 | Vmware, Inc. | Extensible multi-tenant cloud-management system and methods for extending functionalities and services provided by a multi-tenant cloud-managment system |
US9350550B2 (en) | 2013-09-10 | 2016-05-24 | M2M And Iot Technologies, Llc | Power management and security for wireless modules in “machine-to-machine” communications |
US9124575B2 (en) | 2013-11-27 | 2015-09-01 | Sap Se | Self-single sign-on |
US9503447B2 (en) | 2014-01-30 | 2016-11-22 | Sap Se | Secure communication between processes in cloud |
US9756118B2 (en) | 2014-04-28 | 2017-09-05 | Vmware, Inc. | Virtual performance monitoring decoupled from hardware performance-monitoring units |
US9858060B2 (en) | 2014-05-09 | 2018-01-02 | International Business Machines Corporation | Automated deployment of a private modular cloud-computing environment |
US9386079B2 (en) | 2014-06-10 | 2016-07-05 | American Megatrends, Inc. | Method and system of virtual desktop infrastructure deployment studio |
US20160142409A1 (en) | 2014-11-18 | 2016-05-19 | Microsoft Technology Licensing, Llc | Optimized token-based proxy authentication |
US20160285832A1 (en) | 2015-03-23 | 2016-09-29 | Petar D. Petrov | Secure consumption of platform services by applications |
US9773122B2 (en) | 2015-11-27 | 2017-09-26 | International Business Machines Corporation | Automated compliance exception approval |
US10050947B2 (en) | 2016-01-28 | 2018-08-14 | Cisco Technology, Inc. | Key distribution in a distributed network environment |
US10462123B2 (en) | 2016-12-06 | 2019-10-29 | Vmware, Inc. | Systems and methods for cloning an agent in a distributed environment |
US11153297B2 (en) | 2016-12-06 | 2021-10-19 | Vmware, Inc. | Systems and methods to facilitate certificate and trust management across a distributed environment |
US10338981B2 (en) | 2016-12-06 | 2019-07-02 | Vmware, Inc | Systems and methods to facilitate infrastructure installation checks and corrections in a distributed environment |
-
2016
- 2016-12-06 US US15/370,477 patent/US11182203B2/en active Active
-
2021
- 2021-10-29 US US17/514,897 patent/US20220050711A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130232497A1 (en) * | 2012-03-02 | 2013-09-05 | Vmware, Inc. | Execution of a distributed deployment plan for a multi-tier application in a cloud infrastructure |
US20140310258A1 (en) * | 2013-04-15 | 2014-10-16 | Vmware, Inc. | Fault Tolerant Distributed Query Processing Using Query Operator Motion |
US20150347264A1 (en) * | 2014-05-28 | 2015-12-03 | Vmware, Inc. | Tracking application deployment errors via cloud logs |
Also Published As
Publication number | Publication date |
---|---|
US20180159721A1 (en) | 2018-06-07 |
US11182203B2 (en) | 2021-11-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11461125B2 (en) | Methods and apparatus to publish internal commands as an application programming interface in a cloud infrastructure | |
US12074774B2 (en) | Methods, systems, and apparatus to scale in and/or scale out resources managed by a cloud automation system | |
US11265366B2 (en) | Lifecycle management of custom resources in a cloud computing environment | |
US11327821B2 (en) | Systems and methods to facilitate infrastructure installation checks and corrections in a distributed environment | |
US20220050711A1 (en) | Systems and methods to orchestrate infrastructure installation of a hybrid system | |
US20210111957A1 (en) | Methods, systems and apparatus to propagate node configuration changes to services in a distributed environment | |
US11153297B2 (en) | Systems and methods to facilitate certificate and trust management across a distributed environment | |
US11509646B2 (en) | Systems and methods for cloning an agent in a distributed environment | |
US11757730B2 (en) | Methods, systems and apparatus for custom interface specification in a cloud management system | |
US11481239B2 (en) | Apparatus and methods to incorporate external system to approve deployment provisioning | |
US11263058B2 (en) | Methods and apparatus for limiting data transferred over the network by interpreting part of the data as a metaproperty | |
US10353752B2 (en) | Methods and apparatus for event-based extensibility of system logic | |
US11552855B2 (en) | Methods, systems and apparatus for dynamically extending a cloud management system by adding endpoint adapter types | |
US11175901B2 (en) | Distribution and execution of instructions in a distributed computing environment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
AS | Assignment |
Owner name: VMWARE LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:VMWARE, INC.;REEL/FRAME:066692/0103 Effective date: 20231121 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |