Skip to main content
Building Your Private Cloud Infrastructure: Converged Datacenter without a Dedicated Storage Node

Veröffentlicht: April 2012

Letzte Aktualisierung: April 2012

Betrifft: Windows Server 2012

[Dieses Thema ist Teil der Dokumentation für die Vorabversion und kann in zukünftigen Versionen geändert werden. Leere Themen sind als Platzhalter enthalten.]

This document contains the instructions that you need to follow to create a private or public cloud configuration that uses:

  • A converged subnet for live migration, cluster, storage, management and tenant traffic.

  • network adapter Teaming for network bandwidth aggregation and failover.

  • Direct attached storage using Serial Attached SCSI (SAS) just a bunch of disks (JBOD) disks.

Overview

A Windows Server® 2012 cloud is a high-performing and highly available Hyper-V cluster that hosts virtual machines that can be managed to create private or public clouds. This document explains how to configure the basic building blocks for such a cloud. It does not cover the System Center or other management software aspects of deployments; the focus is on configuring the core Windows Server hosts that are used to build cloud infrastructure.

For background information on creating clouds using Windows Server 2012, see Building Infrastructure as a Service Clouds using Windows Server 8.

This cloud configuration consists of the following:

  • Multiple computers in a Hyper-V cluster

    Each computer in the Hyper-V cluster must have at least two network adapters that will be used for the converged network. These network adapters will be teamed by using Windows 8 Load Balancing and Failover. These can be either two for more 10 GbE or 1 GbE network adapters that will be used for live migration, cluster, storage, management and tenant traffic.

  • The appropriate networking hardware to connect all of the computers in the Hyper-V cluster to each other and to a larger network from which the hosted virtual machines are available.

This configuration uses the following technologies and features of Windows Server 2012:

  • Load Balancing with Failover (LBFO): Load Balancing and Failover logically combines multiple network adapters to provide bandwidth aggregation and traffic failover to prevent connectivity loss in the event of a network component failure. Load Balancing with Failover is also known as network adapter Teaming.

  • Quality of Service (QoS): The packet scheduling facility of QoS in Windows Server 2012 can provide performance guarantees of the different types of traffic of the Hyper-V cluster. Windows 8 QoS can be used to ensure that each traffic flow hosted by each cluster member has a guaranteed amount of bandwidth available.

  • Remote Direct Memory Access (RDMA): RDMA is a new feature for Windows Server "8" Beta. RDMA is an accelerated input-output (I/O) delivery model that lets application software bypass most software layers to communicate directly with the computer hardware, which improves application performance and reduces delay.

  • Hyper-V Virtual Switch QoS: In Windows Server 2012, QoS includes new bandwidth management features that lets you provide predictable network performance to virtual machines on a server running Hyper-V.

  • Storage spaces: Storage spaces make it possible for you to create cost-effective disk pools that present themselves as a single mass storage location on which virtual disks or volumes can created and formatted.

Configuration scenario overview

The Converged Datacenter without Dedicated Storage Node scenario includes the following:

  • A single network connection consisting of two or more teamed adapters over which live migration, cluster, management and tenant traffic traverses.

  • All traffic, including both host traffic and tenant traffic, traverses the Hyper-V virtual switch.

  • Local storage, consisting of a SAS connection to a JBOD configuration.

noteHinweis

Although this configuration uses SAS storage, you can easily choose to use other types of storage, such as SAN storage. You can find more information about storage configuration for a non-SAS scenario in the document Building Your Cloud Infrastructure: Non-Converged Enterprise Configuration which describes how to configure the SAN storage and points out that you do not need an RDMA capable network adapter because storage traffic will not use the SMB 3.0 protocol.

The following figure provides a high-level view of the scenario layout. The teamed network adapters are connected to what will be referred to as a converged subnet in this document. We use the term converged subnet to make it clear that all traffic to and from the Hyper-V cluster members must go through the teamed converged subnet network adapter. Both the host operating system and the tenants connect to the network through the Hyper-V virtual switch. The figure also shows an optional network adapter that is RDMA capable that can be used for storage traffic. This document does not discuss this optional configuration option.

Figure 1   High level overview of cluster member networking configuration

noteHinweis

At least one Active Directory Domain Services (AD DS) domain controller is needed for centralized security and management of the cluster member computers (not shown). It must be reachable by all of the cluster member computers, including the members of the shared storage cluster.

Figure 2 provides an overview of traffic flows on each member of the Hyper-V cluster. The figure shows that each cluster node member uses a virtual network adapter to connect to the Hyper-V Extensible Switch, which connects it to the physical network. Each tenant virtual machine is also connected to the Hyper-V Extensible Switch using a virtual network adapter. Network adapters named ConvergedNet1 and ConvergedNet2 participate in a teamed physical network adapter configuration using the Windows Server 2012 Failover and Load Balancing feature. Windows Server 2012 Hyper-V virtual switch QoS is used to assure that each traffic type (such as live migration, cluster, management and tenant) has a predictable amount of bandwidth available. Traffic isolation is enabled by 802.1q VLAN tagging so that host traffic is not visible to the tenants.

noteHinweis

Virtual local area networks (VLANs) are not assigned to each tenant because VLAN-based network isolation is not a scalable solution and is not compatible with Windows Server 2012 network virtualization.

Figure 2   Overview of cluster member traffic flows

The following sections describe how to set up this cloud configuration using UI-based tools and Windows PowerShell.

After the cloud is built, you can do the following:

  • Install and configure virtual machines

  • Migrate running virtual machines between servers in the Hyper-V cluster (live migration)

  • Failover storage from one owner to another

Install and configure

Creating this cloud infrastructure configuration consists of the following steps:

  • Step 1: Initial Node Configuration

  • Step 2: Initial Network Configuration

  • Step 3: Initial Storage Configuration

  • Step 4: Failover Cluster Setup

  • Step 5: Hyper-V Configuration

  • Step 6: Cloud Validation

The table below summarizes the steps that this document describes:

 

Step

Task

Target

Tasks

1

Initial Node Configuration

All Nodes

  • Add appropriate VLANs to interface ports on the physical switch

    • Management (untagged, default)

    • Tenants (tagged)

    • Live migration (tagged)

    • Cluster/cluster shared volumes (CSV) (tagged)

  • Enable BIOS settings required for Hyper-V

  • Clean operating systems install followed by:

    • Set PS execution policy

    • Enable Windows PowerShell remoting

    • Enable Remote Desktop Protocol and Firewall rule

    • Join the domain

  • Install roles and features using default settings, rebooting as needed

    • Hyper-V (plus tools)

    • Storage Services

    • Failover clustering (plus tools)

    • Share and storage management tools

2

Initial Network Configuration

All Nodes

  • Disable unused and disconnected interfaces and rename active connections

  • Create the converged network adapter team (rename as necessary)

  • Create the Hyper-V vSwitch and Management virtual network adapter (PS)

  • Rename Management virtual network adapter

  • Create additional virtual network adapters and assign VLAN IDs (PS)

    • Live migration

    • Cluster

  • Rename network adapters

  • Assign static IPs as necessary

  • Configure traffic performance isolation (QoS) on the converged network across the network adapters

  • Configure the default minimum bandwidth for the switch

3

Initial Storage Configuration

Single Node

  • Present all shared storage to relevant nodes

  • For multipath scenarios, install and configure multipath I/O (MPIO) as necessary

  • All shared disks: Wipe, bring online and initialize

4

Failover Cluster Setup

Single Node

  • Run through the Cluster Validation Wizard

  • Address any indicated warnings and/or errors

  • Complete the Create Cluster Wizard (setting name and IP but do not add eligible storage)

  • Verify cluster quorum configuration and modify as necessary

  • Configure cluster networks to separate traffic via PS (confirm via UI)

  • Create a clustered pool

  • Create a clustered virtual disk in the clustered pool

  • Create a NTFS file system volume (no drive letter) on the clustered virtual disk

  • Add the virtual disk to CSV

  • Add folders to the clustered volume:

    • C:\ClusterStorage\Volume1\Virtual Hard Disks\

    • C:\ClusterStorage\Volume1\Hyper-V\

  • Configure cluster network metrics

5

Hyper-V Configuration

All Nodes

  • Change default file locations, mapping to CSV volumes

    • C:\ClusteredStroage\Volume1\Virtual Hard Disks\

    • C:\ClusterStorage\Volume1\Hyper-V\

  • Verify live migration traffic order

6

Cloud Validation

Single Node

  • Create a virtual machine, attaching an existing operating system VHD and tagging to the appropriate VLAN

  • Startup the virtual machine

  • Quick migrate and/or live migrate to other nodes in the cluster

  • Failover storage (CSV) to other nodes in the cluster

Step 1: Initial node configuration

In step 1, you will perform the following steps on all nodes of the Hyper-V cluster:

  • Add appropriate VLANs to Interface ports on the physical switch.

  • Enable BIOS settings required for Hyper-V.

  • Perform a clean operating system installation.

  • Perform post installation tasks.

  • Install roles and features using the default settings.

1.1 Add appropriate VLANS to the interface ports on the physical switch

Cluster nodes will be configured to use different VLAN for the following traffic types:

  • Management traffic – untagged/default

  • Tenant traffic – tagged

  • Live migration traffic – tagged

  • Cluster and CSV traffic – tagged

VLANs are configured to enable traffic isolation and quality of service policies. Define the VLAN tag numbers for each traffic type and then configure your switch with the appropriate VLAN port numbers. The procedures for doing this will vary with the switch make and model. Please refer to your switch documentation for more information.

1.2 Enable BIOS settings required for Hyper-V

You will need to enable virtualization support in the BIOS of each cluster member prior to installing the Hyper-V server role. The procedure for enabling processor virtualization support will vary with your processors' make and model and the system BIOS. Please refer to your hardware documentation for the appropriate procedures.

1.3 Perform a clean operating system installation

Install Windows Server 2012, naming the computer its designated name.

1.4 Perform post-installation tasks

There are several tasks you need to complete on each node after the operating system installation is complete. These include:

  • Join each node to the domain

  • Enable remote access to each node via the Remote Desktop Protocol.

  • Set the Windows PowerShell execution policy.

  • Enable Windows PowerShell remoting.

Perform the following steps to join each node to the domain:

  1. Press the Windows Key on the keyboard and then press R. Open Control Panel and then click OK.

  2. In the Control Panel window, click System and Security, and then click System.

  3. In the System window under Computer name, domain, and workgroup settings, click Change settings.

  4. In the System Properties dialog box, click Change.

  5. Under Member of, click Domain, type the name of the domain, and then click OK.

Perform the following steps to enable remote access to each node using the Remote Desktop Protocol:

  1. Run the following Windows PowerShell command:

 

(Get-WmiObject Win32_TerminalServiceSetting -Namespace root\cimv2\terminalservices).SetAllowTsConnections(1,1)

Perform the following step to set the Windows PowerShell execution policy:

  1. Run the following Windows PowerShell command:

 

Set-ExecutionPolicy Unrestricted -Force

Perform the following step to enable Windows PowerShell remoting:

  1. Run the Windows PowerShell command:

 

Enable-PSRemoting –Force
1.5 Install roles and features using the default settings

The following roles and features will be installed on each cluster node:

  • Hyper-V and Hyper-V management Tools

  • Data Center Bridging (DCB)

  • Failover cluster and failover cluster management tools

  • Storage management tools

Perform the following steps on each cluster node to install the required roles and features:

  1. In Server Manager, click Dashboard in the console tree.

  2. In Welcome to Server Manager, click 2 Add roles and features, and then click Next.

  3. On the Before You Begin page of the Add Roles and Features Wizard, click Next.

  4. On the Installation Type page, click Next.

  5. On the Server Selection page, click Next.

  6. On the Server Roles page, select Hyper-V from the Roles list. In the Add Roles and Features Wizard dialog box, click Add Features. Click Next.

  7. On the Features page, select Data Center Bridging and Failover Clustering from the Features list. Expand Remote Server Administrator Tools and then expand Role Administration Tools. Expand File Services Tools. Select Share and Storage Management Tools. Click Next.

    noteHinweis
    If you plan to use Multipath I/O for your storage solution, select the Multipath I/O feature.

  8. On the Hyper-V page, click Next.

  9. On the Virtual Switches page, click Next.

  10. On the Migration page, put a checkmark in the Allow this server to send and receive live migrations of virtual machines, and then confirm that Use CredSSP is enabled. Click Next.

  11. On the Default Stores page, click Next.

  12. On the Confirm installation selections page put a checkmark in the Restart the destination server automatically if required checkbox and then in the Add Roles and Features dialog box click Yes, then click Install.

  13. On the Installation progress page, click Close after the installation has succeeded.

  14. Restart the computer. This process might require restarting the computer twice. If so, the installer will trigger the multiple restarts automatically.

After you restart the server, open Server Manager and confirm that the installation completed successfully. Click Close on the Installation Progress page.

Step 2: Initial network configuration

The network configuration on each node needs to be configured to support the converged networking scenario. You will perform the following procedures on each of the nodes to complete the initial network configuration:

  • Disable unused and disconnected interfaces and rename active connections.

  • Create a converged network adapter team.

  • Create the Hyper-V virtual switch and management network adapter.

  • Rename the management virtual network adapter.

  • Create additional virtual network adapters and assign VLAN IDs.

  • Rename virtual network adapters.

  • Assign static IP addresses to the virtual network adapters.

  • Configure traffic performance isolation using Hyper-V virtual switch QoS.

2.1 Disable unused and disconnected interfaces and rename active connections

You can simplify the configuration and avoid errors when running the wizards by disabling all interfaces that are either unused or disconnected. You can disable these interfaces in the Network Connections window.

For the remaining network adapters that will participate in the converged subnet, do the following:

  1. Connect them to the converged network switch ports.

  2. To help you more easily recognize the active network adapters, rename them with names that indicate their use or their connection to the intranet or Internet (for example, ConvergedNet1 and ConvergedNet2). You can do this in the Network Connections window.

2.2 Create a converged network adapter team

Network Load Balancing and Failover (LBFO) enables bandwidth aggregation and network adapter failover to prevent connectivity loss in the event of a network card or port failure. This feature is often referred to as "network adapter teaming". In this scenario you will create one team that will be connected to the ConvergedNet subnet.

To configure the network adapter teams by using Server Manager, do the following on each computer in the cluster:

noteHinweis

Several steps in the following procedure will temporarily interrupt network connectivity. We recommend that all servers be accessible over a KVM so that you can check on the status of these machines if network connectivity is unavailable for more than five minutes.

  1. From Server Manager, click Local Server in the console tree.

  2. In Properties, click Disabled, which you’ll find next to Network adapter Teaming.

  3. In the network adapter Teaming window, click the name of the server computer in Servers

  4. In Teams, click Tasks, and then click New Team.

  5. In the New Team window, in the Team Name text box, enter the name of the network adapter team for the converged traffic subnet (example: ConvergedNet Team).

  6. In the Available adapters list select the two network adapters connected to the converged traffic subnet (in this example, ConvergedNet1 and ConvergedNet2), and then click OK. Note that there may be a delay of several minutes before connectivity is restored after making this change. To ensure that you see the latest state of the configuration, right click your server name in the Servers section and click Refresh Now. There may be a delay before the connection display as Active. You may need to refresh several times before seeing the status change.

  7. Close the network adapter Teaming window.

Configure a static IPv4 address configuration for the new network adapter team connected to the converged traffic subnet (example: ConvergedNet Team). This IP address is the one that you will use when connecting to the host system for management purposes. You can do this in the Properties of the team in the Network Connections window. You will see a new adapter where the name of the teamed network adapter is the name you assigned in step 5. You will lose connectivity for a few moments after assigning the new IP addressing information.

noteHinweis

You might need to manually refresh the display of the network adapter Teaming window to show the new team and there may be a delay in connectivity as the network adapter team is created. If you are managing this server remotely, you might temporarily lose connectivity to the server.

2.3 Create the Hyper-V virtual switch and management virtual network adapter

In this scenario, all traffic will flow through the Hyper-V virtual switch. This includes the host operating system traffic (cluster/CSV, management and live migration) and guest/tenant traffic. You will create the virtual switch in Windows PowerShell instead of using the Hyper-V console. The reason for this is that when you create the virtual switch in the Hyper-V console, you are unable to specify the Minimum Bandwidth Mode, which defaults to Absolute (requiring a bits per second value) as opposed to Weighted mode (configurable via Windows PowerShell) which allows for a relative range between 1 and 100. For more information on the New-VMSwitch cmdlet, please see http://technet.microsoft.com/en-us/library/hh848455.aspx.

Perform the following step to create the Hyper-V virtual switch and the management traffic virtual network adapter:

  1. Run the following Windows PowerShell command:

 

New-VMSwitch "ConvergedNetSwitch" -MinimumBandwidthMode weight -NetAdapterName "ConvergedNetTeam" -AllowManagementOS 1
2.4 Rename the management virtual network adapters

The management virtual network adapter that was created when you created the virtual switch now appears in the Network Connections window and it was assigned a generic name. You should rename this virtual network adapter to make it easier to identify in subsequent operations. Right-click the new virtual network adapter and click Rename and assign the virtual network adapter a new name (for example, Management).

2.5 Create additional virtual network adapters and assign VLAN IDs

The Hyper-V virtual switch now has a single virtual network adapter that will be used for host operating system management traffic. You now will create two additional virtual network adapters: one for live migration traffic and one for cluster traffic.

Perform the following steps to create the live migration and the cluster network adapters and assign these virtual network adapters to a VLAN ID :

  1. Run the following Windows PowerShell command to create the live migration traffic virtual network adapter:

    Add-VMNetworkAdapter -ManagementOS -Name "LiveMigration" -SwitchName "ConvergedNetSwitch"
  2. Run the following Windows PowerShell command to create the cluster traffic virtual network adapter:

    Add-VMNetworkAdapter -ManagementOS -Name "Cluster" -SwitchName "ConvergedNetSwitch"


  3. Run the following Windows PowerShell command to assign the live migration virtual network adapter a VLAN ID:

    Set-VMNetworkAdapterVlan -ManagementOS -VMNetworkAdapterName LiveMigration -Access -VlanId 2160
    In the preceding example, –VMNetworkAdapterName represents the name of the virtual network adapter.

  4. Run the following Windows PowerShell command to assign the cluster virtual network adapter a VLAN ID:

    Set-VMNetworkAdapterVlan -ManagementOS -VMNetworkAdapterName Cluster -Access -VlanId 2161
    In the preceding example, the –VMNetworkAdapterName represents the name of the virtual network adapter.

2.6 Rename virtual network adapters

The live migration and cluster virtual network adapters you created in the previous step now appear in the Network Connections window and they were assigned generic names. You should rename these virtual network adapters to make them easier to identify in subsequent operations. Right-click each of these virtual network adapters, click Rename and assign the virtual network adapter a new name (for example, live migration and cluster)

2.7 Assign static IP addresses to the virtual network adapters

You now need to assign IP addresses to your virtual network adapters. This can be done through DHCP or you can assign static addresses. Make sure that each of the virtual network adapters is assigned an IP address on a different network ID – this will become important later when you configure your cluster networking configuration. You can use the Networking Control Panel applet or Windows PowerShell to assign IP addressing information to the virtual network adapters. For example:

Set-NetIPInterface -InterfaceAlias "LiveMigration" -dhcp Disabled; new-NetIPAddress -PrefixLength 8 -InterfaceAlias "LiveMigration" -IPv4Address 11.0.0.x
2.8 Configure traffic performance isolation by using QoS

In this step you will configure QoS weightings that define the minimal share of bandwidth assigned to each of the virtual network adapters. You can determine the percentage of bandwidth that can be allocated to a particular virtual network adapter by adding all the weight values together and then dividing the individual weight assigned to a virtual network adapter by the total.

Perform the following steps to add weight values to the cluster, management and live migration virtual network adapters:

  1. Run the following Windows PowerShell command to assign a weight value to the cluster virtual network adapter:

    Set-VMNetworkAdapter -ManagementOS -Name "Cluster" -MinimumBandwidthWeight 40
  2. Run the following Windows PowerShell command to assign a weight value to the management virtual network adapter:

    Set-VMNetworkAdapter -ManagementOS -Name ConvergedNetSwitch -MinimumBandwidthWeight 5
    noteHinweis
    VMNetworkAdapter name is listed under 'Device Name' in the Network Connections user interface.

  3. Run the following Windows PowerShell command to assign a weight value to the live migration virtual network adapter:

    Set-VMNetworkAdapter -ManagementOS -Name "LiveMigration" -MinimumBandwidthWeight 20
  4. Run the following Windows PowerShell command to assign a default weight value for any future virtual network adapters you create:

    Set-VMSwitch "ConvergedNetSwitch" -DefaultFlowMinimumBandwidthWeight 10
Step 3: Initial storage configuration

With the initial node configuration complete, you are ready to perform initial storage configuration tasks on all nodes of the cluster. Initial storage configuration tasks include:

  • Present all shared storage to relevant nodes

  • Install and configure MPIO as necessary for multipath scenarios

  • Wipe, bring online, and initialize all shared disks

  • Install required roles and features on the cluster

3.1 Present all shared storage to relevant nodes

In a SAS scenario, connect the SAS adapters to each storage device. Each cluster node should have two adapters in them if high availability of storage access is required.

3.2 Install and configure MPIO as necessary for multipath scenarios

If you have multiple data paths to storage (for example, two SAS cards) make sure to install the Microsoft® Multipath I/O (MPIO) on each node. This step might require you to restart the system. For more information about MPIO, see What's New in Microsoft Multipath I/O.

3.3 Wipe, bring online, and initialize all shared disks

To prevent issues with the storage configuration procedures that are detailed later is this document, confirm that the disks in your storage solution have not been previously provisioned. The disks should have no partitions or volumes. They should also be initialized so that there is a master book file (MBF) and GUID partition table (GPT) record on the disks and then brought online. You can use the Disk Management console, Windows PowerShell or Diskpart.exe to accomplish this task. This task must be completed on each node in the cluster.

noteHinweis

If you have previously configured these disks with Windows 8 Storage Spaces pools, you will need to delete these storage pools prior to proceeding with the storage configuration described in this document

Step 4: Failover cluster setup

You are now ready to complete the failover cluster settings. Failover cluster setup includes the following steps:

  • Complete the Failover Cluster Validation Wizard.

  • Address any indicated warnings and/or errors.

  • Complete the Create Cluster Wizard.

  • Perform manual cluster tuning as needed.

  • Create a clustered pool.

  • Create a clustered virtual disk in the clustered pool.

  • Create a NTFS file system volume on the clustered virtual disk.

  • Add the virtual disk to Cluster Shared Volumes (CSV).

  • Add folders to the clustered volume.

  • Configure cluster networks to separate traffic.

4.1 Run through the Cluster Validation Wizard

The Cluster Validation Wizard will query multiple components in the Hyper-V array and confirm that the hardware and software is ready to support failover clustering. On one of the nodes in the server cluster, perform the following steps to run the Cluster Validation Wizard:

  1. In the Server Manager click Tools and then click Failover Cluster Manager.

  2. In the Failover Cluster Manager console, in the Management section, click Validate Configuration.

  3. On the Before You Begin page of the Validate a Configuration Wizard, click Next.

  4. On the Select Servers or a Cluster page, type the name of the local server, and then click Add. After the name appears in the Selected servers list, type the name of another Hyper-V cluster member computer, and then click Add. Repeat this step for all computers in the Hyper-V cluster. When all of the servers of the Hyper-V cluster appear in the Selected servers list, click Next.

  5. On the Testing Options page, click Next.

  6. On the Confirmation page, click Next. The time to complete the validation process will vary with the number of nodes in the cluster and can take some time to complete.

  7. On the Summary page, the summary text will indicate that the configuration is suitable for clustering. Confirm that there is a checkmark in the Create the cluster now using the validated nodes... checkbox.

4.2 Address any indicated warnings and/or errors

Click the Reports button to see the results of the Cluster Validation. Address any issues that have led to cluster validation failure. After correcting the problems, run the Cluster Validation Wizard again. After the cluster passes validation, then proceed to the next step. Note that you may see errors regarding disk storage. You may see this if you haven't yet initialized the disks. Click Finish.

4.3 Complete the Create Failover Cluster Wizard

After passing cluster validation, you are ready to complete the cluster configuration.

Perform the following steps to complete the cluster configuration:

  1. On the Before You Begin page of the Create Cluster Wizard, click Next.

  2. On the Access Point for Administering the Cluster page, enter a valid NetBIOS name for the cluster, and then select the network you want the cluster on and then type in a static IP address for the cluster, and then click Next.

  3. On the Confirmation page, clear Add all eligible storage to the cluster and then click Next.

  4. On the Creating New Cluster page you will see a progress bar as the cluster is created.

  5. On the Summary page, click Finish.

  6. In the console tree of the Failover Cluster Manager snap-in, open the Networks node under the cluster name.

  7. Right-click the cluster network that corresponds to the Managment network adapter network ID (subnet), and then click Properties. On the General tab, confirm that Allow cluster communications on this network is selected and that Allow clients to connect through this network is enabled. In the Name text box, enter a friendly name for this network (for example, ManagmentNet), and then click OK.

  8. Right-click the cluster network that corresponds to the Cluster network adapter network ID (subnet) and then click Properties. On the General tab, confirm that Allow cluster cluster communications on this network is selected and that Allow clients to connect through this network is not enabled. In the Name text box, enter a friendly name for this network (e.g, ClusterNet), and then click OK.

  9. Right-click the cluster network that corresponds to the live migration network adapter network ID (subnet) and then click Properties. On the General tab, confirm that Allow cluster cluster communications on this network is selected and that Allow clients to connect through this network is not enabled. In the Name text box, enter a friendly name for this network (e.g, LiveMigrationNet), and then click OK.

4.4 Perform manual cluster tuning as needed

In most situations, use the quorum configuration that the cluster software identifies as appropriate for your cluster. Change the quorum configuration only if you have determined that the change is appropriate for your cluster. For more information about quorum configuration see Understanding Quorum Configurations in a Failover Cluster.

4.5 Create a clustered storage pool

Perform the following steps on one of the members of the storage cluster to create the storage pool:

  1. In the left pane of the Failover Cluster Manager, expand the server name and then expand the Storage node. Click Storage Pools.

  2. In the Actions pane, click Create Storage Pool.

  3. On the Before You Begin page, click Next.

  4. On the Storage Pool Name page, enter a name for the storage pool in the Name text box. Enter an option description for the storage pool in the Description text box. In the Available Drives Pool list, select the name you assigned to the cluster (this is the NetBIOS name you assigned to the cluster when you created the cluster). Click Next.

  5. On the Physical Drives page, select the drives that you want to participate in the storage pool. Then click Next.

  6. On the Confirmation page, confirm the settings and click Create.

  7. On the Results page, you should receive the message You have successfully completed the New Storage Pool Wizard. Put a checkmark in the Create a virtual disk when the wizard closes checkbox. Then click Close.

4.6 Create a clustered virtual disk in the clustered pool

Now that you have created the storage pool, you can create a virtual disk within that storage pool.

Perform the following steps to create the virtual disk:

  1. On the Before You Begin page in the New Virtual Disk Wizard, click Next.

  2. On the Storage Pool page, select your server name in the Server section and then select the storage pool you created earlier in the Storage pool section. Click Next.

  3. On the Virtual Disk Name page, enter a name for the virtual disk in the Name text box. You can also enter an optional description in the Description text box. Click Next.

  4. On the Storage Layout page, in the Layout section, select Mirror. Click Next.

  5. On the Provisioning page, select the Fixed option. Click Next. Note that when using Mirrored storage layout, you must use the fixed provisioning option.

  6. On the Size page, in the Virtual disk size text box, enter a size for the new virtual disk. Use the drop down box to select MB, GB or TB. Click Next.

  7. On the Confirmation page, review your settings and click Create.

  8. On the Results page, put a checkmark in the Create a volume when this wizard closes checkbox. Click Close.

4.7 Create an NTFS volume with no drive letter on the clustered virtual disk

Now that the virtual disk is created, you can create a volume on that disk. Perform the following steps to create a volume on your virtual disks:

  1. On the Before You Begin page of the New Volume Wizard, click Next.

  2. On the Server and Disk page, select the name of the cluster from the Servers list. In the Disk section, select the virtual disk you created in the previous step. You can identify this disk by looking in the Virtual Disk column, where you will see the name of the virtual disk you created. Click Next.

  3. On the Size page, accept the default volume size, and click Next.

  4. On the Drive Letter or Folder page, select the Don't assign to a drive letter or folder. Click Next.

  5. On the File System Settings page, from the File system drop down list, select NTFS. Use the default setting in the Allocation unit size list. Note that ReFS is not supported in a cluster shared volume configuration. Click Next.

  6. On the Confirmation page, click Create.

  7. On the Results page, click Close.

4.8 Add the virtual disk to Cluster Shared Volumes

The virtual disk you created is now ready to be added to a Cluster Shared Volume. Perform the following steps to add the virtual disk to a Cluster Shared Volume.

  1. In the Failover Cluster Manager, in the left pane of the console, click the Storage node. In the middle pane of the console, in the Storage section, right click the virtual disk you created in the previous step and then click Add to Cluster Shared Volumes.

  2. Proceed to procedure 4.9.

4.9 Add folders to the clustered volume

Now you need to create the folders on the virtual disk to store the virtual machine files and the virtual machine data files.

Perform the following steps to create a file share to store the running VMs of the Hyper-V cluster:

  1. Open Windows Explorer and navigate to the C: drive and then double-click Cluster Storage and then double-click Volume 1.

Create two folders in Volume 1. One of the folders will contain the .vhd files for the virtual machines (for example, VHDdisks) and one folder will contain the virtual machine configuration files (for example, VHDsettings)

4.10 Configure cluster networks to separate traffic

Add the commands to rename the cluster networks.

The cluster will use the network with the lowest metric for CSV traffic and the second lowest metric for live migration. Windows PowerShell® is the only method available to prescriptively specify the CSV network. You can set the live migration network via the Hyper-V management console, which you will do in section Step 5: Configure Hyper-V settings.

Perform the following steps to configure the cluster network metric values:

  1. Run the following Windows PowerShell command to set the metric for the cluster network traffic:

    (Get-ClusterNetwork "ClusterNet" ).Metric = 100
  2. Run the following Windows PowerShell command run set the metric for the live migration network traffic:

    (Get-ClusterNetwork "LiveMigrationNet" ).Metric = 500
  3. Run the following Windows PowerShell command to set the metric for the management network traffic:

    (Get-ClusterNetwork "ManagementNet" ).Metric = 1000
Step 5: Configure Hyper-V settings

To finalize the Hyper-V configuration, you will need to:

  • Change the default file locations for the .vhdx and configuration files

  • Verify the live migration traffic order

5.1 Change default file locations for virtual machine files

On each Hyper-V cluster member, perform the following steps to change the default file locations for virtual machine files:

  1. In Server Manager, click Tools, then click Hyper-V Manager.

  2. From the console tree of the Hyper-V Manager, right-click the name of the Hyper-V server, and then click Hyper-V Settings.

  3. In the Hyper-V Settings dialog box, click Virtual Hard Disks under Server, type the file share location in Specify the default folder to store virtual hard disk files, and then click Apply.

  4. Click Virtual Machines under Server, type the file folder location in Specify the default folder to store virtual machine configuration files, and then click OK

5.2 Verify Live migration Traffic Order

You can use the Hyper-V management console to configure which networks will be used for live migration traffic.

Perform the following steps to configure the network preference order for live migration traffic:

  1. In the left pane of the Hyper-V Manager console, right-click the server name, and then click Hyper-V Settings.

  2. In the Hyper-V Settings window, in the Server section, click Live Migrations.

  3. In the Incoming Live Migrations section, confirm that the live migration network is listed on the top of the list of network IDs. If not, use the Move Up button to move it to the top.

  4. Click OK and close the Hyper-V Manager.

Step 6: Cloud validation

To verify the configuration of your cloud environment, perform the following operations.

  • Create a new virtual machine in the cluster environment.

  • Test network connectivity from the virtual machine.

  • Perform a live migration.

  • Perform a quick migration.

6.1 Create a new virtual machine

To create a new virtual machine in the cluster environment, perform the following steps.

  1. Open Failover Cluster Manager, click Roles under the cluster name, click Virtual Machines under the Actions pane, and then click New Virtual Machine.

  2. On the New Virtual Machine page, select the cluster node where you want to create the virtual machine, and then click OK.

  3. On the Before you Begin page of the New Virtual Machine Wizard, click Next.

  4. On the Specify Name and Location page, enter a friendly name for this virtual machine and then click Next.

  5. On the Assign Memory page, enter the amount of memory that will be used for this virtual machine (minimum for this lab is 1024 MB RAM) and then click Next.

  6. On the Configuring Networking page, select the ConvergeNetSwitch and then click Next.

  7. On the Connect Virtual Hard Disk page, leave the default options selected and click Next.

  8. On the Installation Options page, select Install an operating system from a boot CD/DVD-ROM and then select the location where the CD/DVD is located. If you are installing the new operating system based on an ISO file, make sure to select the option Image file (.iso) and browse for the file location. After you select the appropriate option for your scenario, click Next.

  9. On the Completing the New Virtual Machine Wizard page, review the options, and then click Finish.

  10. The virtual machine creation process starts. After it is finished, you will see the Summary page, where you can access the report created by the wizard. If the virtual machine was created successfully, click Finish.

At this point your virtual machine is created and you should use the Failover Cluster Manager to start the virtual machine and perform the operating system installation according to the operating system that you choose. For the purpose of this validation, the guest operating system can be any Windows Server version.

6.2 Test network connectivity from the virtual machine

Once you finish installing the operating system in the virtual machine you should log on and verify if this virtual machine was able to obtain IP address from the enterprise network. Assuming that in this network you have a DHCP server, this virtual machine should be able to obtain the IP address. To perform the basic network connectivity test use the following approach.

  • Use ping command for a reachable IP address in the same subnet.

  • Use ping command for the same destination but now using the full qualified domain name for the destination host. The goal here is to test basic name resolution.

noteHinweis

If you installed Windows 8 Developer in this virtual machine you need to open Windows Firewall with Advanced Security and create a new rule to allow Internet Control Message Protocol (ICMP) before performing the previous tests. This may be true for other hosts you want to ping − confirm that the host-based firewall on the target allows for ICMP Echo Requests.

After you confirm that this basic test is working properly, leave a command prompt window open and enter the command ping <Destination_IP_Address_or_FQDN> -t. The goal here is to have a continuous test while you perform the live migration to the second node.

6.3 Perform a live migration

To perform a live migration of this virtual machine from the current cluster node to the other node in the cluster, perform the following steps.

  1. In the Failover Cluster Manager, click Roles under the cluster name. On the Roles pane, right click the virtual machine that you created, click Move, click Live Migration, and then click Select Node.

  2. On the Move Virtual Machine page, select the node that you want to move the virtual machine to and click OK.

You will notice in the Status column when the live migration starts, it will take some time for the Information column to update the current state of the migration. While the migration is taking place you can go back to the virtual machine that has the ping running and observe if there is any packet loss.

6.4 Perform a quick migration

To perform the quick migration of this virtual machine from the current node to the other one, perform the following steps.

  1. On the Failover Cluster Manager, click Roles under the cluster name. In the Roles pane, right-click the virtual machine that you created, click Move, click Quick Migration and then click Select Node.

  2. On the Move Virtual Machine window, select the node that you want to move the virtual machine to, and then click OK.

You will notice in the status that the quick migration will start faster than the live migration did. While the migration is taking place you can go back to the virtual machine that has the ping running and observe if there is any packet loss.

Microsoft führt eine Onlineumfrage durch, um Ihre Meinung zur -Website zu erfahren. Wenn Sie sich zur Teilnahme entscheiden, wird Ihnen die Onlineumfrage angezeigt, sobald Sie die -Website verlassen.

Möchten Sie teilnehmen?