Hi
Is it possible to migrate a VM from another node in cluster to currently active node of hyper V Cluster?
Any clue is highly appreciated.
Thanks
Hi
Is it possible to migrate a VM from another node in cluster to currently active node of hyper V Cluster?
Any clue is highly appreciated.
Thanks
Hi,
I want to use these two commands. 1) Start-ClusterResource and 2) Move-ClusterVirtualMachineRole for first starting the VM and then moving to the active host.
If I move VM without starting it, it does not move. so first I start it and the move it.. It works but
How can i do this using WMI? which are their's WMI equivalents?
Please help
Hi Community,
I want to migrate a VM from node B to node A, where node A is current node.
I used WMI Providers MSCluster_Resource.
I first used BringOnline Method to make the resource online.
Then I am trying to use ExecuteResourceControl method with control code: clusctl-resource-vm-start-migration
(https://docs.microsoft.com/en-us/previous-versions/windows/desktop/mscs/clusctl-resource-vm-start-migration)
but this code seems not be existing while docs show that this is supported from 2012 onwards.
Any help is appreciated on this..
MY IDEA IS TO RUN ONE OS IN MORE HYPER-V MACHINES
I have build two hyper-v machines[VMCSV, VM11] with vhds (sharing in both machines), i installed win server 2019 core inside vhds and build cluster sharing volume (CSV) AND but the OS of vhds in the CSV... it's run good in both machines as same time .....Awesome!!
The problem::
when i try to used one of the machines the other shown BLUE SCREEN ERROR.. why??
and if there is any solutions to used both machines as same time please write down
i have pictures of issues but didnt allow to submit until verify my account!!
Need a little help with my S2D cluster which is not performing as I had expected.
Details:
5 x Supermicro SSG-2028R-NR48N servers with 2 x Xeon E5-2643v4 CPUs and 96GB RAM
Each node has 24 x Intel P3520 1.2TB NVME SSDs
The servers are connected over an Infiniband 40Gb network, RDMA is enabled and working.
All 120 SSDs are added to S2D storage pool as data disks (no cache disks). There are two 30TB CSVs configured with hybrid tiering (3TB 3-way mirror, 27TB Parity)
I know these are read intensive SSDs and that parity write performance is generally pretty bad but I was expecting slightly better numbers then I'm getting:
Tested using CrystalDiskMark and diskspd.exe
Multithreaded Read speeds: < 4GBps (seq) / 150k IOPs (4k rand)
Singlethreaded Read speeds: < 600MBps (seq)
Multithreaded Write speeds: < 400MBps (seq)
Singlethreaded Write speeds: < 200MBps (seq) / 5k IOPS (4k rand)
I did manage to up these numbers by configuring a 4GB CSV cache on the CSVs and forcing write through on the CSVs:
Max Reads: 23GBps/500K IOPs 4K IOPS, Max Writes:2GBps/150K 4KIOPS
That high read performance is due to the CSV cache which uses memory. Write performance is still pretty bad though. In fact it's only slight better than the performance I would get for a single one of these NVME drives. I was expecting much better performance from 120 of them!
I suspect that the issue here is that Storage Spaces is not recognising that these disks have PLP protection which you can see here:
Get-storagepool "*S2D*" | Get-physicaldisk |Get-StorageAdvancedProperty FriendlyName SerialNumber IsPowerProtected IsDeviceCacheEnabled ------------ ------------ ---------------- -------------------- NVMe INTEL SSDPE2MX01 CVPF7165003Y1P2NGN False WARNING: Retrieving IsDeviceCacheEnabled failed with ErrorCode 1. NVMe INTEL SSDPE2MX01 CVPF717000JR1P2NGN False WARNING: Retrieving IsDeviceCacheEnabled failed with ErrorCode 1. NVMe INTEL SSDPE2MX01 CVPF7254009B1P2NGN False WARNING: Retrieving IsDeviceCacheEnabled failed with ErrorCode 1.
Any help with this issue would be appreciated.
Thanks.
Hello,
Recently we started replacing our Windows Server 2016 Hyper-V Clusters for Server 2019. On each cluster we have a Hyper-V Replica broker that allows replication from any authenticated server and stores the Replica Files to a default location of one of the Cluster Shared Volumes.
With WS2019 we run into the issue where we get an error applying the Replication Configuration settings. The error is as follows:
Error applying Replication Configuration changes. Unable to open specified location for replication storage. Failed to add authorization entry. Unable to open specified location to store Replica files 'C:\ClusterStorage\volume1\'. Error: 0x80070057 (One
or more arguments are invalid).
When we target the default location to a CSV where the owner node is the same as the owner node for the Broker role we don't get this error. However I don't expect this to work in production (moving roles to other nodes).
Did anyone ran into the same issue, and what might be a solution for this? Did anything changed between WS2016 & WS2019 what might cause this?
Kind regards,
Malcolm
I created a 2-node failover cluster in my Hyper-V environment.
My concern here is that when I RAN:
Format-Volume -DriveLetter D
The D drives on both nodes were formatted.
When I ran Get-Volume on one of the nodes, I noticed that my D & E drives on each node was listed twice.
I noticed that 'Storage Replica' was added as a Cluster Resource Type and that the following device is installed:
Microsoft ClusPort HBA
Which some cursory research says:
"The Software Storage Bus (SSB) is a virtual storage bus spanning all the servers that make up the cluster. SSB essentially makes it possible for each server to see all disks across all servers in the cluster providing full mesh connectivity. SSB consists of two components on each server in the cluster; ClusPort and ClusBlft. ClusPort implements a virtual HBA that allows the node to connect to disk devices in all the other servers in the cluster. ClusBlft implements virtualization of the disk devices and enclosures in each server for ClusPort in other servers to connect to."
Is this by design? Is there a way to disable this? How do we fix this?
Windows Server 2016 Standard, running on Hyper-V
I'm implementing a product called AppLayering by Citrix in a VMware environment. It creates a unique .vhd for each piece of software you install and want to deploy to end users. We created a Scale-Out
File Server for the share so that we could have 100% up time from crashes and updates/reboots. The end user machines mount the .vhds at login; usually anywhere from 5-15 of these .vhds which range from 1Gb to 12GB in size.
Now that I'm increasing the amount of machines accessing this share, sometimes I experience a very long delay, as much as 6 minutes, before the layers are mounted. They usually mount within seconds.
However, it's not consistently worse the more machines that are logged in, rarely it's still instant, but it does seem to get worse in general the more machines are mounting these layers.
The only performance settings I've tried to tinker with is the MaxThreadsPerQueue from 20 to 64. This reg entry was not in the registry by default, I had to make it myself, so I'm not sure
if that means anything. Also not sure if 64 is even a good number to change it to either, just shooting in the dark here, any help would be much appreciated!
Darin
The following guides have been used as a reference for clustering a certificate authority:
They both indicate the use of a shared disk with a letter for storing the database and log. That disk is then added to the generic cluster role that is created for ADCS. Can a cluster shared volume be used instead?
I am running a windows server 2012 r2 hyper-v cluster with via datacore attached jbod storage (via iscsi).
Situation: within the last 6 month this situation occured 3 times. It happened on different Hyper-V hosts. For some reason the host looses connection to the virtual disk and cannot see other hyper-v server of the cluster. The Failover cluster service is stopped on the host and according to event logs the hyper-v host is removed from failover-cluster and later one re-joined.
The virtual disk gets moved to another Hyper-V hosts and all VMs are moved to other Hyper-V hosts and restart.
When I look at the failover-cluster the hyper-v host that lost the connection is already part of the hyper-v cluster again. It is owner of a virtual disk and several vms are running on this host. So the failed hyper-v hosts is fine again.
I am trying to find out what causes this behavior. All VMs on this host go through a crash restart and that isn't something I want to see on my production environment.
There are no AV running on the hyper-v hosts. The storage system/datacore do not show any errors. No backups or snapshots are being done during that time frame. I checked with our network admin if there were any information on switches/routers/firewall during
that time frame but nothing was found.
The errors I see are:
In event log FailoverClustering - Diagnostic I see the first event
[NETFTAPI] Signaled NetftRemoteUnreachable event, local address x.x.x.x:3343 remote address x.x.x.x:3343 I see this event for all networks.
On System Event logs - at the same time:
The cluster Resource Hosting Subsystem (RHS) process was terminated and will be restarted. This is typically associated with cluster health detection and recovery of a resource. Refer to the System event log to determine which resource and resource DLL is causing the issue. (Error ID 1146).
The next error I see is:
Cluster Shared Volume 'XXX' has entered a paused state because of '(c000000e)'. All I/O will temporarily be queued until a path to the volume is reestablished. (Error ID 5120)
Afterwards I see additional errors ID 1146 and 1135 (about removing hyper-v hosts from cluster) and the error:
Cluster Shared Volume XXX has entered a paused state because of '(c000026e)'. All I/O will temporarily be queued until a path to the volume is reestablished. (Error ID 5120).
But this doesn't provide information why RHS was failing or some ideas why the connection to the storage and other hyper-v hosts was lost.
Any ideas what I could do to determine the cause?
Hi ,
I have setup a Hyper-V Cluster with 2 servers , researching a lot it is recommended to have converged network for best practices.
I have 4 x Physical NIC on each servers
- Created a NIC team with all 4 adapters
- Created vSwitch on the top of that and
-Created 4 x vNIC's on that
Now I have 4 vNIC's per Host and I want to configure the networks properly, I currently have below in failover clustering
1 vNIC network to use live migration
1 vNIC network to use for Cluster Comms
1vNIC network to use for Host Comms ( this is being used for idrac too ) and
1vNIC network for VM comms
My questions is what are the recommended failover clustering configurations are ? which networks do I allow cluster network , client access or no cluster commnunication
And do I really need a separate network for VM communication ? Can I use the Host Comms for host communications and vm communication
Thank you in advance
Please remember to mark the replies as answers if they help and unmark them if they provide no help.
I have a two-node Windows Server 2012 failover cluster. The Windows firewall is disabled on both nodes.
When I log on to one of the nodes (bcs-vmhyperv2), and run the Cluster-Aware Updating tool to analyze the cluster readiness, I receive this result:
When I log on to the other node and run the tool, I received the same two errors. The problem computer is always the local computer.
I know that PowerShell remoting and WINRM are enabled! So, the "resolution" steps don't help.
Here's proof:
If I log on to a different Windows Server 2012 system (not one of the cluster nodes), and run the tool, I receive no errors:
In fact, I setup CAU and used it to apply the latest set of Windows Updates from the third computer.
Why can't I use it from the cluster nodes? How do I fix it?
-Tony
Hello,
I have a 3 node Cluster on Windows Server 2016.
For some weeks one of the nodes (node3, a quorum node) was not part of the domain when it should have been. The remaining nodes continued to work.
Node3 was re-added to the domain thinking that would resolve the errors, but alas no.
The cluster event on node3 shows error “Cluster failed to start. The latest copy of cluster configuration data was not available within the set of nodes attempting to start the cluster. Changes to the cluster occurred while the set of nodes were not in membership and as a result were not able to receive data updates.” and
“Cluster node “Node3” failed to join the cluster because it could not communicate over the network with any other node in the cluster. Verify network connectivity and configuration of any network firewalls”
Firewall is off. All nodes can ping and browse to folders on one another.
The Cluster service has been restarted on all nodes.
Running command “cluster node /status” on node3 returns:
Node1 Down
Node2 Down
Node3 Joining
Running the same command on Node1 and Node2 returns the status as being Up for all.
Couple more logs from the Event viewer:
“The Cluster Service service terminated with the following service-specific error: The wait operation timed out.” and “The Cluster Service service terminated unexpectedly. It has done this 53 time(s). The following corrective action will be taken
in 0 milliseconds: Restart the service.”
A generation of the cluster log file gives the same errors as above but with an added:
“Attempt to start the cluster service on all nodes in the cluster so that nodes with the latest copy of the cluster configuration data can first form the cluster. The cluster will be able to start and the nodes will automatically obtain the updated cluster
configuration data. If there are no nodes available with the latest copy of the cluster configuration data, run the 'Start-ClusterNode -FQ' Windows PowerShell cmdlet. Using the ForceQuorum (FQ) parameter will start the cluster service and mark this node's
copy of the cluster configuration data to be authoritative. Forcing quorum on a node with an outdated copy of the cluster database may result in cluster configuration changes that occurred while the node was not participating in the cluster to be lost.”
How can we add node3 back into the cluster successfully?
Hi Folks,
Getting the below error while testing the cluster failover validation .
Description: Validate that all the nodes have the same domain, domain role, and organizational unit.
Fqdn | Domain | Domain Role | Site Name | Organizational Unit |
---|---|---|---|---|
USTYHPV01..COM | .COM | Member Server | Default-First-Site-Name |
Hello,
We implemented a Windows Server 2012 hyper-V cluster recently. All was working correctly. However we decommissioned one of our 2003 Domain controller recently. Since then whenever I try to move core cluster resource to another, the following error is display in the event viewer.
Cluster network name resource 'Cluster Name' cannot be brought online. Attempt to locate a writeable domain controller (in domain\\nameofdecomissionedomaincontroller.domainname.local) in order to create or update a computer object associated with the resource failed for the following reason:
The server is not operational.
The error code was '8250'. Ensure that a writeable domain controller is accessible to this node within the configured domain. Also ensure that the DNS server is running in order to resolve the name of the domain controller.
But the core resource is brought online without any issue and the cluster is working correctly.
I did a search nameofdecomissionedomaincontroller.domainname.local into the registry, the only entry I found is below.
I guess this is where failover clustering is caching this setting and trying to contact the demoted DC every time I try to move a resource. I already tried to restart each cluster node and checked that the DC was decommissioned correctly.
Is it safe to edit the registry with existing DC name? Or any other solution is most welcomed.
Irfan Goolab SALES ENGINEER (Microsoft UC) MCP, MCSA, MCTS, MCITP, MCT
Dear all,
I have a 2 node windows server 2016 standard edition cluster, which is used for SQL cluster. it's all virtual on Hyper-V 2016, and as far as I can tell everything is configured correctly and working fine.
We have an old application that requires the ability to access the C$ on the clustered nodes, which the cluster role for SQL in this case, so if we go to\\node1\C$ or \\node2\C$ it works fine, however if I go to\\Cluster Role Client Access Name\C$ then we get the following error:
Any help is greatly appreciated.
HelpNeed
Hello,
I have a problem about hyper-v failover cluster live migration. Originally i have a Windows Server 2012R2 hyper-v failover cluster with 3 node members (2 x Dell R420 1 X Dell R430). The live migration works perfect between the nodes. Our company buy 2 Dell R440 server and a new storage and we bulid a new Windows Server 2019 hyper-v failover cluster. After migrate VM's to the new cluster we destroyed the old cluster and reinstall the R430 to Windows Server 2019. Next step we add the R430 to the new cluster.
node1 (R440): Intel® Xeon® Silver 4116 Processor
node 2 (R440): Intel® Xeon® Silver 4116 Processor
node3 (R430): Intel® Xeon® Processor E5-2640 v2
If I try live migraton from node 1 or node 2 to node 3 live migration failed:
Event 21502
Live migration of 'VMNAME' failed.
Virtual machine migration operation for 'VMNAME' failed at migration destination 'NODE03'. (Virtual machine ID 63AFF93A-13F7-40B9-8C4A-32B9E6801448)
The virtual machine 'VMNAME' is using processor-specific features not supported on physical computer 'NODE03'. To allow for migration of this virtual machine to physical computers with different processors, modify the virtual machine settings to limit the processor features used by the virtual machine. (Virtual machine ID 63AFF93A-13F7-40B9-8C4A-32B9E6801448)
Processor compatibility is already turned on every VM!!
If I turned off VM I can migrate to node3. After offline migration I turend on VM on node3 (Dell R430) I can move the VM between all nodes but if I restart VM node1 or node2 live migration to node3 fail again.
All nodes is updated with SUU and the OS is up to date.