Quantcast
Channel: High Availability (Clustering) forum
Viewing all 5654 articles
Browse latest View live

S2D in a lab - questions about node failure.....

$
0
0

Set up a 2 node S2D cluster with nested Hyper-V and was doing a few tests. Live and quick migration work fine, but if I 'pull the power' on one of the S2D nodes (to simulate a node failure), the machine on that node never migrates. In Failover Cluster Manager the role shows as 'Unmonitored' and the VM is dead in the water. I do have a file share witness on a machine not impacted by my testing.

I would think that if I pulled the power on a host those VMs would 'figure out' the node is offline and for the other node to pick up the load.

Did I miss a configuration step somewhere?

[EDIT] After a few minutes the machine came back online, but like it had been restarted. Is that expected behavior? Was hoping that would be faster or in a state where the machine wasn't reset. But I may need to set my expectations!


SOFS and connections

$
0
0

Hey

I would like to create an active/active file server for User Profile Disks (vhdx)

I have create a SOFS cluster (Virtual machines + shared disk + CSV).

To my understandig the SOFS setup still uses one (disk owner) server to do the IO to the disk (shared disk on SAN)

If its true - how do I create an "real" active/active cluster? 

(Using Windows Server 2019)

I have approx 1000 concurrent connections....

Mike

Failover Cluster Manager mmc and 4K display issue

$
0
0

Running Win 10 Enterprise v1703 with RSAT installed, when launching the Failover Cluster Manager mmc the right pane is scaled down and superimposed across the middle of the window (see image).  Does anyone else have this problem?  Any solutions? I've found nothing in my searches so far, very frustrating and impossible to use.

4kmmc

thanks

Storage Spaces Direct - No disks with supported bus types found to be used for S2D

$
0
0

Hello,

I am trying to setup a 3 nodes Windows Cluster to take advantage of SQL Always On failover feature. 

I have 3 VM's running on VMWare, inside my company’s datacenter (not Azure), with Windows Server 2016 DataCenter installed on each. I can create the cluster with those 3 nodes, they are not joining any Active Directory (DNS only). I want to use Storage Spaces Direct as shared storage, and this is where I am stuck.

On each 3 nodes, I have 4 disks. From the “Get-PhysicalDisk” PS command result, for all disks, MediaType is SSD and BusType is SAS. I have one disk as the boot volume, one disk to store various files, and 2 disks with an unused partition. These 2 last disks are the ones I want to use with S2D, and they are marked as CanPool=True.
When I run the “Get-PhysicalDisk” PS command from the first node, the disks showing up in the list are : the boot disk and the file disk from node 1, and 6 poolable disks (2 disks from each 3 nodes).

From the S2D validation report (launched from the Failover cluster manager), the 6 poolable disks are marked as "eligible for validation=True" with these characteristics :
Disk partition style is MBR. Disk has an Unused Partition. Disk type is BASIC.

while the others disks (boot volume and file disk) are reporting a warning (I am not sure if it is a problem preventing the enability of S2D...) :
Failed to get SCSI page 83h VPD descriptors for physical disk 0.

and have those characteristics :

Disk 1 : Disk is a boot volume. Disk is a system volume. Disk is used for paging files. Disk partition style is MBR. Disk has an Unused Partition. Disk has an IFS Partition. Cannot cluster a disk with an IFS Partition. Disk type is BASIC. The required inquiry data (SCSI page 83h VPD descriptor) was reported as not being supported.

Disk 2 : 
 Disk partition style is MBR. Disk has an Unused Partition. Disk has an IFS Partition. Cannot cluster a disk with an IFS Partition. Disk type is BASIC. The required inquiry data (SCSI page 83h VPD descriptor) was reported as not being supported.  


When I want to enable S2D from PowerShell command prompt, I receive an error saying : "No disks with supported bus types found to be used for S2D", even if the bus type is SAS.


I am not an expert on managing servers, and I may have overlooked something during the setup. If more information on my setup is needed, I can provide them to the best of knowledge. I wanted to put some screenshots but since my account is not verified yet it was impossible, I have put as many details as I could.

Thank you for any advice provided.


windows 2016 cluster QuarantineThreshold

An error pccurred while creating the cluster. Could not determine Management point Network Type -- The procedure number is out of range

$
0
0
I am receiving the error in the subject when i try to create a cluster. The two nodes of the cluster are WS2016 Core.

The Validation Wizard doesn't show any Error or Warning.
The two servers are set-up to prefer IPv4 over IPv6 as this has been pointed out in some other posts as a possible cause of the type of error that i am  getting.

Thanks in advance for your help, and your suggestions!

Issue with adding a file share witness

$
0
0

Hi,

I've a 2 SQL Servers clustered but no file share witness has been added to them, so I'm trying to add a file share witness now and I got the following error:

Configure Cluster Quorum Settings

Witness Type:
File Share Witness
Witness Resource:
\\nassrv\sql\Censusdbclusterwitness
Cluster Managed Voting:
Enabled

Errors

An error was encountered while modifying the quorum settings.
Your cluster quorum settings have not been changed.

There was an error configuring the file share witness '\\nassrv\sql'.

Unable to save property changes for 'File Share Witness'.

The user name or password is incorrect

Putting into considerations that the cluster hosts computer accounts have a full control permissions over the shared folder, also the Custer Virtual Object.

The shared folder on a separate host that can be accessed by the cluster hosts.

Windows 2012 R2 crashes during backup - Bug Check code 0x00000133

$
0
0
Hi,
We have an HP DL580G7 server, OS Windows Server 2012 R2 Datacenter Core with Hyper-V that is crashing repeatedly due to Bug Check code  0x00000133 . The error most occurs during backup. The server is one of the Hyper-V cluster nodes. Both servers have the same configuration (drivers, firmware, hotfix...), with the only difference that the problematic server is connected to the autoloader.
If you require more information please let me know.
Has anyone else had this issue, if so is there any resolve for it?

Thank

creating cluster log file with C#

$
0
0

I can't find internet help on creating Windows 2008 cluster log via C#.

can some help with samples?

Understand Clustering

$
0
0

Hello Team,

Could you please help me with Clustering articles. I have seen other article of Ask DS , can you please suggest which one to start off so that I could work and troubleshoot the issue ?

Thanks..

add node to 2016 cluster no longer has validation option

$
0
0

using RSAT on a 2016 GUI server to remotely administer a 2016 cluster (still running v8) to add the last 2016 node to it before upgrading to v9.

going through the add node wizard in failover cluster manager no longer seems to have the option of running cluster validation unlike 2012 R2 failover cluster manager - is this by design? i know i can run it after the node is added but would prefer it be done prior to as well.

SOFS and load balancing

$
0
0

Hey

Just create a Cluster running Scale-Out Fileserver.

I need to have an active/active system due to many connection.

When looking into open files - it seems users only connect to one server in the cluster. (I have enable continuous avalibility)

Why?

Mike


Sysadm

2016 to 2008 R2 cluster

$
0
0

Can I add a Windows Server 2016 node to a Windows Server 2008 R2 cluster?


Reza Negarestani

Cannot add cluster disk on windows 2012

$
0
0

I have 2 servers connected to a shared storage array using Fibre Channel.  Currently there are 2 luns on the storage device.  

From the Failover Cluster Manager, when I select “Add Disk” from Storage – Disks, both luns show up and I can add them without problems.  But if I choose to only add one of them first (it doesn't make a difference which one is added first), then it will not allow adding the second one later.  I get a message: "No disks suitable for cluster disks were found. For diagnostic information about disks available to the cluster, use Validate a Configuration Wizard to run Storage tests."

When I do add one (or both disks at the same time), they work just fine for failover clustering.

I can’t imagine this is by design. Is this a known/unknown issue, or is there something special that needs to be done?

Thanks

Data move from ClusterA (iSCSI) to ClusterB (FC)

$
0
0

Have existing Server 2012R2 cluster with iSCSI storage, need to move to new Server 2019 cluster with FC storage.

Do I have it right that one could do any:

1) Downtime (VMs moved via LAN from old cluster storage to new)

2) Add new 2019 hosts to existing cluster & configure at least one 2019 host to have iSCSI & then do LM?

3) Add new 2019 hosts to existing cluster & configure at least one old 2012 R2 host to have FC & then do LM?

How well 2019 adds to 2012R2 cluster?

If 2 or 3 then after data move, remove 2012R2 hosts & upgrade cluster version?

Any other options?

Seb


VMs Failing to Automatically Migrate

$
0
0
I come in every morning to find a hand full of my VMs indicating "Live Migration was canceled." This seems to be happening around 12:00 - 1:00 AM, but I can't find anything configured to tell it to migrate so I'm not sure why it is happening to begin with. The event logs are not helpful... Cluster Event ID is 1155 "The pending move for the role 'server name' did not complete." The Hyper-V-High-Availability log shows Event ID 21150 "'Virtual Machine Cluster WMI' successfully taken the cluster WMI provider offline." which was right before the 21111 Event ID "Live migration of 'VM Instance Name' failed. It is typically the same VMs, but not always. I see the error on both Nodes (2 node cluster, 2 CSVs). Hyper-V-VMMS logs show 1940 "The WMI provider 'VmmsWmiInstanceAndMethodProvider' has shut down." Then 20413 "The Virtual Machine Management service initiated the live migration of virtual machine  'VM Name' to destination host 'Other Node' (VMID)." for each of the VMs running on that node. Some are successful, but a few get 21014 "Virtual machine migration for 'VM Name' was not finished because the operation was canceled. (Virtual machine ID)" and finally 21024 "Virtual machine migration operation for 'VM Name' failed at migration source 'Host Name'. (Virtual machine ID)". I can manually live migrate all VMs back and forth all day. I have plenty of resources on both nodes (RAM & CPU), and I have turned off the Hyper-V cluster balancer to automatically move machines. We used to have SCVMM installed but it was overkill for our small environment so it was decommissioned. While I would like to resolve the failures, I would be happy just knowing what was causing the VMs to migrate in the first place since it isn't necessary for them to do this every night. The cluster is not configured with CAU. Any guidance would be greatly appreciated!!

Can't live migrate multiple machines from FCM, but can from powershell

$
0
0

Hey everyone,

I started experiencing a weird issue this week.  We have a 2 node cluster (server 2016) setup with hyper-v vdi and some pooled desktops.  We cannot live migrate these vdi machines consistently.  At some point during the live migration one fails which causes the rest to stop migrating.  The errors are few and far between, but the main one just says that the live migration failed.  However there isn't an error message attached to that event (in FCM).  There is an error message in event viewer, in the hyper-v vmms log, but the description can't be found.  The event id is 22040 and the error code at the end of the message is 0x800705B4, which from my research refers to a timeout issue.  

There are two weird issues with this problem.  The first is that even though the machines fail to migrate I can migrate them one at a time.  (I tested with draining the roles and it still fails).  If I migrate them one at a time there are no errors, ever, and every machine migrates perfectly fine.  The other issue is I wrote a powershell script to move the VMs in parallel, with a foreach command and all of the machines migrate just fine.  I believe that is due to the script calling one command at a time to migrate each virtual machine, but I am not sure why that would work.  

We are currently in the process of rebuilding our master image to see if something has gone wrong with it, however I don't have much faith in that.  I think the issue lies somewhere in the FCM, but I am also not sure.  

I have already checked the simultaneous migrations setup in hyper-v settings.  We use Kerberos, but have tried CredSSP as well.  Since the machines are live migrating one by one I don't think that is the issue.  The servers are connected with a 10GB direct attached link which is the only network setup for live migration traffic.  We also have a duplicate system in our primary location, identical servers with identical peripherals, and it doesn't have an issue.  The only difference is the pools/master images.  Both servers are connected to an iSCSI nimble SAN, but so is the duplicate system, just a different hardware piece.  Everything is identical from the switches they are connecting to, to the coax that is directly attached, between the two different setups.  

One other note the servers in each location are slightly different from each other though.  One server is running a V2 of the same processor and has 8 cores as one is running V1 with 6 cores.  However the exact same situation in our other site exists and it works fine.  

Thanks for anything that you all can provide.  For now I can utilize the powershell script, but need to figure out what this issue is in case it is a pre-cursor to what's to come.  



ClusterStorage Volume appear as Folder and empty on the host

$
0
0

Hello,

hyper-v 2016 cluster volume folder appear in yellow Folderinstead of cluster icon , in addition this folder empty , and if i try to live migrate any VM has storage on this folder  from another host to this host it will fail .

i think is a permeation issue , the node cant brows all the volume with the right permissions...




Osama Yotman Systems Consultant

Get-Volume returns all volumes within Windows Failover Cluster instead of just local

$
0
0
Hello all.

This is my first entry in the forums, so apologies if I miss something or have this in the wrong place.

I am using the Get-Volume command in PowerShell to return all the volumes located on the server I am running it from.

However, our servers are members of Windows Failover Clusters.

On one of our clusters it does what I would expect. We get a list of all the volumes on this particular node. On the other cluster we get a list of all volumes within the cluster.

Does anyone know of any setting in the windows failover cluster (or anywhere else) that could explain the difference in behavior.



In addition if I try to create a new volume using New-Volume (PowerShell) in the cluster that behaves as expected it works without issue.

If I try to create a New-Volume using New-Volume (PowerShell) in the cluster that shows all volumes I get the below error:



Failover clustering could not be enabled for this storage object.
Activity ID: {<blanked>}
    + CategoryInfo          : NotSpecified: (:) [New-Volume], CimException
    + FullyQualifiedErrorId : StorageWMI 46008,Microsoft.Management.Infrastructure.CimCmdlets.InvokeCimMethodCommand,New-Volume
    + PSComputerName        : <blanked>




Any help on this would be greatly appreciated.

Thank you.


error FailoverClustering-Manager - I don't have a cluster

$
0
0

Hi everybody.

a few days ago known a long series of errors as i said in the title ...
I checked that the associated service is disabled because i do not have a cluster but nothing changes.
I checked that there is no cluster configuration and they do not actually exist.
I still clean chache, old cluster configurations (although this machine has never seen a cluster)

As I have known of the strong slowdowns from a few days i noticed that it is concurrent with this error but I just can not get it back to normal.

The error is as follows:

-System
-Provider
[ Name]Microsoft-Windows-FailoverClustering-Manager
[ Guid]{11B3C6B7-E06F-4191-BBB9-7099FFF55614}
EventID4657
Version0
Level2
Task2
Opcode0
Keywords0x8000000000000000
-TimeCreated
[ SystemTime]2018-12-14T10:29:33.170788700Z
EventRecordID2448336
-Correlation
[ ActivityID]{5D23092D-9393-0005-2381-235D9393D401}
-Execution
[ ProcessID]2792
[ ThreadID]7864
ChannelMicrosoft-Windows-FailoverClustering-Manager/Admin
ComputerENPAPNAS03.Ente_Enpap.local
-Security
[ UserID]S-1-5-18
-EventData
Parameter1Get-ClusterNode
Parameter2Il Servizio cluster non è in esecuzione. Verificare che il servizio sia in esecuzione su tutti i nodi del cluster.

thanks for your supports


Viewing all 5654 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>