SOLVED
Home

WS2016 failover cluster issues

%3CLINGO-SUB%20id%3D%22lingo-sub-59224%22%20slang%3D%22en-US%22%3EWS2016%20failover%20cluster%20issues%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-59224%22%20slang%3D%22en-US%22%3E%3CP%3EHello%20all%2C%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3EAt%20work%20we%20have%20recently%20purchased%204%20new%20and%20shiny%20Dell%20R730's.%20We%20have%20got%202%20of%20the%20new%20servers%20in%20the%20new%20cluster%20and%20in%20operation%20at%20the%20moment.%20We%20have%20got%204%2010gbe%20connections%20to%20the%20switch%20and%20SAN%2C%20on%20the%202nd%20node%20we%20have%20only%20got%20one%20resilient%20link%20to%20switch%2FSAN.%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3EWe%20have%20had%20to%20migrate%20our%20production%20exchange%20server%20to%20it%20over%20the%20weekend%20(even%20though%20the%20new%20cluster%20is%20still%20being%20tested)%20as%20everytime%20we%20tried%20to%20restore%20using%20VEEAM%20it%20caused%20the%20host%20in%20the%20old%20(overcommitted)%20cluster%20to%20BSOD.%20So%20as%20our%20options%20were%20limited%20we%20placed%20it%20on%20the%20new%20cluster.%20Everything%20worked%20fine%20until%20the%20moment%20I%20walked%20through%20the%20door%20this%20morning%2C%20when%20the%20iSCSI%20nic%20on%20the%201st%20node%20ended%20up%20being%20disabled%20(attempts%20to%20reenable%20were%20unsuccessful).%20I%20managed%20to%20sort%20of%20bring%20it%20back%20by%20enabling%20the%202nd%20iSCSI%20link%20(Dell%20Host%20Integration%20Tools%20for%20WS2016%20haven'%20been%20released%20yet%20%3A(%3C%2Fimg%3E%20)%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3EI%20had%20to%20reboot%20the%202nd%20node%20this%20afternoon%20as%20there%20was%20locked%20resources%20on%20it%2C%20as%20the%202nd%20node%20rebooted.%20The%201st%20nodes%20connection%20to%20the%20SAN%20dropped%20again%20(killing%20exchange)%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3EIs%20there%20any%20particular%20reason%20why%20this%20would%20have%20happened%3F%20and%20what%20is%20the%20best%20thing%20to%20do%20regarding%20Exchange.%20I%20don't%20want%20to%20move%20it%20back%20as%20we%20will%20only%20need%20to%20move%20it%20again%20soon%20anyway%3C%2FP%3E%3C%2FLINGO-BODY%3E%3CLINGO-SUB%20id%3D%22lingo-sub-59534%22%20slang%3D%22en-US%22%3ERe%3A%20WS2016%20failover%20cluster%20issues%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-59534%22%20slang%3D%22en-US%22%3EHi%20Erwin%2C%3CBR%20%2F%3E%3CBR%20%2F%3EI%20logged%20a%20support%20ticket%20with%20the%20company%20helping%20us%20install%20it.%20The%20MPIO%20feature%20was%20not%20installed.%20We%20have%20two%20dedicated%20nics%20for%20iSCSI%20and%20two%20for%20data.%3CBR%20%2F%3E%3CBR%20%2F%3ECheers%2C%3CBR%20%2F%3E%3CBR%20%2F%3EAndrew%3C%2FLINGO-BODY%3E%3CLINGO-SUB%20id%3D%22lingo-sub-59510%22%20slang%3D%22en-US%22%3ERe%3A%20WS2016%20failover%20cluster%20issues%3C%2FLINGO-SUB%3E%3CLINGO-BODY%20id%3D%22lingo-body-59510%22%20slang%3D%22en-US%22%3E%3CP%3EHello%20Andrew%2C%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3ELocking%20problems%20could%20be%20due%20to%20missing%20or%20wrong%20configuration%20of%20MPIO%3F%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3E%3CSPAN%20class%3D%22lia-inline-image-display-wrapper%20lia-image-align-inline%22%20style%3D%22width%3A%20400px%3B%22%3E%3CIMG%20src%3D%22https%3A%2F%2Fgxcuf89792.i.lithium.com%2Ft5%2Fimage%2Fserverpage%2Fimage-id%2F12775i3E120B7C438F4421%2Fimage-size%2Flarge%3Fv%3D1.0%26amp%3Bpx%3D999%22%20alt%3D%22MPIO.png%22%20title%3D%22MPIO.png%22%20%2F%3E%3CSPAN%20class%3D%22lia-inline-image-caption%22%20onclick%3D%22event.preventDefault()%3B%22%3EMPIO%20settings%3C%2FSPAN%3E%3C%2FSPAN%3E%3C%2FP%3E%3CP%3EDo%20you%20have%20two%20dedicated%20NICs%20for%20iSCSI%20and%20two%20for%20the%26nbsp%3BHyper-V%20switch%3F%20Are%20you%20using%20Switch%20Embedded%20Teaming%20or%20regular%20NIC%20teaming%20on%20the%20Hyper-V%20switch%3F%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3ERegards%2C%3C%2FP%3E%3CP%3E%26nbsp%3B%3C%2FP%3E%3CP%3EErwin%3C%2FP%3E%3C%2FLINGO-BODY%3E
Andrew Baker
Occasional Contributor

Hello all,

 

At work we have recently purchased 4 new and shiny Dell R730's. We have got 2 of the new servers in the new cluster and in operation at the moment. We have got 4 10gbe connections to the switch and SAN, on the 2nd node we have only got one resilient link to switch/SAN.

 

We have had to migrate our production exchange server to it over the weekend (even though the new cluster is still being tested) as everytime we tried to restore using VEEAM it caused the host in the old (overcommitted) cluster to BSOD. So as our options were limited we placed it on the new cluster. Everything worked fine until the moment I walked through the door this morning, when the iSCSI nic on the 1st node ended up being disabled (attempts to reenable were unsuccessful). I managed to sort of bring it back by enabling the 2nd iSCSI link (Dell Host Integration Tools for WS2016 haven' been released yet :( )

 

I had to reboot the 2nd node this afternoon as there was locked resources on it, as the 2nd node rebooted. The 1st nodes connection to the SAN dropped again (killing exchange)

 

Is there any particular reason why this would have happened? and what is the best thing to do regarding Exchange. I don't want to move it back as we will only need to move it again soon anyway

2 Replies
Solution

Hello Andrew,

 

Locking problems could be due to missing or wrong configuration of MPIO?

 

MPIO.pngMPIO settings

Do you have two dedicated NICs for iSCSI and two for the Hyper-V switch? Are you using Switch Embedded Teaming or regular NIC teaming on the Hyper-V switch?

 

Regards,

 

Erwin

Hi Erwin,

I logged a support ticket with the company helping us install it. The MPIO feature was not installed. We have two dedicated nics for iSCSI and two for data.

Cheers,

Andrew
Related Conversations
Stable version of Edge insider browser
HotCakeX in Discussions on
35 Replies
Tabs and Dark Mode
cjc2112 in Discussions on
35 Replies
Extentions Synchronization
ChirmyRam in Discussions on
3 Replies
flashing a white screen while open new tab
Deleted in Discussions on
14 Replies