Hi everyone, I am back with another Storage Spaces Direct post. This time I would like to talk to you about planning S2D clusters from 2 node to a 16 node cluster. What you need to think about when doing your planning. This will be a multi part series covering 4 nodes, adding additional nodes. All Flash, Hybrid, MRV and the big one a 16 Node setup.
So i have been working on this case for about 4 weeks. Where a client have been having issues with CSV’s becoming unhealthy and Cache drives going into Lost Communication during backup. We have a MS Support case but as the client do not have Premier Support it’s taking a while to get trough 1st line and getting it escalated.
I have been starting over and over on this blog post since nov 2017 wich coinsided with me chaning jobs, wich i again just did in april of 2018 to work soly with S2D, VMM, Azure Stak and other datacenter products. So i thought it was time to get this done.
So a while ago i started on a blog post series i had hoped to be able to update with you on a regular basis. That did not happen as the new job was quite hectic and i was not able to get the funding i needed for the new S2D lab. So i ended up getting the old HP DL380 G6 servers from my then old job.
After the Nordic Infrastructure Conference i was approached by a company about a new exciting new job working on S2D, VMM, SCOM and Azure Stack as main focus. After some interview’s i decided to join CTGlobal which i started with in April. So now i thought it was time to revive this series. Since then my S2D cluster has been running very good. Upgrading it to Insider Build 17083 which is the last Insider Build that is supported on VMM 1801. New blog post cumming as soon as the VMM team releases a build for RS5 Windows Server 2019.
4 weeks ago i started my new job at CTGlobal here in Norway. Where i will be focusing on S2D, VMM, OMS, Azure Stack and other datacenter solutions. As we are deploying S2D with VMM we wanted to build an easy to use but robust way of configuring VMM and deploy the physical hosts for S2D and configure them all the way til we create the cluster and enable S2D. Deploying the host require BMC deep discovery of the host. It detects everything from disks, controller, network cards and so on. In my home lab i have Mellanox CX3 cards and are using them for host mgmt and SMB, so i am using a setSwitch with this. And configured this in the script.
So a client had patched there 2 node S2D cluster this Sunday. When the 2nd node came up it would not join the cluster. So the client started troubleshooting, without realizing that the Quorum file share Witness was not working. Someone had managed to delete the folder where the File Share Witness was on. At first, the first node that was patched and booted was working. But at some point both nodes where booted over and over again.
A few days ago a friend of mine asked me if i had any idea on how to get his SSD’s and HDD’s to be attached to the NVME cache devices. As he had added alot of disks over the last 7 months to his S2D cluster. And the normal behavior is that any new Disk will be auto bound to the Cache
A while ago i gave u the first look into Dataon Must, Dataon’s monitoring system that comes with their S2D servers.
Today i want to give you an insight into a new offering that is coming, Barton Glass. Barton Glass is built by Barton Systems member of the Cronos group and 2016 Microsoft Partner of the year in Belgium.
So i built a 2 node S2D cluster a while ago at home on some old HP G6 nodes i got cheap. But have decided to get rid of that and setup a new 3 node cluster with bit’s i can find on ebay. And reuse disks i have. This will be a multipart blog post as the parts are orderd and they come in. And during the build process. I will provide a step by step guide in building it, installing, configuring, monitoring and troubleshooting Storage Spaces Direct. Including switch config.
After patching both our S2D clusters today, i have had the same error after resuming nodes and failing back roles. This happens after installing KB4038782
Update, after installing CU10 october patch. Rebooting a S2D node will not cause this issue, after the initial update boot.
The Physical disk’s stays in maintenance mode.