@bensmrs @candlerb I am still in constant pain thinking about this. I am wanting to stick to LTS for the sake of stability as I will be using this cluster to teach classes at a University.
I guess if you don’t mind maybe helping me think whether its worth the wait or possibly just doing Ceph with NVMe-OF gateway.
I have a 9 node cluster of which 3 nodes will be my primary storage nodes.
Of those 3 nodes , each will carry 8 Micron 7300 MAX U.2 drives (800GB each).
My original plan was to just put all 24 drives in one of our servers and using NVMe-OverRDMA and providing all 24 drives to the other 8 nodes and just using the lvmcluster driver. Unfortunately I think I lose out on too many features for backup and snapshots, and also I just am too open to a single point of failure for the cluster.
incus LTS is incredibllyyy stable and I love it. I would like to stay with it, but I thought about changing to the stable branch with the addition of the linstor driver being implemented.
If I decide to stay with the LTS version, what would you recommend my best course of action be ?
The network is a mix of 40Gb and 100Gb nodes (all Mellanox CX3 Pro and CX6-DX cards with Mellanox SN2700 switch), and will soon be all 100Gb in about 6 months.
Using Ceph as my storage backend is a bit meh in my opinion because I know that NVMe-OverTCP will be CPU intensive as the CX3-PRO does not do any offloading for that, but I know its capable of “SOME” RDMA offloading.
Linbit just seems so nice and easy to use, and so thats why I thought using the linbit driver with DRDB and NVMe-OverRDMA would be my best option.
There is just something I don’t understand from your previous messages…
If I would use 3 storage nodes with Linstor then I would have the storage capacity of a single node with 2 copies , right?
Would it be right to say that if I go about it like that then I could deploy that capacity to all my nodes (that being the storage capacity of 1 storage node which is 8 of the Micron 7300 MAX drives) via whatever protocol I want (NVMe-OverTCP or NVMe-OverRDMA or whatever)?
If those 8 storage drives are deployed to the rest of my nodes as a LVM block storage device , why would I have issues with LVM extents being aligned correctly?
I would like to have the ability that if one of my storage nodes goes down, then one of the other 2 nodes (out of the 3 storage nodes) picks up for the incus cluster and continues on with operations.
Is this correct to think I can achieve this with linstor?