Forums

Home / Forums

You need to log in to create posts and topics. Login · Register

Pre Production Questions?

Hello,

As you know I am running a three node cluster each node with a 3TB disk totaling 9TB on the cluster. I also have a standalone 800G SSD as well on each host I was going to use for cache even though each host has 128Gb Ram. I'm just about to put her in production and had some questions. My questions are,

  1. Is a journal recommended?
  2. Is journaling or cacheing preferred?
  3. What happens if a node fails or shuts down, will it effect the other nodes?
  4. What happens if a data disk fails, will it effect the other nodes?
  5. I have three 3TB disks for a 9TB cluster, is it best to only use 3TB out of the 9 or is it safe to use more?
  6. What happens if the MGMT network is disconnected?
  7. What happens if the Backend network is connected?
  8. Can you safely remove a node from a cluster?
  9. What is the process to replace a failed DATA disk on a node?

I appreciate the answers as I'd like to know what I'm gonna have to deal with some day 🙂

  1. Journals are recommended when using HDD. Some users use nvme as journals to SSDs.
  2. Journals speed both read and write, writecache is better for writes. Do test yourself as it depends on your workload.
  3. When a node shuts, io will stall for approx 20 s then resume as the cluster updates itself so clients do not deal with downed disks, then after 10 min a recovery will start to create the lost replicas on some other nodes.
  4. If a disk fails, it is similar to a node failing, you can think of a node as a container of disks.
  5. For replicated pools, it is recommended to use 3 replicas, 2 is not safe. You can however use EC pools, which gives much lower overhead (an 4--2 EC pool will have 1.5 % storage overhead), you get the same redundancy but at the cost of speed.
  6. If Management node disconnects, you will not be able to access the node on this subnet for ui and management purpose but you storage and iSCSI is working. you can also use interface bonding to make it highly available.
  7. If backend network goes down you have a problem and your cluster will not work. you should use interface bonding to make it highly available.
  8. If a disks fails, you will be able to delete it from ui, you can put a new disk and add it as an OSD. The system has no concept of replace, rather adding and deleting
  9. you can remove non-management nodes. management nodes should not be removed.  if you wish to change hardware you can "replace" a management node with a new box, it will still use the same hostname/ips as old node

Perfect thank you!! and Great support!

Last questions,

  1. Under Pool> Add Pool> "size" and "Min size" what are these? The replica counts?
  2. Under CRUSH> Rules> does the default replicate rule place replicas on each separate node by default? If not what do i need to add to achieve this?

Thank You!