Process to rotate Gluster bricks in 'Replicated' volume type cluster #2156
-
Hello, To give some background of our use case: This week we hit some weirdness with files not healing despite the self-heal daemon running and triggering a full heal. The files did not say they were split brain, and despite getting it fixed using split-brain fix techniques, I can't quite pinpoint where our issue lies that it got in that state to begin with. Is there any suggestions on how we should do these rotations based on our use case? Should we add the new brick, make sure everything heals and then remove the old? Do we need to even wait to remove the old brick once the new one is added? Should we stop any processes or anything? I want to make sure our process can be removed from the "root cause" equation so we can keep poking at it and see if it occurs again. While on the topic of healing, could you also clarify the heal output for me? Brick server1:/gfs/test-volume_0
Number of entries: 0
Brick server2:/gfs/test-volume_1
Number of entries: 101
/95.txt
/32.txt
/66.txt With the above example output, does this essentially say that the listed files on server2 do not match whats on the other servers and they need to be updated or perhaps they are missing? Is there any diagram/flowchart of whats happening between the bricks communication-wise in a replicated volume when healing is occuring? Just wanted to understand the nuts and bolts a bit more. Thank you in advance! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Based on more digging, I'm gonna go with Red Hat's methods outlined. I'm assuming this is the best approach and will mark this answered unless someone has a different opinion. |
Beta Was this translation helpful? Give feedback.
Based on more digging, I'm gonna go with Red Hat's methods outlined. I'm assuming this is the best approach and will mark this answered unless someone has a different opinion.
https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.4/html/administration_guide/sect-replacing_hosts#Replacing_a_Host_Machine_with_a_Different_Hostname