Garage fails to count to 3? #597
Labels
No Label
AdminAPI
Bug
Check AWS
CI
Correctness
Critical
Documentation
Ideas
Improvement
Low priority
Newcomer
Performance
S3 Compatibility
Testing
Usability
No Milestone
No Assignees
3 Participants
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: Deuxfleurs/garage#597
Loading…
Reference in New Issue
No description provided.
Delete Branch "%!s(<nil>)"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
For some reason, garage cannot reach a quorum despite 2 nodes being available in the region and replicas being available outside the region.
Any suggestions on how to fix this (the failed node is basically fubar -- see #595 -- and I'm trying to reach a good state again, so it will be removed completely).
Presumably, this would be an issue with the fact that
capital
andcantor
are on IPv4 - and thus have no way to communicate with any of the other nodes as they have IPv6 addresses registered.You would have to readd them to the deployment with their IPv6 addresses.
Are you using replication mode 2 or 3 ? Your logs look like you are using replication mode 2, which would explain why a single unavailable node breaks your cluster. If that's the case, try setting it to 2-dangerous to restore write capability to your cluster.
Also, if your plan is to eventually remove the failed node from the cluster, you can remove it now from the layout to rebuild copies of all your data, there is no particular reason to wait.