[peer-metrics] Add metrics for cluster health, like GetClusterHealth admin API
All checks were successful
ci/woodpecker/push/debug Pipeline was successful
ci/woodpecker/pr/debug Pipeline was successful

This commit is contained in:
Alex 2024-02-20 12:37:55 +01:00
parent 00d479358d
commit 3cdf69f079
Signed by: lx
GPG key ID: 0E496D15096376BE
2 changed files with 172 additions and 28 deletions

View file

@ -104,7 +104,7 @@ pub struct System {
#[cfg(feature = "kubernetes-discovery")] #[cfg(feature = "kubernetes-discovery")]
kubernetes_discovery: Option<KubernetesDiscoveryConfig>, kubernetes_discovery: Option<KubernetesDiscoveryConfig>,
_metrics: SystemMetrics, metrics: SystemMetrics,
replication_mode: ReplicationMode, replication_mode: ReplicationMode,
replication_factor: usize, replication_factor: usize,
@ -168,7 +168,7 @@ pub struct ClusterHealth {
pub partitions_all_ok: usize, pub partitions_all_ok: usize,
} }
#[derive(Debug, Clone, Copy)] #[derive(Debug, Clone, Copy, Eq, PartialEq)]
pub enum ClusterHealthStatus { pub enum ClusterHealthStatus {
/// All nodes are available /// All nodes are available
Healthy, Healthy,
@ -376,7 +376,7 @@ impl System {
consul_discovery, consul_discovery,
#[cfg(feature = "kubernetes-discovery")] #[cfg(feature = "kubernetes-discovery")]
kubernetes_discovery: config.kubernetes_discovery.clone(), kubernetes_discovery: config.kubernetes_discovery.clone(),
_metrics: metrics, metrics,
ring, ring,
update_ring: Mutex::new(update_ring), update_ring: Mutex::new(update_ring),
@ -698,7 +698,13 @@ impl System {
while !*stop_signal.borrow() { while !*stop_signal.borrow() {
let restart_at = Instant::now() + STATUS_EXCHANGE_INTERVAL; let restart_at = Instant::now() + STATUS_EXCHANGE_INTERVAL;
// Update local node status that is exchanged.
// Status variables are exported into Prometheus in SystemMetrics,
// so we take the opportunity to also update here the health status
// that is reported in those metrics.
self.update_local_status(); self.update_local_status();
*self.metrics.health.write().unwrap() = Some(self.health());
let local_status: NodeStatus = self.local_status.read().unwrap().clone(); let local_status: NodeStatus = self.local_status.read().unwrap().clone();
let _ = self let _ = self
.rpc .rpc

View file

@ -2,22 +2,40 @@ use std::sync::{Arc, RwLock};
use opentelemetry::{global, metrics::*, KeyValue}; use opentelemetry::{global, metrics::*, KeyValue};
use crate::system::NodeStatus; use crate::system::{ClusterHealth, ClusterHealthStatus, NodeStatus};
/// TableMetrics reference all counter used for metrics /// TableMetrics reference all counter used for metrics
pub struct SystemMetrics { pub struct SystemMetrics {
pub(crate) health: Arc<RwLock<Option<ClusterHealth>>>,
// Static values
pub(crate) _garage_build_info: ValueObserver<u64>, pub(crate) _garage_build_info: ValueObserver<u64>,
pub(crate) _replication_factor: ValueObserver<u64>, pub(crate) _replication_factor: ValueObserver<u64>,
// Disk space values from System::local_status
pub(crate) _disk_avail: ValueObserver<u64>, pub(crate) _disk_avail: ValueObserver<u64>,
pub(crate) _disk_total: ValueObserver<u64>, pub(crate) _disk_total: ValueObserver<u64>,
// Health report from System::health()
pub(crate) _cluster_healthy: ValueObserver<u64>,
pub(crate) _cluster_available: ValueObserver<u64>,
pub(crate) _known_nodes: ValueObserver<u64>,
pub(crate) _connected_nodes: ValueObserver<u64>,
pub(crate) _storage_nodes: ValueObserver<u64>,
pub(crate) _storage_nodes_ok: ValueObserver<u64>,
pub(crate) _partitions: ValueObserver<u64>,
pub(crate) _partitions_quorum: ValueObserver<u64>,
pub(crate) _partitions_all_ok: ValueObserver<u64>,
} }
impl SystemMetrics { impl SystemMetrics {
pub fn new(replication_factor: usize, local_status: Arc<RwLock<NodeStatus>>) -> Self { pub fn new(replication_factor: usize, local_status: Arc<RwLock<NodeStatus>>) -> Self {
let meter = global::meter("garage_system"); let meter = global::meter("garage_system");
let st1 = local_status.clone(); let health = Arc::new(RwLock::new(None));
let st2 = local_status.clone();
Self { Self {
health: health.clone(),
// Static values
_garage_build_info: meter _garage_build_info: meter
.u64_value_observer("garage_build_info", move |observer| { .u64_value_observer("garage_build_info", move |observer| {
observer.observe( observer.observe(
@ -36,30 +54,150 @@ impl SystemMetrics {
}) })
.with_description("Garage replication factor setting") .with_description("Garage replication factor setting")
.init(), .init(),
_disk_avail: meter
.u64_value_observer("garage_local_disk_avail", move |observer| { // Disk space values from System::local_status
let st = st1.read().unwrap(); _disk_avail: {
if let Some((avail, _total)) = st.data_disk_avail { let status = local_status.clone();
observer.observe(avail, &[KeyValue::new("volume", "data")]); meter
} .u64_value_observer("garage_local_disk_avail", move |observer| {
if let Some((avail, _total)) = st.meta_disk_avail { let st = status.read().unwrap();
observer.observe(avail, &[KeyValue::new("volume", "metadata")]); if let Some((avail, _total)) = st.data_disk_avail {
observer.observe(avail, &[KeyValue::new("volume", "data")]);
}
if let Some((avail, _total)) = st.meta_disk_avail {
observer.observe(avail, &[KeyValue::new("volume", "metadata")]);
}
})
.with_description("Garage available disk space on each node")
.init()
},
_disk_total: {
let status = local_status.clone();
meter
.u64_value_observer("garage_local_disk_total", move |observer| {
let st = status.read().unwrap();
if let Some((_avail, total)) = st.data_disk_avail {
observer.observe(total, &[KeyValue::new("volume", "data")]);
}
if let Some((_avail, total)) = st.meta_disk_avail {
observer.observe(total, &[KeyValue::new("volume", "metadata")]);
}
})
.with_description("Garage total disk space on each node")
.init()
},
// Health report from System::health()
_cluster_healthy: {
let health = health.clone();
meter
.u64_value_observer("cluster_healthy", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
if h.status == ClusterHealthStatus::Healthy {
observer.observe(1, &[]);
} else {
observer.observe(0, &[]);
}
}
})
.with_description("Whether all storage nodes are connected")
.init()
},
_cluster_available: {
let health = health.clone();
meter.u64_value_observer("cluster_available", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
if h.status != ClusterHealthStatus::Unavailable {
observer.observe(1, &[]);
} else {
observer.observe(0, &[]);
}
} }
}) })
.with_description("Garage available disk space on each node") .with_description("Whether all requests can be served, even if some storage nodes are disconnected")
.init(), .init()
_disk_total: meter },
.u64_value_observer("garage_local_disk_total", move |observer| { _known_nodes: {
let st = st2.read().unwrap(); let health = health.clone();
if let Some((_avail, total)) = st.data_disk_avail { meter
observer.observe(total, &[KeyValue::new("volume", "data")]); .u64_value_observer("cluster_known_nodes", move |observer| {
} if let Some(h) = health.read().unwrap().as_ref() {
if let Some((_avail, total)) = st.meta_disk_avail { observer.observe(h.known_nodes as u64, &[]);
observer.observe(total, &[KeyValue::new("volume", "metadata")]); }
} })
}) .with_description("Number of nodes already seen once in the cluster")
.with_description("Garage total disk space on each node") .init()
.init(), },
_connected_nodes: {
let health = health.clone();
meter
.u64_value_observer("cluster_connected_nodes", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
observer.observe(h.connected_nodes as u64, &[]);
}
})
.with_description("Number of nodes currently connected")
.init()
},
_storage_nodes: {
let health = health.clone();
meter
.u64_value_observer("cluster_storage_nodes", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
observer.observe(h.storage_nodes as u64, &[]);
}
})
.with_description("Number of storage nodes declared in the current layout")
.init()
},
_storage_nodes_ok: {
let health = health.clone();
meter
.u64_value_observer("cluster_storage_nodes_ok", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
observer.observe(h.storage_nodes_ok as u64, &[]);
}
})
.with_description("Number of storage nodes currently connected")
.init()
},
_partitions: {
let health = health.clone();
meter
.u64_value_observer("cluster_partitions", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
observer.observe(h.partitions as u64, &[]);
}
})
.with_description("Number of partitions in the layout")
.init()
},
_partitions_quorum: {
let health = health.clone();
meter
.u64_value_observer("cluster_partitions_quorum", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
observer.observe(h.partitions_quorum as u64, &[]);
}
})
.with_description(
"Number of partitions for which we have a quorum of connected nodes",
)
.init()
},
_partitions_all_ok: {
let health = health.clone();
meter
.u64_value_observer("cluster_partitions_all_ok", move |observer| {
if let Some(h) = health.read().unwrap().as_ref() {
observer.observe(h.partitions_all_ok as u64, &[]);
}
})
.with_description(
"Number of partitions for which all storage nodes are connected",
)
.init()
},
} }
} }
} }