2021-11-09 11:24:04 +00:00
|
|
|
use std::cmp::Ordering;
|
2022-05-01 07:57:05 +00:00
|
|
|
use std::collections::HashMap;
|
2022-09-21 12:39:59 +00:00
|
|
|
use std::collections::HashSet;
|
|
|
|
|
2022-11-07 20:12:11 +00:00
|
|
|
use bytesize::ByteSize;
|
2022-09-22 17:30:01 +00:00
|
|
|
use itertools::Itertools;
|
2021-11-09 11:24:04 +00:00
|
|
|
|
|
|
|
use serde::{Deserialize, Serialize};
|
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
use garage_util::crdt::{AutoCrdt, Crdt, Lww, LwwMap};
|
2021-11-09 11:24:04 +00:00
|
|
|
use garage_util::data::*;
|
2022-05-24 10:16:39 +00:00
|
|
|
use garage_util::error::*;
|
2022-05-01 07:54:19 +00:00
|
|
|
|
2022-09-21 12:39:59 +00:00
|
|
|
use crate::graph_algo::*;
|
2021-11-09 11:24:04 +00:00
|
|
|
|
|
|
|
use crate::ring::*;
|
|
|
|
|
2022-09-21 12:39:59 +00:00
|
|
|
use std::convert::TryInto;
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
const NB_PARTITIONS: usize = 1usize << PARTITION_BITS;
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// The Message type will be used to collect information on the algorithm.
|
2022-09-21 12:39:59 +00:00
|
|
|
type Message = Vec<String>;
|
|
|
|
|
2021-11-09 11:24:04 +00:00
|
|
|
/// The layout of the cluster, i.e. the list of roles
|
|
|
|
/// which are assigned to each cluster node
|
|
|
|
#[derive(Clone, Debug, Serialize, Deserialize)]
|
|
|
|
pub struct ClusterLayout {
|
|
|
|
pub version: u64,
|
|
|
|
|
|
|
|
pub replication_factor: usize,
|
2022-10-10 15:21:13 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This attribute is only used to retain the previously computed partition size,
|
|
|
|
/// to know to what extent does it change with the layout update.
|
2022-11-07 20:12:11 +00:00
|
|
|
pub partition_size: u64,
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Parameters used to compute the assignation currently given by
|
|
|
|
/// ring_assignation_data
|
2022-10-10 15:21:13 +00:00
|
|
|
pub parameters: LayoutParameters,
|
2022-09-21 12:39:59 +00:00
|
|
|
|
2021-11-09 11:24:04 +00:00
|
|
|
pub roles: LwwMap<Uuid, NodeRoleV>,
|
|
|
|
|
|
|
|
/// node_id_vec: a vector of node IDs with a role assigned
|
|
|
|
/// in the system (this includes gateway nodes).
|
|
|
|
/// The order here is different than the vec stored by `roles`, because:
|
2022-09-21 12:39:59 +00:00
|
|
|
/// 1. non-gateway nodes are first so that they have lower numbers holding
|
2022-10-10 15:21:13 +00:00
|
|
|
/// in u8 (the number of non-gateway nodes is at most 256).
|
2021-11-09 11:24:04 +00:00
|
|
|
/// 2. nodes that don't have a role are excluded (but they need to
|
|
|
|
/// stay in the CRDT as tombstones)
|
|
|
|
pub node_id_vec: Vec<Uuid>,
|
|
|
|
/// the assignation of data partitions to node, the values
|
|
|
|
/// are indices in node_id_vec
|
|
|
|
#[serde(with = "serde_bytes")]
|
|
|
|
pub ring_assignation_data: Vec<CompactNodeType>,
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
/// Parameters to be used in the next partition assignation computation.
|
2022-11-07 18:34:40 +00:00
|
|
|
pub staging_parameters: Lww<LayoutParameters>,
|
2022-10-11 16:29:21 +00:00
|
|
|
/// Role changes which are staged for the next version of the layout
|
2022-11-07 18:34:40 +00:00
|
|
|
pub staging_roles: LwwMap<Uuid, NodeRoleV>,
|
2021-11-09 11:24:04 +00:00
|
|
|
pub staging_hash: Hash,
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This struct is used to set the parameters to be used in the assignation computation
|
|
|
|
/// algorithm. It is stored as a Crdt.
|
2022-10-05 13:29:48 +00:00
|
|
|
#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)]
|
|
|
|
pub struct LayoutParameters {
|
2022-10-10 15:21:13 +00:00
|
|
|
pub zone_redundancy: usize,
|
2022-10-05 13:29:48 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
impl AutoCrdt for LayoutParameters {
|
|
|
|
const WARN_IF_DIFFERENT: bool = true;
|
|
|
|
}
|
|
|
|
|
2021-11-09 11:24:04 +00:00
|
|
|
#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)]
|
|
|
|
pub struct NodeRoleV(pub Option<NodeRole>);
|
|
|
|
|
|
|
|
impl AutoCrdt for NodeRoleV {
|
|
|
|
const WARN_IF_DIFFERENT: bool = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// The user-assigned roles of cluster nodes
|
|
|
|
#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)]
|
|
|
|
pub struct NodeRole {
|
2022-10-11 16:29:21 +00:00
|
|
|
/// Datacenter at which this entry belong. This information is used to
|
|
|
|
/// perform a better geodistribution
|
2021-11-09 11:24:04 +00:00
|
|
|
pub zone: String,
|
2022-10-11 16:29:21 +00:00
|
|
|
/// The capacity of the node
|
2021-11-09 11:24:04 +00:00
|
|
|
/// If this is set to None, the node does not participate in storing data for the system
|
|
|
|
/// and is only active as an API gateway to other nodes
|
2022-11-07 20:12:11 +00:00
|
|
|
pub capacity: Option<u64>,
|
2021-11-09 11:24:04 +00:00
|
|
|
/// A set of tags to recognize the node
|
|
|
|
pub tags: Vec<String>,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl NodeRole {
|
|
|
|
pub fn capacity_string(&self) -> String {
|
|
|
|
match self.capacity {
|
2022-11-07 20:12:11 +00:00
|
|
|
Some(c) => ByteSize::b(c).to_string_as(false),
|
2021-11-09 11:24:04 +00:00
|
|
|
None => "gateway".to_string(),
|
|
|
|
}
|
|
|
|
}
|
2022-09-21 12:39:59 +00:00
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
pub fn tags_string(&self) -> String {
|
2022-11-08 15:15:45 +00:00
|
|
|
self.tags.join(",")
|
2022-10-10 15:21:13 +00:00
|
|
|
}
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// Implementation of the ClusterLayout methods unrelated to the assignation algorithm.
|
2021-11-09 11:24:04 +00:00
|
|
|
impl ClusterLayout {
|
2022-10-05 13:29:48 +00:00
|
|
|
pub fn new(replication_factor: usize) -> Self {
|
2022-11-07 18:34:40 +00:00
|
|
|
// We set the default zone redundancy to be equal to the replication factor,
|
|
|
|
// i.e. as strict as possible.
|
2022-10-10 15:21:13 +00:00
|
|
|
let parameters = LayoutParameters {
|
|
|
|
zone_redundancy: replication_factor,
|
|
|
|
};
|
2022-11-07 18:34:40 +00:00
|
|
|
let staging_parameters = Lww::<LayoutParameters>::new(parameters.clone());
|
2022-10-05 13:29:48 +00:00
|
|
|
|
2021-11-09 11:24:04 +00:00
|
|
|
let empty_lwwmap = LwwMap::new();
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
let mut ret = ClusterLayout {
|
2021-11-09 11:24:04 +00:00
|
|
|
version: 0,
|
|
|
|
replication_factor,
|
2022-10-10 15:21:13 +00:00
|
|
|
partition_size: 0,
|
2021-11-09 11:24:04 +00:00
|
|
|
roles: LwwMap::new(),
|
|
|
|
node_id_vec: Vec::new(),
|
|
|
|
ring_assignation_data: Vec::new(),
|
2022-10-10 15:21:13 +00:00
|
|
|
parameters,
|
2022-11-07 18:34:40 +00:00
|
|
|
staging_parameters,
|
|
|
|
staging_roles: empty_lwwmap,
|
|
|
|
staging_hash: [0u8; 32].into(),
|
|
|
|
};
|
|
|
|
ret.staging_hash = ret.calculate_staging_hash();
|
|
|
|
ret
|
|
|
|
}
|
|
|
|
|
|
|
|
fn calculate_staging_hash(&self) -> Hash {
|
|
|
|
let hashed_tuple = (&self.staging_roles, &self.staging_parameters);
|
|
|
|
blake2sum(&rmp_to_vec_all_named(&hashed_tuple).unwrap()[..])
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pub fn merge(&mut self, other: &ClusterLayout) -> bool {
|
|
|
|
match other.version.cmp(&self.version) {
|
|
|
|
Ordering::Greater => {
|
|
|
|
*self = other.clone();
|
|
|
|
true
|
|
|
|
}
|
|
|
|
Ordering::Equal => {
|
2022-11-07 18:34:40 +00:00
|
|
|
self.staging_parameters.merge(&other.staging_parameters);
|
|
|
|
self.staging_roles.merge(&other.staging_roles);
|
2021-11-09 11:24:04 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
let new_staging_hash = self.calculate_staging_hash();
|
|
|
|
let changed = new_staging_hash != self.staging_hash;
|
2021-11-09 11:24:04 +00:00
|
|
|
|
|
|
|
self.staging_hash = new_staging_hash;
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
changed
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
Ordering::Less => false,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
pub fn apply_staged_changes(mut self, version: Option<u64>) -> Result<(Self, Message), Error> {
|
2022-05-24 10:16:39 +00:00
|
|
|
match version {
|
|
|
|
None => {
|
|
|
|
let error = r#"
|
|
|
|
Please pass the new layout version number to ensure that you are writing the correct version of the cluster layout.
|
|
|
|
To know the correct value of the new layout version, invoke `garage layout show` and review the proposed changes.
|
|
|
|
"#;
|
|
|
|
return Err(Error::Message(error.into()));
|
|
|
|
}
|
|
|
|
Some(v) => {
|
|
|
|
if v != self.version + 1 {
|
|
|
|
return Err(Error::Message("Invalid new layout version".into()));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
self.roles.merge(&self.staging_roles);
|
2022-05-24 10:16:39 +00:00
|
|
|
self.roles.retain(|(_, _, v)| v.0.is_some());
|
2022-11-07 18:34:40 +00:00
|
|
|
self.parameters = self.staging_parameters.get().clone();
|
2022-05-24 10:16:39 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
self.staging_roles.clear();
|
|
|
|
self.staging_hash = self.calculate_staging_hash();
|
2022-05-24 10:16:39 +00:00
|
|
|
|
2022-11-08 13:23:08 +00:00
|
|
|
let msg = self.calculate_partition_assignation()?;
|
|
|
|
|
2022-05-24 10:16:39 +00:00
|
|
|
self.version += 1;
|
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
Ok((self, msg))
|
2022-05-24 10:16:39 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pub fn revert_staged_changes(mut self, version: Option<u64>) -> Result<Self, Error> {
|
|
|
|
match version {
|
|
|
|
None => {
|
|
|
|
let error = r#"
|
|
|
|
Please pass the new layout version number to ensure that you are writing the correct version of the cluster layout.
|
|
|
|
To know the correct value of the new layout version, invoke `garage layout show` and review the proposed changes.
|
|
|
|
"#;
|
|
|
|
return Err(Error::Message(error.into()));
|
|
|
|
}
|
|
|
|
Some(v) => {
|
|
|
|
if v != self.version + 1 {
|
|
|
|
return Err(Error::Message("Invalid new layout version".into()));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
self.staging_roles.clear();
|
|
|
|
self.staging_parameters.update(self.parameters.clone());
|
2022-11-08 13:23:08 +00:00
|
|
|
self.staging_hash = self.calculate_staging_hash();
|
2022-05-24 10:16:39 +00:00
|
|
|
|
|
|
|
self.version += 1;
|
|
|
|
|
|
|
|
Ok(self)
|
|
|
|
}
|
|
|
|
|
2021-11-09 11:24:04 +00:00
|
|
|
/// Returns a list of IDs of nodes that currently have
|
|
|
|
/// a role in the cluster
|
|
|
|
pub fn node_ids(&self) -> &[Uuid] {
|
|
|
|
&self.node_id_vec[..]
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn num_nodes(&self) -> usize {
|
|
|
|
self.node_id_vec.len()
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Returns the role of a node in the layout
|
|
|
|
pub fn node_role(&self, node: &Uuid) -> Option<&NodeRole> {
|
|
|
|
match self.roles.get(node) {
|
|
|
|
Some(NodeRoleV(Some(v))) => Some(v),
|
|
|
|
_ => None,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Returns the uuids of the non_gateway nodes in self.node_id_vec.
|
2022-11-08 15:15:45 +00:00
|
|
|
fn nongateway_nodes(&self) -> Vec<Uuid> {
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut result = Vec::<Uuid>::new();
|
|
|
|
for uuid in self.node_id_vec.iter() {
|
|
|
|
match self.node_role(uuid) {
|
|
|
|
Some(role) if role.capacity != None => result.push(*uuid),
|
|
|
|
_ => (),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
result
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Given a node uuids, this function returns the label of its zone
|
2022-11-08 15:15:45 +00:00
|
|
|
fn get_node_zone(&self, uuid: &Uuid) -> Result<String, Error> {
|
2022-10-10 15:21:13 +00:00
|
|
|
match self.node_role(uuid) {
|
|
|
|
Some(role) => Ok(role.zone.clone()),
|
|
|
|
_ => Err(Error::Message(
|
|
|
|
"The Uuid does not correspond to a node present in the cluster.".into(),
|
|
|
|
)),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Given a node uuids, this function returns its capacity or fails if it does not have any
|
2022-11-07 20:12:11 +00:00
|
|
|
pub fn get_node_capacity(&self, uuid: &Uuid) -> Result<u64, Error> {
|
2022-10-10 15:21:13 +00:00
|
|
|
match self.node_role(uuid) {
|
|
|
|
Some(NodeRole {
|
|
|
|
capacity: Some(cap),
|
|
|
|
zone: _,
|
|
|
|
tags: _,
|
|
|
|
}) => Ok(*cap),
|
|
|
|
_ => Err(Error::Message(
|
|
|
|
"The Uuid does not correspond to a node present in the \
|
|
|
|
cluster or this node does not have a positive capacity."
|
|
|
|
.into(),
|
|
|
|
)),
|
|
|
|
}
|
|
|
|
}
|
2022-09-21 12:39:59 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Returns the number of partitions associated to this node in the ring
|
2022-10-11 15:17:13 +00:00
|
|
|
pub fn get_node_usage(&self, uuid: &Uuid) -> Result<usize, Error> {
|
|
|
|
for (i, id) in self.node_id_vec.iter().enumerate() {
|
|
|
|
if id == uuid {
|
|
|
|
let mut count = 0;
|
|
|
|
for nod in self.ring_assignation_data.iter() {
|
|
|
|
if i as u8 == *nod {
|
|
|
|
count += 1
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return Ok(count);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Err(Error::Message(
|
|
|
|
"The Uuid does not correspond to a node present in the \
|
|
|
|
cluster or this node does not have a positive capacity."
|
|
|
|
.into(),
|
|
|
|
))
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Returns the sum of capacities of non gateway nodes in the cluster
|
2022-11-08 15:15:45 +00:00
|
|
|
fn get_total_capacity(&self) -> Result<u64, Error> {
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut total_capacity = 0;
|
2022-10-11 16:29:21 +00:00
|
|
|
for uuid in self.nongateway_nodes().iter() {
|
2022-10-10 15:21:13 +00:00
|
|
|
total_capacity += self.get_node_capacity(uuid)?;
|
|
|
|
}
|
|
|
|
Ok(total_capacity)
|
|
|
|
}
|
2022-09-21 12:39:59 +00:00
|
|
|
|
2021-11-09 11:24:04 +00:00
|
|
|
/// Check a cluster layout for internal consistency
|
2022-10-11 16:29:21 +00:00
|
|
|
/// (assignation, roles, parameters, partition size)
|
2021-11-09 11:24:04 +00:00
|
|
|
/// returns true if consistent, false if error
|
2022-11-08 13:23:08 +00:00
|
|
|
pub fn check(&self) -> Result<(), String> {
|
2021-11-09 11:24:04 +00:00
|
|
|
// Check that the hash of the staging data is correct
|
2022-11-07 18:34:40 +00:00
|
|
|
let staging_hash = self.calculate_staging_hash();
|
2021-11-09 11:24:04 +00:00
|
|
|
if staging_hash != self.staging_hash {
|
2022-11-08 13:23:08 +00:00
|
|
|
return Err("staging_hash is incorrect".into());
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Check that node_id_vec contains the correct list of nodes
|
|
|
|
let mut expected_nodes = self
|
|
|
|
.roles
|
|
|
|
.items()
|
|
|
|
.iter()
|
|
|
|
.filter(|(_, _, v)| v.0.is_some())
|
|
|
|
.map(|(id, _, _)| *id)
|
|
|
|
.collect::<Vec<_>>();
|
|
|
|
expected_nodes.sort();
|
|
|
|
let mut node_id_vec = self.node_id_vec.clone();
|
|
|
|
node_id_vec.sort();
|
|
|
|
if expected_nodes != node_id_vec {
|
2022-11-08 13:23:08 +00:00
|
|
|
return Err(format!("node_id_vec does not contain the correct set of nodes\nnode_id_vec: {:?}\nexpected: {:?}", node_id_vec, expected_nodes));
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Check that the assignation data has the correct length
|
2022-11-08 13:23:08 +00:00
|
|
|
let expected_assignation_data_len = (1 << PARTITION_BITS) * self.replication_factor;
|
|
|
|
if self.ring_assignation_data.len() != expected_assignation_data_len {
|
|
|
|
return Err(format!(
|
|
|
|
"ring_assignation_data has incorrect length {} instead of {}",
|
|
|
|
self.ring_assignation_data.len(),
|
|
|
|
expected_assignation_data_len
|
|
|
|
));
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Check that the assigned nodes are correct identifiers
|
|
|
|
// of nodes that are assigned a role
|
|
|
|
// and that role is not the role of a gateway nodes
|
|
|
|
for x in self.ring_assignation_data.iter() {
|
|
|
|
if *x as usize >= self.node_id_vec.len() {
|
2022-11-08 13:23:08 +00:00
|
|
|
return Err(format!(
|
|
|
|
"ring_assignation_data contains invalid node id {}",
|
|
|
|
*x
|
|
|
|
));
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
let node = self.node_id_vec[*x as usize];
|
|
|
|
match self.roles.get(&node) {
|
|
|
|
Some(NodeRoleV(Some(x))) if x.capacity.is_some() => (),
|
2022-11-08 13:23:08 +00:00
|
|
|
_ => return Err("ring_assignation_data contains id of a gateway node".into()),
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// Check that every partition is associated to distinct nodes
|
2022-10-10 15:21:13 +00:00
|
|
|
let rf = self.replication_factor;
|
|
|
|
for p in 0..(1 << PARTITION_BITS) {
|
2022-12-11 17:30:02 +00:00
|
|
|
let nodes_of_p = self.ring_assignation_data[rf * p..rf * (p + 1)].to_vec();
|
2022-10-10 15:21:13 +00:00
|
|
|
if nodes_of_p.iter().unique().count() != rf {
|
2022-11-08 13:23:08 +00:00
|
|
|
return Err(format!("partition does not contain {} unique node ids", rf));
|
2022-10-10 15:21:13 +00:00
|
|
|
}
|
2022-11-07 18:34:40 +00:00
|
|
|
// Check that every partition is spread over at least zone_redundancy zones.
|
2022-12-11 17:30:02 +00:00
|
|
|
let zones_of_p = nodes_of_p
|
2022-11-07 19:29:25 +00:00
|
|
|
.iter()
|
|
|
|
.map(|n| {
|
|
|
|
self.get_node_zone(&self.node_id_vec[*n as usize])
|
|
|
|
.expect("Zone not found.")
|
|
|
|
})
|
|
|
|
.collect::<Vec<_>>();
|
2022-10-10 15:21:13 +00:00
|
|
|
let redundancy = self.parameters.zone_redundancy;
|
2022-11-07 19:29:25 +00:00
|
|
|
if zones_of_p.iter().unique().count() < redundancy {
|
2022-11-08 13:23:08 +00:00
|
|
|
return Err(format!(
|
|
|
|
"nodes of partition are in less than {} distinct zones",
|
|
|
|
redundancy
|
|
|
|
));
|
2022-10-10 15:21:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// Check that the nodes capacities is consistent with the stored partitions
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut node_usage = vec![0; MAX_NODE_NUMBER];
|
|
|
|
for n in self.ring_assignation_data.iter() {
|
|
|
|
node_usage[*n as usize] += 1;
|
|
|
|
}
|
|
|
|
for (n, usage) in node_usage.iter().enumerate() {
|
|
|
|
if *usage > 0 {
|
|
|
|
let uuid = self.node_id_vec[n];
|
2022-11-08 13:23:08 +00:00
|
|
|
let partusage = usage * self.partition_size;
|
|
|
|
let nodecap = self.get_node_capacity(&uuid).unwrap();
|
|
|
|
if partusage > nodecap {
|
|
|
|
return Err(format!(
|
|
|
|
"node usage ({}) is bigger than node capacity ({})",
|
|
|
|
usage * self.partition_size,
|
|
|
|
nodecap
|
|
|
|
));
|
2022-10-10 15:21:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2022-09-22 17:30:01 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// Check that the partition size stored is the one computed by the asignation
|
|
|
|
// algorithm.
|
2022-10-10 15:21:13 +00:00
|
|
|
let cl2 = self.clone();
|
2022-11-07 19:29:25 +00:00
|
|
|
let (_, zone_to_id) = cl2.generate_nongateway_zone_ids().unwrap();
|
2022-10-11 15:17:13 +00:00
|
|
|
match cl2.compute_optimal_partition_size(&zone_to_id) {
|
2022-11-08 13:23:08 +00:00
|
|
|
Ok(s) if s != self.partition_size => {
|
|
|
|
return Err(format!(
|
|
|
|
"partition_size ({}) is different than optimal value ({})",
|
|
|
|
self.partition_size, s
|
|
|
|
))
|
|
|
|
}
|
|
|
|
Err(e) => return Err(format!("could not calculate optimal partition size: {}", e)),
|
2022-10-11 15:17:13 +00:00
|
|
|
_ => (),
|
2022-10-10 15:21:13 +00:00
|
|
|
}
|
2022-09-22 17:30:01 +00:00
|
|
|
|
2022-11-08 13:23:08 +00:00
|
|
|
Ok(())
|
2021-11-09 11:24:04 +00:00
|
|
|
}
|
2022-09-21 12:39:59 +00:00
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// Implementation of the ClusterLayout methods related to the assignation algorithm.
|
2022-09-21 12:39:59 +00:00
|
|
|
impl ClusterLayout {
|
2022-05-01 07:57:05 +00:00
|
|
|
/// This function calculates a new partition-to-node assignation.
|
2022-09-21 12:39:59 +00:00
|
|
|
/// The computed assignation respects the node replication factor
|
2022-10-10 15:21:13 +00:00
|
|
|
/// and the zone redundancy parameter It maximizes the capacity of a
|
2022-05-01 07:57:05 +00:00
|
|
|
/// partition (assuming all partitions have the same size).
|
|
|
|
/// Among such optimal assignation, it minimizes the distance to
|
|
|
|
/// the former assignation (if any) to minimize the amount of
|
2022-10-06 10:54:51 +00:00
|
|
|
/// data to be moved.
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Staged role changes must be merged with nodes roles before calling this function,
|
|
|
|
/// hence it must only be called from apply_staged_changes() and hence is not public.
|
2022-10-11 15:17:13 +00:00
|
|
|
fn calculate_partition_assignation(&mut self) -> Result<Message, Error> {
|
2022-11-07 18:34:40 +00:00
|
|
|
// We update the node ids, since the node role list might have changed with the
|
|
|
|
// changes in the layout. We retrieve the old_assignation reframed with new ids
|
2022-10-10 15:21:13 +00:00
|
|
|
let old_assignation_opt = self.update_node_id_vec()?;
|
|
|
|
|
|
|
|
let mut msg = Message::new();
|
2022-10-11 15:17:13 +00:00
|
|
|
msg.push("==== COMPUTATION OF A NEW PARTITION ASSIGNATION ====".into());
|
|
|
|
msg.push("".into());
|
2022-10-10 15:21:13 +00:00
|
|
|
msg.push(format!(
|
2022-10-11 15:17:13 +00:00
|
|
|
"Partitions are \
|
2022-10-10 15:21:13 +00:00
|
|
|
replicated {} times on at least {} distinct zones.",
|
2022-10-11 15:17:13 +00:00
|
|
|
self.replication_factor, self.parameters.zone_redundancy
|
2022-10-10 15:21:13 +00:00
|
|
|
));
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We generate for once numerical ids for the zones of non gateway nodes,
|
|
|
|
// to use them as indices in the flow graphs.
|
2022-10-11 16:29:21 +00:00
|
|
|
let (id_to_zone, zone_to_id) = self.generate_nongateway_zone_ids()?;
|
2022-10-10 15:21:13 +00:00
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
let nb_nongateway_nodes = self.nongateway_nodes().len();
|
|
|
|
if nb_nongateway_nodes < self.replication_factor {
|
2022-10-10 15:21:13 +00:00
|
|
|
return Err(Error::Message(format!(
|
|
|
|
"The number of nodes with positive \
|
2022-10-06 10:54:51 +00:00
|
|
|
capacity ({}) is smaller than the replication factor ({}).",
|
2022-10-11 16:29:21 +00:00
|
|
|
nb_nongateway_nodes, self.replication_factor
|
2022-10-10 15:21:13 +00:00
|
|
|
)));
|
|
|
|
}
|
2022-10-11 15:17:13 +00:00
|
|
|
if id_to_zone.len() < self.parameters.zone_redundancy {
|
2022-10-10 15:21:13 +00:00
|
|
|
return Err(Error::Message(format!(
|
|
|
|
"The number of zones with non-gateway \
|
2022-10-06 10:54:51 +00:00
|
|
|
nodes ({}) is smaller than the redundancy parameter ({})",
|
2022-10-10 15:21:13 +00:00
|
|
|
id_to_zone.len(),
|
2022-10-11 15:17:13 +00:00
|
|
|
self.parameters.zone_redundancy
|
2022-10-10 15:21:13 +00:00
|
|
|
)));
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We compute the optimal partition size
|
|
|
|
// Capacities should be given in a unit so that partition size is at least 100.
|
|
|
|
// In this case, integer rounding plays a marginal role in the percentages of
|
|
|
|
// optimality.
|
2022-10-10 15:21:13 +00:00
|
|
|
let partition_size = self.compute_optimal_partition_size(&zone_to_id)?;
|
|
|
|
|
|
|
|
if old_assignation_opt != None {
|
|
|
|
msg.push(format!(
|
2022-10-11 15:17:13 +00:00
|
|
|
"Optimal size of a partition: {} (was {} in the previous layout).",
|
2022-11-07 20:12:11 +00:00
|
|
|
ByteSize::b(partition_size).to_string_as(false),
|
|
|
|
ByteSize::b(self.partition_size).to_string_as(false)
|
2022-10-10 15:21:13 +00:00
|
|
|
));
|
|
|
|
} else {
|
|
|
|
msg.push(format!(
|
|
|
|
"Given the replication and redundancy constraints, the \
|
|
|
|
optimal size of a partition is {}.",
|
2022-11-07 20:12:11 +00:00
|
|
|
ByteSize::b(partition_size).to_string_as(false)
|
2022-10-10 15:21:13 +00:00
|
|
|
));
|
|
|
|
}
|
2022-11-07 18:34:40 +00:00
|
|
|
// We write the partition size.
|
2022-10-10 15:21:13 +00:00
|
|
|
self.partition_size = partition_size;
|
|
|
|
|
|
|
|
if partition_size < 100 {
|
|
|
|
msg.push(
|
2022-11-08 15:15:45 +00:00
|
|
|
"WARNING: The partition size is low (< 100), make sure the capacities of your nodes are correct and are of at least a few MB"
|
2022-10-10 15:21:13 +00:00
|
|
|
.into(),
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We compute a first flow/assignation that is heuristically close to the previous
|
|
|
|
// assignation
|
2022-10-11 16:29:21 +00:00
|
|
|
let mut gflow = self.compute_candidate_assignation(&zone_to_id, &old_assignation_opt)?;
|
2022-10-10 15:21:13 +00:00
|
|
|
if let Some(assoc) = &old_assignation_opt {
|
2022-11-07 18:34:40 +00:00
|
|
|
// We minimize the distance to the previous assignation.
|
2022-10-10 15:21:13 +00:00
|
|
|
self.minimize_rebalance_load(&mut gflow, &zone_to_id, assoc)?;
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We display statistics of the computation
|
2022-11-07 19:29:25 +00:00
|
|
|
msg.extend(self.output_stat(&gflow, &old_assignation_opt, &zone_to_id, &id_to_zone)?);
|
2022-10-10 15:21:13 +00:00
|
|
|
msg.push("".to_string());
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We update the layout structure
|
2022-10-10 15:21:13 +00:00
|
|
|
self.update_ring_from_flow(id_to_zone.len(), &gflow)?;
|
2022-10-11 15:17:13 +00:00
|
|
|
|
2022-11-08 13:23:08 +00:00
|
|
|
if let Err(e) = self.check() {
|
2022-10-11 15:17:13 +00:00
|
|
|
return Err(Error::Message(
|
2022-11-08 13:23:08 +00:00
|
|
|
format!("Layout check returned an error: {}\nOriginal result of computation: <<<<\n{}\n>>>>", e, msg.join("\n"))
|
2022-10-11 15:17:13 +00:00
|
|
|
));
|
|
|
|
}
|
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
Ok(msg)
|
|
|
|
}
|
2022-05-01 07:57:05 +00:00
|
|
|
|
|
|
|
/// The LwwMap of node roles might have changed. This function updates the node_id_vec
|
|
|
|
/// and returns the assignation given by ring, with the new indices of the nodes, and
|
2022-09-21 12:39:59 +00:00
|
|
|
/// None if the node is not present anymore.
|
2022-05-01 07:57:05 +00:00
|
|
|
/// We work with the assumption that only this function and calculate_new_assignation
|
|
|
|
/// do modify assignation_ring and node_id_vec.
|
2022-10-10 15:21:13 +00:00
|
|
|
fn update_node_id_vec(&mut self) -> Result<Option<Vec<Vec<usize>>>, Error> {
|
|
|
|
// (1) We compute the new node list
|
2022-11-07 18:34:40 +00:00
|
|
|
// Non gateway nodes should be coded on 8bits, hence they must be first in the list
|
|
|
|
// We build the new node ids
|
2022-11-08 15:15:45 +00:00
|
|
|
let new_non_gateway_nodes: Vec<Uuid> = self
|
2022-10-10 15:21:13 +00:00
|
|
|
.roles
|
|
|
|
.items()
|
|
|
|
.iter()
|
|
|
|
.filter(|(_, _, v)| matches!(&v.0, Some(r) if r.capacity != None))
|
|
|
|
.map(|(k, _, _)| *k)
|
|
|
|
.collect();
|
|
|
|
|
|
|
|
if new_non_gateway_nodes.len() > MAX_NODE_NUMBER {
|
|
|
|
return Err(Error::Message(format!(
|
|
|
|
"There are more than {} non-gateway nodes in the new \
|
|
|
|
layout. This is not allowed.",
|
|
|
|
MAX_NODE_NUMBER
|
|
|
|
)));
|
|
|
|
}
|
|
|
|
|
2022-11-08 15:15:45 +00:00
|
|
|
let new_gateway_nodes: Vec<Uuid> = self
|
2022-10-10 15:21:13 +00:00
|
|
|
.roles
|
|
|
|
.items()
|
|
|
|
.iter()
|
|
|
|
.filter(|(_, _, v)| matches!(v, NodeRoleV(Some(r)) if r.capacity == None))
|
|
|
|
.map(|(k, _, _)| *k)
|
|
|
|
.collect();
|
|
|
|
|
|
|
|
let mut new_node_id_vec = Vec::<Uuid>::new();
|
2022-11-08 15:15:45 +00:00
|
|
|
new_node_id_vec.extend(new_non_gateway_nodes);
|
|
|
|
new_node_id_vec.extend(new_gateway_nodes);
|
2022-10-10 15:21:13 +00:00
|
|
|
|
|
|
|
let old_node_id_vec = self.node_id_vec.clone();
|
|
|
|
self.node_id_vec = new_node_id_vec.clone();
|
|
|
|
|
|
|
|
// (2) We retrieve the old association
|
2022-11-07 18:34:40 +00:00
|
|
|
// We rewrite the old association with the new indices. We only consider partition
|
|
|
|
// to node assignations where the node is still in use.
|
2022-10-10 15:21:13 +00:00
|
|
|
if self.ring_assignation_data.is_empty() {
|
2022-11-07 18:34:40 +00:00
|
|
|
// This is a new association
|
2022-10-10 15:21:13 +00:00
|
|
|
return Ok(None);
|
|
|
|
}
|
2022-11-08 15:15:45 +00:00
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
if self.ring_assignation_data.len() != NB_PARTITIONS * self.replication_factor {
|
2022-10-10 15:21:13 +00:00
|
|
|
return Err(Error::Message(
|
|
|
|
"The old assignation does not have a size corresponding to \
|
|
|
|
the old replication factor or the number of partitions."
|
|
|
|
.into(),
|
|
|
|
));
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We build a translation table between the uuid and new ids
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut uuid_to_new_id = HashMap::<Uuid, usize>::new();
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We add the indices of only the new non-gateway nodes that can be used in the
|
|
|
|
// association ring
|
2022-10-10 15:21:13 +00:00
|
|
|
for (i, uuid) in new_node_id_vec.iter().enumerate() {
|
|
|
|
uuid_to_new_id.insert(*uuid, i);
|
|
|
|
}
|
|
|
|
|
2022-11-08 15:15:45 +00:00
|
|
|
let mut old_assignation = vec![Vec::<usize>::new(); NB_PARTITIONS];
|
2022-10-10 15:21:13 +00:00
|
|
|
let rf = self.replication_factor;
|
2022-11-08 15:15:45 +00:00
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
for (p, old_assign_p) in old_assignation.iter_mut().enumerate() {
|
|
|
|
for old_id in &self.ring_assignation_data[p * rf..(p + 1) * rf] {
|
|
|
|
let uuid = old_node_id_vec[*old_id as usize];
|
|
|
|
if uuid_to_new_id.contains_key(&uuid) {
|
|
|
|
old_assign_p.push(uuid_to_new_id[&uuid]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We write the ring
|
2022-10-10 15:21:13 +00:00
|
|
|
self.ring_assignation_data = Vec::<CompactNodeType>::new();
|
|
|
|
|
|
|
|
Ok(Some(old_assignation))
|
2022-05-01 07:57:05 +00:00
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This function generates ids for the zone of the nodes appearing in
|
|
|
|
/// self.node_id_vec.
|
2022-10-11 16:29:21 +00:00
|
|
|
fn generate_nongateway_zone_ids(&self) -> Result<(Vec<String>, HashMap<String, usize>), Error> {
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut id_to_zone = Vec::<String>::new();
|
|
|
|
let mut zone_to_id = HashMap::<String, usize>::new();
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
for uuid in self.nongateway_nodes().iter() {
|
2022-11-08 15:15:45 +00:00
|
|
|
let r = self.node_role(uuid).unwrap();
|
|
|
|
if !zone_to_id.contains_key(&r.zone) && r.capacity != None {
|
|
|
|
zone_to_id.insert(r.zone.clone(), id_to_zone.len());
|
|
|
|
id_to_zone.push(r.zone.clone());
|
2022-10-10 15:21:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
Ok((id_to_zone, zone_to_id))
|
|
|
|
}
|
2022-05-01 07:57:05 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This function computes by dichotomy the largest realizable partition size, given
|
|
|
|
/// the layout roles and parameters.
|
2022-10-10 15:21:13 +00:00
|
|
|
fn compute_optimal_partition_size(
|
|
|
|
&self,
|
|
|
|
zone_to_id: &HashMap<String, usize>,
|
2022-11-07 20:12:11 +00:00
|
|
|
) -> Result<u64, Error> {
|
2022-10-10 15:21:13 +00:00
|
|
|
let empty_set = HashSet::<(usize, usize)>::new();
|
|
|
|
let mut g = self.generate_flow_graph(1, zone_to_id, &empty_set)?;
|
|
|
|
g.compute_maximal_flow()?;
|
2022-11-08 15:15:45 +00:00
|
|
|
if g.get_flow_value()? < (NB_PARTITIONS * self.replication_factor) as i64 {
|
2022-10-10 15:21:13 +00:00
|
|
|
return Err(Error::Message(
|
|
|
|
"The storage capacity of he cluster is to small. It is \
|
|
|
|
impossible to store partitions of size 1."
|
|
|
|
.into(),
|
|
|
|
));
|
|
|
|
}
|
|
|
|
|
|
|
|
let mut s_down = 1;
|
|
|
|
let mut s_up = self.get_total_capacity()?;
|
|
|
|
while s_down + 1 < s_up {
|
|
|
|
g = self.generate_flow_graph((s_down + s_up) / 2, zone_to_id, &empty_set)?;
|
|
|
|
g.compute_maximal_flow()?;
|
2022-11-08 15:15:45 +00:00
|
|
|
if g.get_flow_value()? < (NB_PARTITIONS * self.replication_factor) as i64 {
|
2022-10-10 15:21:13 +00:00
|
|
|
s_up = (s_down + s_up) / 2;
|
|
|
|
} else {
|
|
|
|
s_down = (s_down + s_up) / 2;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Ok(s_down)
|
|
|
|
}
|
|
|
|
|
|
|
|
fn generate_graph_vertices(nb_zones: usize, nb_nodes: usize) -> Vec<Vertex> {
|
|
|
|
let mut vertices = vec![Vertex::Source, Vertex::Sink];
|
|
|
|
for p in 0..NB_PARTITIONS {
|
|
|
|
vertices.push(Vertex::Pup(p));
|
|
|
|
vertices.push(Vertex::Pdown(p));
|
|
|
|
for z in 0..nb_zones {
|
|
|
|
vertices.push(Vertex::PZ(p, z));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for n in 0..nb_nodes {
|
|
|
|
vertices.push(Vertex::N(n));
|
|
|
|
}
|
|
|
|
vertices
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// Generates the graph to compute the maximal flow corresponding to the optimal
|
|
|
|
/// partition assignation.
|
|
|
|
/// exclude_assoc is the set of (partition, node) association that we are forbidden
|
|
|
|
/// to use (hence we do not add the corresponding edge to the graph). This parameter
|
|
|
|
/// is used to compute a first flow that uses only edges appearing in the previous
|
|
|
|
/// assignation. This produces a solution that heuristically should be close to the
|
|
|
|
/// previous one.
|
2022-10-10 15:21:13 +00:00
|
|
|
fn generate_flow_graph(
|
|
|
|
&self,
|
2022-11-07 20:12:11 +00:00
|
|
|
partition_size: u64,
|
2022-10-10 15:21:13 +00:00
|
|
|
zone_to_id: &HashMap<String, usize>,
|
|
|
|
exclude_assoc: &HashSet<(usize, usize)>,
|
|
|
|
) -> Result<Graph<FlowEdge>, Error> {
|
|
|
|
let vertices =
|
2022-10-11 16:29:21 +00:00
|
|
|
ClusterLayout::generate_graph_vertices(zone_to_id.len(), self.nongateway_nodes().len());
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut g = Graph::<FlowEdge>::new(&vertices);
|
|
|
|
let nb_zones = zone_to_id.len();
|
2022-10-11 15:17:13 +00:00
|
|
|
let redundancy = self.parameters.zone_redundancy;
|
2022-10-10 15:21:13 +00:00
|
|
|
for p in 0..NB_PARTITIONS {
|
2022-11-07 20:12:11 +00:00
|
|
|
g.add_edge(Vertex::Source, Vertex::Pup(p), redundancy as u64)?;
|
2022-10-10 15:21:13 +00:00
|
|
|
g.add_edge(
|
|
|
|
Vertex::Source,
|
|
|
|
Vertex::Pdown(p),
|
2022-11-07 20:12:11 +00:00
|
|
|
(self.replication_factor - redundancy) as u64,
|
2022-10-10 15:21:13 +00:00
|
|
|
)?;
|
|
|
|
for z in 0..nb_zones {
|
|
|
|
g.add_edge(Vertex::Pup(p), Vertex::PZ(p, z), 1)?;
|
|
|
|
g.add_edge(
|
|
|
|
Vertex::Pdown(p),
|
|
|
|
Vertex::PZ(p, z),
|
2022-11-07 20:12:11 +00:00
|
|
|
self.replication_factor as u64,
|
2022-10-10 15:21:13 +00:00
|
|
|
)?;
|
|
|
|
}
|
|
|
|
}
|
2022-10-11 16:29:21 +00:00
|
|
|
for n in 0..self.nongateway_nodes().len() {
|
2022-10-10 15:21:13 +00:00
|
|
|
let node_capacity = self.get_node_capacity(&self.node_id_vec[n])?;
|
|
|
|
let node_zone = zone_to_id[&self.get_node_zone(&self.node_id_vec[n])?];
|
2022-10-11 16:29:21 +00:00
|
|
|
g.add_edge(Vertex::N(n), Vertex::Sink, node_capacity / partition_size)?;
|
2022-10-10 15:21:13 +00:00
|
|
|
for p in 0..NB_PARTITIONS {
|
|
|
|
if !exclude_assoc.contains(&(p, n)) {
|
|
|
|
g.add_edge(Vertex::PZ(p, node_zone), Vertex::N(n), 1)?;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Ok(g)
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This function computes a first optimal assignation (in the form of a flow graph).
|
2022-10-11 16:29:21 +00:00
|
|
|
fn compute_candidate_assignation(
|
2022-10-10 15:21:13 +00:00
|
|
|
&self,
|
|
|
|
zone_to_id: &HashMap<String, usize>,
|
2022-10-11 16:29:21 +00:00
|
|
|
prev_assign_opt: &Option<Vec<Vec<usize>>>,
|
2022-10-10 15:21:13 +00:00
|
|
|
) -> Result<Graph<FlowEdge>, Error> {
|
2022-11-07 18:34:40 +00:00
|
|
|
// We list the (partition,node) associations that are not used in the
|
|
|
|
// previous assignation
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut exclude_edge = HashSet::<(usize, usize)>::new();
|
2022-10-11 16:29:21 +00:00
|
|
|
if let Some(prev_assign) = prev_assign_opt {
|
|
|
|
let nb_nodes = self.nongateway_nodes().len();
|
|
|
|
for (p, prev_assign_p) in prev_assign.iter().enumerate() {
|
2022-10-10 15:21:13 +00:00
|
|
|
for n in 0..nb_nodes {
|
|
|
|
exclude_edge.insert((p, n));
|
|
|
|
}
|
2022-10-11 16:29:21 +00:00
|
|
|
for n in prev_assign_p.iter() {
|
2022-10-10 15:21:13 +00:00
|
|
|
exclude_edge.remove(&(p, *n));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We compute the best flow using only the edges used in the previous assignation
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut g = self.generate_flow_graph(self.partition_size, zone_to_id, &exclude_edge)?;
|
|
|
|
g.compute_maximal_flow()?;
|
2022-10-11 16:29:21 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We add the excluded edges and compute the maximal flow with the full graph.
|
|
|
|
// The algorithm is such that it will start with the flow that we just computed
|
|
|
|
// and find ameliorating paths from that.
|
2022-10-10 15:21:13 +00:00
|
|
|
for (p, n) in exclude_edge.iter() {
|
|
|
|
let node_zone = zone_to_id[&self.get_node_zone(&self.node_id_vec[*n])?];
|
|
|
|
g.add_edge(Vertex::PZ(*p, node_zone), Vertex::N(*n), 1)?;
|
|
|
|
}
|
|
|
|
g.compute_maximal_flow()?;
|
|
|
|
Ok(g)
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This function updates the flow graph gflow to minimize the distance between
|
|
|
|
/// its corresponding assignation and the previous one
|
2022-10-10 15:21:13 +00:00
|
|
|
fn minimize_rebalance_load(
|
|
|
|
&self,
|
|
|
|
gflow: &mut Graph<FlowEdge>,
|
|
|
|
zone_to_id: &HashMap<String, usize>,
|
2022-10-11 16:29:21 +00:00
|
|
|
prev_assign: &[Vec<usize>],
|
2022-10-10 15:21:13 +00:00
|
|
|
) -> Result<(), Error> {
|
2022-11-07 18:34:40 +00:00
|
|
|
// We define a cost function on the edges (pairs of vertices) corresponding
|
|
|
|
// to the distance between the two assignations.
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut cost = CostFunction::new();
|
2022-10-11 16:29:21 +00:00
|
|
|
for (p, assoc_p) in prev_assign.iter().enumerate() {
|
2022-10-10 15:21:13 +00:00
|
|
|
for n in assoc_p.iter() {
|
|
|
|
let node_zone = zone_to_id[&self.get_node_zone(&self.node_id_vec[*n])?];
|
|
|
|
cost.insert((Vertex::PZ(p, node_zone), Vertex::N(*n)), -1);
|
|
|
|
}
|
|
|
|
}
|
2022-10-11 16:29:21 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We compute the maximal length of a simple path in gflow. It is used in the
|
|
|
|
// Bellman-Ford algorithm in optimize_flow_with_cost to set the number
|
|
|
|
// of iterations.
|
2022-10-11 16:29:21 +00:00
|
|
|
let nb_nodes = self.nongateway_nodes().len();
|
2022-10-10 15:21:13 +00:00
|
|
|
let path_length = 4 * nb_nodes;
|
|
|
|
gflow.optimize_flow_with_cost(&cost, path_length)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This function updates the assignation ring from the flow graph.
|
2022-10-10 15:21:13 +00:00
|
|
|
fn update_ring_from_flow(
|
|
|
|
&mut self,
|
|
|
|
nb_zones: usize,
|
|
|
|
gflow: &Graph<FlowEdge>,
|
|
|
|
) -> Result<(), Error> {
|
|
|
|
self.ring_assignation_data = Vec::<CompactNodeType>::new();
|
|
|
|
for p in 0..NB_PARTITIONS {
|
|
|
|
for z in 0..nb_zones {
|
|
|
|
let assoc_vertex = gflow.get_positive_flow_from(Vertex::PZ(p, z))?;
|
|
|
|
for vertex in assoc_vertex.iter() {
|
|
|
|
if let Vertex::N(n) = vertex {
|
|
|
|
self.ring_assignation_data.push((*n).try_into().unwrap());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if self.ring_assignation_data.len() != NB_PARTITIONS * self.replication_factor {
|
|
|
|
return Err(Error::Message(
|
|
|
|
"Critical Error : the association ring we produced does not \
|
|
|
|
have the right size."
|
|
|
|
.into(),
|
|
|
|
));
|
|
|
|
}
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
/// This function returns a message summing up the partition repartition of the new
|
|
|
|
/// layout, and other statistics of the partition assignation computation.
|
2022-10-10 15:21:13 +00:00
|
|
|
fn output_stat(
|
|
|
|
&self,
|
|
|
|
gflow: &Graph<FlowEdge>,
|
2022-10-11 16:29:21 +00:00
|
|
|
prev_assign_opt: &Option<Vec<Vec<usize>>>,
|
2022-10-10 15:21:13 +00:00
|
|
|
zone_to_id: &HashMap<String, usize>,
|
|
|
|
id_to_zone: &[String],
|
|
|
|
) -> Result<Message, Error> {
|
|
|
|
let mut msg = Message::new();
|
|
|
|
|
2022-11-07 20:12:11 +00:00
|
|
|
let used_cap = self.partition_size * NB_PARTITIONS as u64 * self.replication_factor as u64;
|
2022-10-10 15:21:13 +00:00
|
|
|
let total_cap = self.get_total_capacity()?;
|
|
|
|
let percent_cap = 100.0 * (used_cap as f32) / (total_cap as f32);
|
2022-10-11 15:17:13 +00:00
|
|
|
msg.push("".into());
|
2022-10-10 15:21:13 +00:00
|
|
|
msg.push(format!(
|
2022-10-11 15:17:13 +00:00
|
|
|
"Usable capacity / Total cluster capacity: {} / {} ({:.1} %)",
|
2022-11-07 20:12:11 +00:00
|
|
|
ByteSize::b(used_cap).to_string_as(false),
|
|
|
|
ByteSize::b(total_cap).to_string_as(false),
|
|
|
|
percent_cap
|
2022-10-10 15:21:13 +00:00
|
|
|
));
|
|
|
|
msg.push("".into());
|
|
|
|
msg.push(
|
2022-11-07 20:12:11 +00:00
|
|
|
"If the percentage is too low, it might be that the \
|
2022-09-22 17:30:01 +00:00
|
|
|
replication/redundancy constraints force the use of nodes/zones with small \
|
|
|
|
storage capacities. \
|
|
|
|
You might want to rebalance the storage capacities or relax the constraints. \
|
2022-10-10 15:21:13 +00:00
|
|
|
See the detailed statistics below and look for saturated nodes/zones."
|
|
|
|
.into(),
|
|
|
|
);
|
|
|
|
msg.push(format!(
|
|
|
|
"Recall that because of the replication factor, the actual available \
|
|
|
|
storage capacity is {} / {} = {}.",
|
2022-11-07 20:12:11 +00:00
|
|
|
ByteSize::b(used_cap).to_string_as(false),
|
2022-10-10 15:21:13 +00:00
|
|
|
self.replication_factor,
|
2022-11-07 20:12:11 +00:00
|
|
|
ByteSize::b(used_cap / self.replication_factor as u64).to_string_as(false)
|
2022-10-10 15:21:13 +00:00
|
|
|
));
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We define and fill in the following tables
|
2022-10-11 16:29:21 +00:00
|
|
|
let storing_nodes = self.nongateway_nodes();
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut new_partitions = vec![0; storing_nodes.len()];
|
|
|
|
let mut stored_partitions = vec![0; storing_nodes.len()];
|
|
|
|
|
|
|
|
let mut new_partitions_zone = vec![0; id_to_zone.len()];
|
|
|
|
let mut stored_partitions_zone = vec![0; id_to_zone.len()];
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
for p in 0..NB_PARTITIONS {
|
2022-10-10 15:21:13 +00:00
|
|
|
for z in 0..id_to_zone.len() {
|
|
|
|
let pz_nodes = gflow.get_positive_flow_from(Vertex::PZ(p, z))?;
|
|
|
|
if !pz_nodes.is_empty() {
|
|
|
|
stored_partitions_zone[z] += 1;
|
2022-10-11 16:29:21 +00:00
|
|
|
if let Some(prev_assign) = prev_assign_opt {
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut old_zones_of_p = Vec::<usize>::new();
|
2022-10-11 16:29:21 +00:00
|
|
|
for n in prev_assign[p].iter() {
|
2022-10-10 15:21:13 +00:00
|
|
|
old_zones_of_p
|
|
|
|
.push(zone_to_id[&self.get_node_zone(&self.node_id_vec[*n])?]);
|
|
|
|
}
|
|
|
|
if !old_zones_of_p.contains(&z) {
|
|
|
|
new_partitions_zone[z] += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for vert in pz_nodes.iter() {
|
|
|
|
if let Vertex::N(n) = *vert {
|
|
|
|
stored_partitions[n] += 1;
|
2022-10-11 16:29:21 +00:00
|
|
|
if let Some(prev_assign) = prev_assign_opt {
|
|
|
|
if !prev_assign[p].contains(&n) {
|
2022-10-10 15:21:13 +00:00
|
|
|
new_partitions[n] += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
if *prev_assign_opt == None {
|
2022-10-10 15:21:13 +00:00
|
|
|
new_partitions = stored_partitions.clone();
|
|
|
|
new_partitions_zone = stored_partitions_zone.clone();
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// We display the statistics
|
2022-10-10 15:21:13 +00:00
|
|
|
|
|
|
|
msg.push("".into());
|
2022-10-11 16:29:21 +00:00
|
|
|
if *prev_assign_opt != None {
|
2022-10-10 15:21:13 +00:00
|
|
|
let total_new_partitions: usize = new_partitions.iter().sum();
|
|
|
|
msg.push(format!(
|
|
|
|
"A total of {} new copies of partitions need to be \
|
|
|
|
transferred.",
|
|
|
|
total_new_partitions
|
|
|
|
));
|
|
|
|
}
|
|
|
|
msg.push("".into());
|
|
|
|
msg.push("==== DETAILED STATISTICS BY ZONES AND NODES ====".into());
|
|
|
|
|
|
|
|
for z in 0..id_to_zone.len() {
|
|
|
|
let mut nodes_of_z = Vec::<usize>::new();
|
|
|
|
for n in 0..storing_nodes.len() {
|
|
|
|
if self.get_node_zone(&self.node_id_vec[n])? == id_to_zone[z] {
|
|
|
|
nodes_of_z.push(n);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let replicated_partitions: usize =
|
|
|
|
nodes_of_z.iter().map(|n| stored_partitions[*n]).sum();
|
|
|
|
msg.push("".into());
|
|
|
|
|
|
|
|
msg.push(format!(
|
|
|
|
"Zone {}: {} distinct partitions stored ({} new, \
|
|
|
|
{} partition copies) ",
|
|
|
|
id_to_zone[z],
|
|
|
|
stored_partitions_zone[z],
|
|
|
|
new_partitions_zone[z],
|
|
|
|
replicated_partitions
|
|
|
|
));
|
|
|
|
|
2022-11-07 20:12:11 +00:00
|
|
|
let available_cap_z: u64 = self.partition_size * replicated_partitions as u64;
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut total_cap_z = 0;
|
|
|
|
for n in nodes_of_z.iter() {
|
|
|
|
total_cap_z += self.get_node_capacity(&self.node_id_vec[*n])?;
|
|
|
|
}
|
|
|
|
let percent_cap_z = 100.0 * (available_cap_z as f32) / (total_cap_z as f32);
|
|
|
|
msg.push(format!(
|
2022-11-07 20:12:11 +00:00
|
|
|
" Usable capacity / Total capacity: {} / {} ({:.1}%).",
|
|
|
|
ByteSize::b(available_cap_z).to_string_as(false),
|
|
|
|
ByteSize::b(total_cap_z).to_string_as(false),
|
|
|
|
percent_cap_z
|
2022-10-10 15:21:13 +00:00
|
|
|
));
|
|
|
|
|
|
|
|
for n in nodes_of_z.iter() {
|
2022-11-07 20:12:11 +00:00
|
|
|
let available_cap_n = stored_partitions[*n] as u64 * self.partition_size;
|
2022-10-10 15:21:13 +00:00
|
|
|
let total_cap_n = self.get_node_capacity(&self.node_id_vec[*n])?;
|
|
|
|
let tags_n = (self
|
|
|
|
.node_role(&self.node_id_vec[*n])
|
|
|
|
.ok_or("Node not found."))?
|
|
|
|
.tags_string();
|
|
|
|
msg.push(format!(
|
2022-11-07 20:12:11 +00:00
|
|
|
" Node {:?}: {} partitions ({} new) ; \
|
2022-10-11 15:17:13 +00:00
|
|
|
usable/total capacity: {} / {} ({:.1}%) ; tags:{}",
|
2022-11-07 20:12:11 +00:00
|
|
|
self.node_id_vec[*n],
|
2022-10-10 15:21:13 +00:00
|
|
|
stored_partitions[*n],
|
|
|
|
new_partitions[*n],
|
2022-11-07 20:12:11 +00:00
|
|
|
ByteSize::b(available_cap_n).to_string_as(false),
|
|
|
|
ByteSize::b(total_cap_n).to_string_as(false),
|
2022-10-10 15:21:13 +00:00
|
|
|
(available_cap_n as f32) / (total_cap_n as f32) * 100.0,
|
|
|
|
tags_n
|
|
|
|
));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
Ok(msg)
|
|
|
|
}
|
2022-05-01 07:57:05 +00:00
|
|
|
}
|
2021-11-09 11:24:04 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// ====================================================================================
|
2022-09-21 12:39:59 +00:00
|
|
|
|
2022-05-01 07:54:19 +00:00
|
|
|
#[cfg(test)]
|
|
|
|
mod tests {
|
2022-10-10 15:21:13 +00:00
|
|
|
use super::{Error, *};
|
|
|
|
use std::cmp::min;
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// This function checks that the partition size S computed is at least better than the
|
|
|
|
// one given by a very naive algorithm. To do so, we try to run the naive algorithm
|
|
|
|
// assuming a partion size of S+1. If we succed, it means that the optimal assignation
|
|
|
|
// was not optimal. The naive algorithm is the following :
|
|
|
|
// - we compute the max number of partitions associated to every node, capped at the
|
|
|
|
// partition number. It gives the number of tokens of every node.
|
|
|
|
// - every zone has a number of tokens equal to the sum of the tokens of its nodes.
|
|
|
|
// - we cycle over the partitions and associate zone tokens while respecting the
|
|
|
|
// zone redundancy constraint.
|
|
|
|
// NOTE: the naive algorithm is not optimal. Counter example:
|
|
|
|
// take nb_partition = 3 ; replication_factor = 5; redundancy = 4;
|
|
|
|
// number of tokens by zone : (A, 4), (B,1), (C,4), (D, 4), (E, 2)
|
|
|
|
// With these parameters, the naive algo fails, whereas there is a solution:
|
|
|
|
// (A,A,C,D,E) , (A,B,C,D,D) (A,C,C,D,E)
|
2022-10-10 15:21:13 +00:00
|
|
|
fn check_against_naive(cl: &ClusterLayout) -> Result<bool, Error> {
|
|
|
|
let over_size = cl.partition_size + 1;
|
|
|
|
let mut zone_token = HashMap::<String, usize>::new();
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
let (zones, zone_to_id) = cl.generate_nongateway_zone_ids()?;
|
2022-10-10 15:21:13 +00:00
|
|
|
|
|
|
|
if zones.is_empty() {
|
|
|
|
return Ok(false);
|
|
|
|
}
|
|
|
|
|
|
|
|
for z in zones.iter() {
|
|
|
|
zone_token.insert(z.clone(), 0);
|
|
|
|
}
|
2022-10-11 16:29:21 +00:00
|
|
|
for uuid in cl.nongateway_nodes().iter() {
|
2022-10-10 15:21:13 +00:00
|
|
|
let z = cl.get_node_zone(uuid)?;
|
|
|
|
let c = cl.get_node_capacity(uuid)?;
|
|
|
|
zone_token.insert(
|
|
|
|
z.clone(),
|
2022-10-11 16:29:21 +00:00
|
|
|
zone_token[&z] + min(NB_PARTITIONS, (c / over_size) as usize),
|
2022-10-10 15:21:13 +00:00
|
|
|
);
|
|
|
|
}
|
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
// For every partition, we count the number of zone already associated and
|
|
|
|
// the name of the last zone associated
|
2022-10-10 15:21:13 +00:00
|
|
|
|
|
|
|
let mut id_zone_token = vec![0; zones.len()];
|
|
|
|
for (z, t) in zone_token.iter() {
|
|
|
|
id_zone_token[zone_to_id[z]] = *t;
|
|
|
|
}
|
|
|
|
|
2022-10-11 16:29:21 +00:00
|
|
|
let mut nb_token = vec![0; NB_PARTITIONS];
|
|
|
|
let mut last_zone = vec![zones.len(); NB_PARTITIONS];
|
2022-10-10 15:21:13 +00:00
|
|
|
|
|
|
|
let mut curr_zone = 0;
|
|
|
|
|
|
|
|
let redundancy = cl.parameters.zone_redundancy;
|
|
|
|
|
|
|
|
for replic in 0..cl.replication_factor {
|
2022-10-11 16:29:21 +00:00
|
|
|
for p in 0..NB_PARTITIONS {
|
2022-10-10 15:21:13 +00:00
|
|
|
while id_zone_token[curr_zone] == 0
|
|
|
|
|| (last_zone[p] == curr_zone
|
|
|
|
&& redundancy - nb_token[p] <= cl.replication_factor - replic)
|
|
|
|
{
|
|
|
|
curr_zone += 1;
|
|
|
|
if curr_zone >= zones.len() {
|
|
|
|
return Ok(true);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
id_zone_token[curr_zone] -= 1;
|
|
|
|
if last_zone[p] != curr_zone {
|
|
|
|
nb_token[p] += 1;
|
|
|
|
last_zone[p] = curr_zone;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return Ok(false);
|
|
|
|
}
|
|
|
|
|
|
|
|
fn show_msg(msg: &Message) {
|
|
|
|
for s in msg.iter() {
|
|
|
|
println!("{}", s);
|
|
|
|
}
|
|
|
|
}
|
2022-05-01 07:57:05 +00:00
|
|
|
|
|
|
|
fn update_layout(
|
|
|
|
cl: &mut ClusterLayout,
|
|
|
|
node_id_vec: &Vec<u8>,
|
2022-11-07 20:12:11 +00:00
|
|
|
node_capacity_vec: &Vec<u64>,
|
2022-05-01 07:57:05 +00:00
|
|
|
node_zone_vec: &Vec<String>,
|
2022-10-10 15:21:13 +00:00
|
|
|
zone_redundancy: usize,
|
2022-05-01 07:57:05 +00:00
|
|
|
) {
|
|
|
|
for i in 0..node_id_vec.len() {
|
|
|
|
if let Some(x) = FixedBytes32::try_from(&[i as u8; 32]) {
|
|
|
|
cl.node_id_vec.push(x);
|
|
|
|
}
|
2021-11-09 11:24:04 +00:00
|
|
|
|
2022-11-07 18:34:40 +00:00
|
|
|
let update = cl.staging_roles.update_mutator(
|
2022-05-01 07:57:05 +00:00
|
|
|
cl.node_id_vec[i],
|
|
|
|
NodeRoleV(Some(NodeRole {
|
|
|
|
zone: (node_zone_vec[i].to_string()),
|
|
|
|
capacity: (Some(node_capacity_vec[i])),
|
|
|
|
tags: (vec![]),
|
|
|
|
})),
|
|
|
|
);
|
2022-11-07 18:34:40 +00:00
|
|
|
cl.staging_roles.merge(&update);
|
2022-05-01 07:57:05 +00:00
|
|
|
}
|
2022-11-07 18:34:40 +00:00
|
|
|
cl.staging_parameters
|
2022-10-11 15:17:13 +00:00
|
|
|
.update(LayoutParameters { zone_redundancy });
|
2022-11-08 13:23:08 +00:00
|
|
|
cl.staging_hash = cl.calculate_staging_hash();
|
2022-05-01 07:57:05 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_assignation() {
|
2022-10-10 15:21:13 +00:00
|
|
|
let mut node_id_vec = vec![1, 2, 3];
|
2022-05-01 07:57:05 +00:00
|
|
|
let mut node_capacity_vec = vec![4000, 1000, 2000];
|
|
|
|
let mut node_zone_vec = vec!["A", "B", "C"]
|
|
|
|
.into_iter()
|
|
|
|
.map(|x| x.to_string())
|
|
|
|
.collect();
|
|
|
|
|
2022-10-10 15:19:25 +00:00
|
|
|
let mut cl = ClusterLayout::new(3);
|
|
|
|
update_layout(&mut cl, &node_id_vec, &node_capacity_vec, &node_zone_vec, 3);
|
2022-10-11 15:17:13 +00:00
|
|
|
let v = cl.version;
|
|
|
|
let (mut cl, msg) = cl.apply_staged_changes(Some(v + 1)).unwrap();
|
|
|
|
show_msg(&msg);
|
2022-11-08 13:23:08 +00:00
|
|
|
assert_eq!(cl.check(), Ok(()));
|
2022-10-10 15:19:25 +00:00
|
|
|
assert!(matches!(check_against_naive(&cl), Ok(true)));
|
2022-05-01 07:57:05 +00:00
|
|
|
|
|
|
|
node_id_vec = vec![1, 2, 3, 4, 5, 6, 7, 8, 9];
|
|
|
|
node_capacity_vec = vec![4000, 1000, 1000, 3000, 1000, 1000, 2000, 10000, 2000];
|
|
|
|
node_zone_vec = vec!["A", "B", "C", "C", "C", "B", "G", "H", "I"]
|
|
|
|
.into_iter()
|
|
|
|
.map(|x| x.to_string())
|
|
|
|
.collect();
|
2022-10-10 15:19:25 +00:00
|
|
|
update_layout(&mut cl, &node_id_vec, &node_capacity_vec, &node_zone_vec, 2);
|
2022-10-11 15:17:13 +00:00
|
|
|
let v = cl.version;
|
|
|
|
let (mut cl, msg) = cl.apply_staged_changes(Some(v + 1)).unwrap();
|
|
|
|
show_msg(&msg);
|
2022-11-08 13:23:08 +00:00
|
|
|
assert_eq!(cl.check(), Ok(()));
|
2022-10-10 15:19:25 +00:00
|
|
|
assert!(matches!(check_against_naive(&cl), Ok(true)));
|
2022-05-01 07:57:05 +00:00
|
|
|
|
|
|
|
node_capacity_vec = vec![4000, 1000, 2000, 7000, 1000, 1000, 2000, 10000, 2000];
|
2022-10-10 15:19:25 +00:00
|
|
|
update_layout(&mut cl, &node_id_vec, &node_capacity_vec, &node_zone_vec, 3);
|
2022-10-11 15:17:13 +00:00
|
|
|
let v = cl.version;
|
|
|
|
let (mut cl, msg) = cl.apply_staged_changes(Some(v + 1)).unwrap();
|
|
|
|
show_msg(&msg);
|
2022-11-08 13:23:08 +00:00
|
|
|
assert_eq!(cl.check(), Ok(()));
|
2022-10-10 15:19:25 +00:00
|
|
|
assert!(matches!(check_against_naive(&cl), Ok(true)));
|
2022-05-01 07:57:05 +00:00
|
|
|
|
2022-10-10 15:21:13 +00:00
|
|
|
node_capacity_vec = vec![
|
|
|
|
4000000, 4000000, 2000000, 7000000, 1000000, 9000000, 2000000, 10000, 2000000,
|
|
|
|
];
|
2022-10-10 15:19:25 +00:00
|
|
|
update_layout(&mut cl, &node_id_vec, &node_capacity_vec, &node_zone_vec, 1);
|
2022-10-11 15:17:13 +00:00
|
|
|
let v = cl.version;
|
|
|
|
let (cl, msg) = cl.apply_staged_changes(Some(v + 1)).unwrap();
|
|
|
|
show_msg(&msg);
|
2022-11-08 13:23:08 +00:00
|
|
|
assert_eq!(cl.check(), Ok(()));
|
2022-10-10 15:19:25 +00:00
|
|
|
assert!(matches!(check_against_naive(&cl), Ok(true)));
|
2022-05-01 07:57:05 +00:00
|
|
|
}
|
|
|
|
}
|