Merge pull request 'Changed all instances of assignation to assignment' (#465) from jpds/garage:assignments-correction into next
All checks were successful
continuous-integration/drone/push Build is passing

Reviewed-on: #465
This commit is contained in:
Alex 2023-01-11 16:05:27 +00:00
commit 4f409f73dc
5 changed files with 73 additions and 73 deletions

View file

@ -210,7 +210,7 @@ pub async fn cmd_show_layout(
v + 1) v + 1)
} }
Err(e) => { Err(e) => {
println!("Error while trying to compute the assignation: {}", e); println!("Error while trying to compute the assignment: {}", e);
println!("This new layout cannot yet be applied."); println!("This new layout cannot yet be applied.");
println!( println!(
"You can also revert all proposed changes with: garage layout revert --version {}", "You can also revert all proposed changes with: garage layout revert --version {}",
@ -236,7 +236,7 @@ pub async fn cmd_apply_layout(
send_layout(rpc_cli, rpc_host, layout).await?; send_layout(rpc_cli, rpc_host, layout).await?;
println!("New cluster layout with updated role assignation has been applied in cluster."); println!("New cluster layout with updated role assignment has been applied in cluster.");
println!("Data will now be moved around between nodes accordingly."); println!("Data will now be moved around between nodes accordingly.");
Ok(()) Ok(())

View file

@ -17,7 +17,7 @@ pub enum Command {
#[structopt(name = "node", version = garage_version())] #[structopt(name = "node", version = garage_version())]
Node(NodeOperation), Node(NodeOperation),
/// Operations on the assignation of node roles in the cluster layout /// Operations on the assignment of node roles in the cluster layout
#[structopt(name = "layout", version = garage_version())] #[structopt(name = "layout", version = garage_version())]
Layout(LayoutOperation), Layout(LayoutOperation),

View file

@ -1,5 +1,5 @@
//! This module deals with graph algorithms. //! This module deals with graph algorithms.
//! It is used in layout.rs to build the partition to node assignation. //! It is used in layout.rs to build the partition to node assignment.
use rand::prelude::SliceRandom; use rand::prelude::SliceRandom;
use std::cmp::{max, min}; use std::cmp::{max, min};
@ -177,7 +177,7 @@ impl Graph<FlowEdge> {
let flow_upper_bound = self.flow_upper_bound()?; let flow_upper_bound = self.flow_upper_bound()?;
// To ensure the dispersion of the associations generated by the // To ensure the dispersion of the associations generated by the
// assignation, we shuffle the neighbours of the nodes. Hence, // assignment, we shuffle the neighbours of the nodes. Hence,
// the vertices do not consider their neighbours in the same order. // the vertices do not consider their neighbours in the same order.
self.shuffle_edges(); self.shuffle_edges();

View file

@ -34,8 +34,8 @@ pub struct ClusterLayout {
/// This attribute is only used to retain the previously computed partition size, /// This attribute is only used to retain the previously computed partition size,
/// to know to what extent does it change with the layout update. /// to know to what extent does it change with the layout update.
pub partition_size: u64, pub partition_size: u64,
/// Parameters used to compute the assignation currently given by /// Parameters used to compute the assignment currently given by
/// ring_assignation_data /// ring_assignment_data
pub parameters: LayoutParameters, pub parameters: LayoutParameters,
pub roles: LwwMap<Uuid, NodeRoleV>, pub roles: LwwMap<Uuid, NodeRoleV>,
@ -48,12 +48,12 @@ pub struct ClusterLayout {
/// 2. nodes that don't have a role are excluded (but they need to /// 2. nodes that don't have a role are excluded (but they need to
/// stay in the CRDT as tombstones) /// stay in the CRDT as tombstones)
pub node_id_vec: Vec<Uuid>, pub node_id_vec: Vec<Uuid>,
/// the assignation of data partitions to node, the values /// the assignment of data partitions to node, the values
/// are indices in node_id_vec /// are indices in node_id_vec
#[serde(with = "serde_bytes")] #[serde(with = "serde_bytes")]
pub ring_assignation_data: Vec<CompactNodeType>, pub ring_assignment_data: Vec<CompactNodeType>,
/// Parameters to be used in the next partition assignation computation. /// Parameters to be used in the next partition assignment computation.
pub staging_parameters: Lww<LayoutParameters>, pub staging_parameters: Lww<LayoutParameters>,
/// Role changes which are staged for the next version of the layout /// Role changes which are staged for the next version of the layout
pub staging_roles: LwwMap<Uuid, NodeRoleV>, pub staging_roles: LwwMap<Uuid, NodeRoleV>,
@ -61,7 +61,7 @@ pub struct ClusterLayout {
} }
impl garage_util::migrate::InitialFormat for ClusterLayout {} impl garage_util::migrate::InitialFormat for ClusterLayout {}
/// This struct is used to set the parameters to be used in the assignation computation /// This struct is used to set the parameters to be used in the assignment computation
/// algorithm. It is stored as a Crdt. /// algorithm. It is stored as a Crdt.
#[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)] #[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)]
pub struct LayoutParameters { pub struct LayoutParameters {
@ -106,7 +106,7 @@ impl NodeRole {
} }
} }
// Implementation of the ClusterLayout methods unrelated to the assignation algorithm. // Implementation of the ClusterLayout methods unrelated to the assignment algorithm.
impl ClusterLayout { impl ClusterLayout {
pub fn new(replication_factor: usize) -> Self { pub fn new(replication_factor: usize) -> Self {
// We set the default zone redundancy to be equal to the replication factor, // We set the default zone redundancy to be equal to the replication factor,
@ -124,7 +124,7 @@ impl ClusterLayout {
partition_size: 0, partition_size: 0,
roles: LwwMap::new(), roles: LwwMap::new(),
node_id_vec: Vec::new(), node_id_vec: Vec::new(),
ring_assignation_data: Vec::new(), ring_assignment_data: Vec::new(),
parameters, parameters,
staging_parameters, staging_parameters,
staging_roles: empty_lwwmap, staging_roles: empty_lwwmap,
@ -183,7 +183,7 @@ To know the correct value of the new layout version, invoke `garage layout show`
self.staging_roles.clear(); self.staging_roles.clear();
self.staging_hash = self.calculate_staging_hash(); self.staging_hash = self.calculate_staging_hash();
let msg = self.calculate_partition_assignation()?; let msg = self.calculate_partition_assignment()?;
self.version += 1; self.version += 1;
@ -276,7 +276,7 @@ To know the correct value of the new layout version, invoke `garage layout show`
for (i, id) in self.node_id_vec.iter().enumerate() { for (i, id) in self.node_id_vec.iter().enumerate() {
if id == uuid { if id == uuid {
let mut count = 0; let mut count = 0;
for nod in self.ring_assignation_data.iter() { for nod in self.ring_assignment_data.iter() {
if i as u8 == *nod { if i as u8 == *nod {
count += 1 count += 1
} }
@ -301,7 +301,7 @@ To know the correct value of the new layout version, invoke `garage layout show`
} }
/// Check a cluster layout for internal consistency /// Check a cluster layout for internal consistency
/// (assignation, roles, parameters, partition size) /// (assignment, roles, parameters, partition size)
/// returns true if consistent, false if error /// returns true if consistent, false if error
pub fn check(&self) -> Result<(), String> { pub fn check(&self) -> Result<(), String> {
// Check that the hash of the staging data is correct // Check that the hash of the staging data is correct
@ -325,37 +325,37 @@ To know the correct value of the new layout version, invoke `garage layout show`
return Err(format!("node_id_vec does not contain the correct set of nodes\nnode_id_vec: {:?}\nexpected: {:?}", node_id_vec, expected_nodes)); return Err(format!("node_id_vec does not contain the correct set of nodes\nnode_id_vec: {:?}\nexpected: {:?}", node_id_vec, expected_nodes));
} }
// Check that the assignation data has the correct length // Check that the assignment data has the correct length
let expected_assignation_data_len = (1 << PARTITION_BITS) * self.replication_factor; let expected_assignment_data_len = (1 << PARTITION_BITS) * self.replication_factor;
if self.ring_assignation_data.len() != expected_assignation_data_len { if self.ring_assignment_data.len() != expected_assignment_data_len {
return Err(format!( return Err(format!(
"ring_assignation_data has incorrect length {} instead of {}", "ring_assignment_data has incorrect length {} instead of {}",
self.ring_assignation_data.len(), self.ring_assignment_data.len(),
expected_assignation_data_len expected_assignment_data_len
)); ));
} }
// Check that the assigned nodes are correct identifiers // Check that the assigned nodes are correct identifiers
// of nodes that are assigned a role // of nodes that are assigned a role
// and that role is not the role of a gateway nodes // and that role is not the role of a gateway nodes
for x in self.ring_assignation_data.iter() { for x in self.ring_assignment_data.iter() {
if *x as usize >= self.node_id_vec.len() { if *x as usize >= self.node_id_vec.len() {
return Err(format!( return Err(format!(
"ring_assignation_data contains invalid node id {}", "ring_assignment_data contains invalid node id {}",
*x *x
)); ));
} }
let node = self.node_id_vec[*x as usize]; let node = self.node_id_vec[*x as usize];
match self.roles.get(&node) { match self.roles.get(&node) {
Some(NodeRoleV(Some(x))) if x.capacity.is_some() => (), Some(NodeRoleV(Some(x))) if x.capacity.is_some() => (),
_ => return Err("ring_assignation_data contains id of a gateway node".into()), _ => return Err("ring_assignment_data contains id of a gateway node".into()),
} }
} }
// Check that every partition is associated to distinct nodes // Check that every partition is associated to distinct nodes
let rf = self.replication_factor; let rf = self.replication_factor;
for p in 0..(1 << PARTITION_BITS) { for p in 0..(1 << PARTITION_BITS) {
let nodes_of_p = self.ring_assignation_data[rf * p..rf * (p + 1)].to_vec(); let nodes_of_p = self.ring_assignment_data[rf * p..rf * (p + 1)].to_vec();
if nodes_of_p.iter().unique().count() != rf { if nodes_of_p.iter().unique().count() != rf {
return Err(format!("partition does not contain {} unique node ids", rf)); return Err(format!("partition does not contain {} unique node ids", rf));
} }
@ -378,7 +378,7 @@ To know the correct value of the new layout version, invoke `garage layout show`
// Check that the nodes capacities is consistent with the stored partitions // Check that the nodes capacities is consistent with the stored partitions
let mut node_usage = vec![0; MAX_NODE_NUMBER]; let mut node_usage = vec![0; MAX_NODE_NUMBER];
for n in self.ring_assignation_data.iter() { for n in self.ring_assignment_data.iter() {
node_usage[*n as usize] += 1; node_usage[*n as usize] += 1;
} }
for (n, usage) in node_usage.iter().enumerate() { for (n, usage) in node_usage.iter().enumerate() {
@ -415,21 +415,21 @@ To know the correct value of the new layout version, invoke `garage layout show`
} }
} }
// Implementation of the ClusterLayout methods related to the assignation algorithm. // Implementation of the ClusterLayout methods related to the assignment algorithm.
impl ClusterLayout { impl ClusterLayout {
/// This function calculates a new partition-to-node assignation. /// This function calculates a new partition-to-node assignment.
/// The computed assignation respects the node replication factor /// The computed assignment respects the node replication factor
/// and the zone redundancy parameter It maximizes the capacity of a /// and the zone redundancy parameter It maximizes the capacity of a
/// partition (assuming all partitions have the same size). /// partition (assuming all partitions have the same size).
/// Among such optimal assignation, it minimizes the distance to /// Among such optimal assignment, it minimizes the distance to
/// the former assignation (if any) to minimize the amount of /// the former assignment (if any) to minimize the amount of
/// data to be moved. /// data to be moved.
/// Staged role changes must be merged with nodes roles before calling this function, /// Staged role changes must be merged with nodes roles before calling this function,
/// hence it must only be called from apply_staged_changes() and hence is not public. /// hence it must only be called from apply_staged_changes() and hence is not public.
fn calculate_partition_assignation(&mut self) -> Result<Message, Error> { fn calculate_partition_assignment(&mut self) -> Result<Message, Error> {
// We update the node ids, since the node role list might have changed with the // We update the node ids, since the node role list might have changed with the
// changes in the layout. We retrieve the old_assignation reframed with new ids // changes in the layout. We retrieve the old_assignment reframed with new ids
let old_assignation_opt = self.update_node_id_vec()?; let old_assignment_opt = self.update_node_id_vec()?;
let mut msg = Message::new(); let mut msg = Message::new();
msg.push("==== COMPUTATION OF A NEW PARTITION ASSIGNATION ====".into()); msg.push("==== COMPUTATION OF A NEW PARTITION ASSIGNATION ====".into());
@ -467,7 +467,7 @@ impl ClusterLayout {
// optimality. // optimality.
let partition_size = self.compute_optimal_partition_size(&zone_to_id)?; let partition_size = self.compute_optimal_partition_size(&zone_to_id)?;
if old_assignation_opt != None { if old_assignment_opt != None {
msg.push(format!( msg.push(format!(
"Optimal size of a partition: {} (was {} in the previous layout).", "Optimal size of a partition: {} (was {} in the previous layout).",
ByteSize::b(partition_size).to_string_as(false), ByteSize::b(partition_size).to_string_as(false),
@ -490,16 +490,16 @@ impl ClusterLayout {
); );
} }
// We compute a first flow/assignation that is heuristically close to the previous // We compute a first flow/assignment that is heuristically close to the previous
// assignation // assignment
let mut gflow = self.compute_candidate_assignation(&zone_to_id, &old_assignation_opt)?; let mut gflow = self.compute_candidate_assignment(&zone_to_id, &old_assignment_opt)?;
if let Some(assoc) = &old_assignation_opt { if let Some(assoc) = &old_assignment_opt {
// We minimize the distance to the previous assignation. // We minimize the distance to the previous assignment.
self.minimize_rebalance_load(&mut gflow, &zone_to_id, assoc)?; self.minimize_rebalance_load(&mut gflow, &zone_to_id, assoc)?;
} }
// We display statistics of the computation // We display statistics of the computation
msg.extend(self.output_stat(&gflow, &old_assignation_opt, &zone_to_id, &id_to_zone)?); msg.extend(self.output_stat(&gflow, &old_assignment_opt, &zone_to_id, &id_to_zone)?);
msg.push("".to_string()); msg.push("".to_string());
// We update the layout structure // We update the layout structure
@ -515,10 +515,10 @@ impl ClusterLayout {
} }
/// The LwwMap of node roles might have changed. This function updates the node_id_vec /// The LwwMap of node roles might have changed. This function updates the node_id_vec
/// and returns the assignation given by ring, with the new indices of the nodes, and /// and returns the assignment given by ring, with the new indices of the nodes, and
/// None if the node is not present anymore. /// None if the node is not present anymore.
/// We work with the assumption that only this function and calculate_new_assignation /// We work with the assumption that only this function and calculate_new_assignment
/// do modify assignation_ring and node_id_vec. /// do modify assignment_ring and node_id_vec.
fn update_node_id_vec(&mut self) -> Result<Option<Vec<Vec<usize>>>, Error> { fn update_node_id_vec(&mut self) -> Result<Option<Vec<Vec<usize>>>, Error> {
// (1) We compute the new node list // (1) We compute the new node list
// Non gateway nodes should be coded on 8bits, hence they must be first in the list // Non gateway nodes should be coded on 8bits, hence they must be first in the list
@ -556,15 +556,15 @@ impl ClusterLayout {
// (2) We retrieve the old association // (2) We retrieve the old association
// We rewrite the old association with the new indices. We only consider partition // We rewrite the old association with the new indices. We only consider partition
// to node assignations where the node is still in use. // to node assignments where the node is still in use.
if self.ring_assignation_data.is_empty() { if self.ring_assignment_data.is_empty() {
// This is a new association // This is a new association
return Ok(None); return Ok(None);
} }
if self.ring_assignation_data.len() != NB_PARTITIONS * self.replication_factor { if self.ring_assignment_data.len() != NB_PARTITIONS * self.replication_factor {
return Err(Error::Message( return Err(Error::Message(
"The old assignation does not have a size corresponding to \ "The old assignment does not have a size corresponding to \
the old replication factor or the number of partitions." the old replication factor or the number of partitions."
.into(), .into(),
)); ));
@ -579,11 +579,11 @@ impl ClusterLayout {
uuid_to_new_id.insert(*uuid, i); uuid_to_new_id.insert(*uuid, i);
} }
let mut old_assignation = vec![Vec::<usize>::new(); NB_PARTITIONS]; let mut old_assignment = vec![Vec::<usize>::new(); NB_PARTITIONS];
let rf = self.replication_factor; let rf = self.replication_factor;
for (p, old_assign_p) in old_assignation.iter_mut().enumerate() { for (p, old_assign_p) in old_assignment.iter_mut().enumerate() {
for old_id in &self.ring_assignation_data[p * rf..(p + 1) * rf] { for old_id in &self.ring_assignment_data[p * rf..(p + 1) * rf] {
let uuid = old_node_id_vec[*old_id as usize]; let uuid = old_node_id_vec[*old_id as usize];
if uuid_to_new_id.contains_key(&uuid) { if uuid_to_new_id.contains_key(&uuid) {
old_assign_p.push(uuid_to_new_id[&uuid]); old_assign_p.push(uuid_to_new_id[&uuid]);
@ -592,9 +592,9 @@ impl ClusterLayout {
} }
// We write the ring // We write the ring
self.ring_assignation_data = Vec::<CompactNodeType>::new(); self.ring_assignment_data = Vec::<CompactNodeType>::new();
Ok(Some(old_assignation)) Ok(Some(old_assignment))
} }
/// This function generates ids for the zone of the nodes appearing in /// This function generates ids for the zone of the nodes appearing in
@ -661,11 +661,11 @@ impl ClusterLayout {
} }
/// Generates the graph to compute the maximal flow corresponding to the optimal /// Generates the graph to compute the maximal flow corresponding to the optimal
/// partition assignation. /// partition assignment.
/// exclude_assoc is the set of (partition, node) association that we are forbidden /// exclude_assoc is the set of (partition, node) association that we are forbidden
/// to use (hence we do not add the corresponding edge to the graph). This parameter /// to use (hence we do not add the corresponding edge to the graph). This parameter
/// is used to compute a first flow that uses only edges appearing in the previous /// is used to compute a first flow that uses only edges appearing in the previous
/// assignation. This produces a solution that heuristically should be close to the /// assignment. This produces a solution that heuristically should be close to the
/// previous one. /// previous one.
fn generate_flow_graph( fn generate_flow_graph(
&self, &self,
@ -707,14 +707,14 @@ impl ClusterLayout {
Ok(g) Ok(g)
} }
/// This function computes a first optimal assignation (in the form of a flow graph). /// This function computes a first optimal assignment (in the form of a flow graph).
fn compute_candidate_assignation( fn compute_candidate_assignment(
&self, &self,
zone_to_id: &HashMap<String, usize>, zone_to_id: &HashMap<String, usize>,
prev_assign_opt: &Option<Vec<Vec<usize>>>, prev_assign_opt: &Option<Vec<Vec<usize>>>,
) -> Result<Graph<FlowEdge>, Error> { ) -> Result<Graph<FlowEdge>, Error> {
// We list the (partition,node) associations that are not used in the // We list the (partition,node) associations that are not used in the
// previous assignation // previous assignment
let mut exclude_edge = HashSet::<(usize, usize)>::new(); let mut exclude_edge = HashSet::<(usize, usize)>::new();
if let Some(prev_assign) = prev_assign_opt { if let Some(prev_assign) = prev_assign_opt {
let nb_nodes = self.nongateway_nodes().len(); let nb_nodes = self.nongateway_nodes().len();
@ -728,7 +728,7 @@ impl ClusterLayout {
} }
} }
// We compute the best flow using only the edges used in the previous assignation // We compute the best flow using only the edges used in the previous assignment
let mut g = self.generate_flow_graph(self.partition_size, zone_to_id, &exclude_edge)?; let mut g = self.generate_flow_graph(self.partition_size, zone_to_id, &exclude_edge)?;
g.compute_maximal_flow()?; g.compute_maximal_flow()?;
@ -744,7 +744,7 @@ impl ClusterLayout {
} }
/// This function updates the flow graph gflow to minimize the distance between /// This function updates the flow graph gflow to minimize the distance between
/// its corresponding assignation and the previous one /// its corresponding assignment and the previous one
fn minimize_rebalance_load( fn minimize_rebalance_load(
&self, &self,
gflow: &mut Graph<FlowEdge>, gflow: &mut Graph<FlowEdge>,
@ -752,7 +752,7 @@ impl ClusterLayout {
prev_assign: &[Vec<usize>], prev_assign: &[Vec<usize>],
) -> Result<(), Error> { ) -> Result<(), Error> {
// We define a cost function on the edges (pairs of vertices) corresponding // We define a cost function on the edges (pairs of vertices) corresponding
// to the distance between the two assignations. // to the distance between the two assignments.
let mut cost = CostFunction::new(); let mut cost = CostFunction::new();
for (p, assoc_p) in prev_assign.iter().enumerate() { for (p, assoc_p) in prev_assign.iter().enumerate() {
for n in assoc_p.iter() { for n in assoc_p.iter() {
@ -771,25 +771,25 @@ impl ClusterLayout {
Ok(()) Ok(())
} }
/// This function updates the assignation ring from the flow graph. /// This function updates the assignment ring from the flow graph.
fn update_ring_from_flow( fn update_ring_from_flow(
&mut self, &mut self,
nb_zones: usize, nb_zones: usize,
gflow: &Graph<FlowEdge>, gflow: &Graph<FlowEdge>,
) -> Result<(), Error> { ) -> Result<(), Error> {
self.ring_assignation_data = Vec::<CompactNodeType>::new(); self.ring_assignment_data = Vec::<CompactNodeType>::new();
for p in 0..NB_PARTITIONS { for p in 0..NB_PARTITIONS {
for z in 0..nb_zones { for z in 0..nb_zones {
let assoc_vertex = gflow.get_positive_flow_from(Vertex::PZ(p, z))?; let assoc_vertex = gflow.get_positive_flow_from(Vertex::PZ(p, z))?;
for vertex in assoc_vertex.iter() { for vertex in assoc_vertex.iter() {
if let Vertex::N(n) = vertex { if let Vertex::N(n) = vertex {
self.ring_assignation_data.push((*n).try_into().unwrap()); self.ring_assignment_data.push((*n).try_into().unwrap());
} }
} }
} }
} }
if self.ring_assignation_data.len() != NB_PARTITIONS * self.replication_factor { if self.ring_assignment_data.len() != NB_PARTITIONS * self.replication_factor {
return Err(Error::Message( return Err(Error::Message(
"Critical Error : the association ring we produced does not \ "Critical Error : the association ring we produced does not \
have the right size." have the right size."
@ -800,7 +800,7 @@ impl ClusterLayout {
} }
/// This function returns a message summing up the partition repartition of the new /// This function returns a message summing up the partition repartition of the new
/// layout, and other statistics of the partition assignation computation. /// layout, and other statistics of the partition assignment computation.
fn output_stat( fn output_stat(
&self, &self,
gflow: &Graph<FlowEdge>, gflow: &Graph<FlowEdge>,
@ -960,7 +960,7 @@ mod tests {
// This function checks that the partition size S computed is at least better than the // This function checks that the partition size S computed is at least better than the
// one given by a very naive algorithm. To do so, we try to run the naive algorithm // one given by a very naive algorithm. To do so, we try to run the naive algorithm
// assuming a partion size of S+1. If we succed, it means that the optimal assignation // assuming a partion size of S+1. If we succed, it means that the optimal assignment
// was not optimal. The naive algorithm is the following : // was not optimal. The naive algorithm is the following :
// - we compute the max number of partitions associated to every node, capped at the // - we compute the max number of partitions associated to every node, capped at the
// partition number. It gives the number of tokens of every node. // partition number. It gives the number of tokens of every node.
@ -1065,7 +1065,7 @@ mod tests {
} }
#[test] #[test]
fn test_assignation() { fn test_assignment() {
let mut node_id_vec = vec![1, 2, 3]; let mut node_id_vec = vec![1, 2, 3];
let mut node_capacity_vec = vec![4000, 1000, 2000]; let mut node_capacity_vec = vec![4000, 1000, 2000];
let mut node_zone_vec = vec!["A", "B", "C"] let mut node_zone_vec = vec!["A", "B", "C"]

View file

@ -63,12 +63,12 @@ struct RingEntry {
impl Ring { impl Ring {
pub(crate) fn new(layout: ClusterLayout, replication_factor: usize) -> Self { pub(crate) fn new(layout: ClusterLayout, replication_factor: usize) -> Self {
if replication_factor != layout.replication_factor { if replication_factor != layout.replication_factor {
warn!("Could not build ring: replication factor does not match between local configuration and network role assignation."); warn!("Could not build ring: replication factor does not match between local configuration and network role assignment.");
return Self::empty(layout, replication_factor); return Self::empty(layout, replication_factor);
} }
if layout.ring_assignation_data.len() != replication_factor * (1 << PARTITION_BITS) { if layout.ring_assignment_data.len() != replication_factor * (1 << PARTITION_BITS) {
warn!("Could not build ring: network role assignation data has invalid length"); warn!("Could not build ring: network role assignment data has invalid length");
return Self::empty(layout, replication_factor); return Self::empty(layout, replication_factor);
} }
@ -78,7 +78,7 @@ impl Ring {
let top = (i as u16) << (16 - PARTITION_BITS); let top = (i as u16) << (16 - PARTITION_BITS);
let mut nodes_buf = [0u8; MAX_REPLICATION]; let mut nodes_buf = [0u8; MAX_REPLICATION];
nodes_buf[..replication_factor].copy_from_slice( nodes_buf[..replication_factor].copy_from_slice(
&layout.ring_assignation_data &layout.ring_assignment_data
[replication_factor * i..replication_factor * (i + 1)], [replication_factor * i..replication_factor * (i + 1)],
); );
RingEntry { RingEntry {