forked from Deuxfleurs/garage
329 lines
8.3 KiB
Rust
329 lines
8.3 KiB
Rust
use std::collections::{BTreeMap, HashMap};
|
|
use std::sync::Arc;
|
|
use std::time::Duration;
|
|
|
|
use futures::stream::*;
|
|
use serde::{Deserialize, Serialize};
|
|
use serde_bytes::ByteBuf;
|
|
|
|
use garage_util::data::*;
|
|
use garage_util::error::Error;
|
|
|
|
use garage_rpc::membership::System;
|
|
use garage_rpc::rpc_client::*;
|
|
use garage_rpc::rpc_server::*;
|
|
|
|
use crate::crdt::CRDT;
|
|
use crate::data::*;
|
|
use crate::schema::*;
|
|
use crate::table_sync::*;
|
|
use crate::replication::*;
|
|
|
|
const TABLE_RPC_TIMEOUT: Duration = Duration::from_secs(10);
|
|
|
|
pub struct TableAux<F: TableSchema, R: TableReplication> {
|
|
pub system: Arc<System>,
|
|
pub replication: R,
|
|
pub(crate) rpc_client: Arc<RpcClient<TableRPC<F>>>,
|
|
}
|
|
|
|
pub struct Table<F: TableSchema, R: TableReplication> {
|
|
pub data: Arc<TableData<F>>,
|
|
pub aux: Arc<TableAux<F, R>>,
|
|
pub syncer: Arc<TableSyncer<F, R>>,
|
|
}
|
|
|
|
#[derive(Serialize, Deserialize)]
|
|
pub(crate) enum TableRPC<F: TableSchema> {
|
|
Ok,
|
|
|
|
ReadEntry(F::P, F::S),
|
|
ReadEntryResponse(Option<ByteBuf>),
|
|
|
|
// Read range: read all keys in partition P, possibly starting at a certain sort key offset
|
|
ReadRange(F::P, Option<F::S>, Option<F::Filter>, usize),
|
|
|
|
Update(Vec<Arc<ByteBuf>>),
|
|
|
|
SyncRPC(SyncRPC),
|
|
}
|
|
|
|
impl<F: TableSchema> RpcMessage for TableRPC<F> {}
|
|
|
|
|
|
impl<F, R> Table<F, R>
|
|
where
|
|
F: TableSchema + 'static,
|
|
R: TableReplication + 'static,
|
|
{
|
|
// =============== PUBLIC INTERFACE FUNCTIONS (new, insert, get, etc) ===============
|
|
|
|
pub fn new(
|
|
instance: F,
|
|
replication: R,
|
|
system: Arc<System>,
|
|
db: &sled::Db,
|
|
name: String,
|
|
rpc_server: &mut RpcServer,
|
|
) -> Arc<Self> {
|
|
let rpc_path = format!("table_{}", name);
|
|
let rpc_client = system.rpc_client::<TableRPC<F>>(&rpc_path);
|
|
|
|
let data = TableData::new(
|
|
name,
|
|
instance,
|
|
db,
|
|
system.background.clone(),
|
|
);
|
|
|
|
let aux = Arc::new(TableAux{
|
|
system,
|
|
replication,
|
|
rpc_client,
|
|
});
|
|
|
|
let syncer = TableSyncer::launch(
|
|
data.clone(),
|
|
aux.clone(),
|
|
);
|
|
|
|
let table = Arc::new(Self {
|
|
data,
|
|
aux,
|
|
syncer,
|
|
});
|
|
|
|
table.clone().register_handler(rpc_server, rpc_path);
|
|
|
|
table
|
|
}
|
|
|
|
pub async fn insert(&self, e: &F::E) -> Result<(), Error> {
|
|
let hash = e.partition_key().hash();
|
|
let who = self.aux.replication.write_nodes(&hash, &self.aux.system);
|
|
//eprintln!("insert who: {:?}", who);
|
|
|
|
let e_enc = Arc::new(ByteBuf::from(rmp_to_vec_all_named(e)?));
|
|
let rpc = TableRPC::<F>::Update(vec![e_enc]);
|
|
|
|
self.aux.rpc_client
|
|
.try_call_many(
|
|
&who[..],
|
|
rpc,
|
|
RequestStrategy::with_quorum(self.aux.replication.write_quorum(&self.aux.system))
|
|
.with_timeout(TABLE_RPC_TIMEOUT),
|
|
)
|
|
.await?;
|
|
Ok(())
|
|
}
|
|
|
|
pub async fn insert_many(&self, entries: &[F::E]) -> Result<(), Error> {
|
|
let mut call_list = HashMap::new();
|
|
|
|
for entry in entries.iter() {
|
|
let hash = entry.partition_key().hash();
|
|
let who = self.aux.replication.write_nodes(&hash, &self.aux.system);
|
|
let e_enc = Arc::new(ByteBuf::from(rmp_to_vec_all_named(entry)?));
|
|
for node in who {
|
|
if !call_list.contains_key(&node) {
|
|
call_list.insert(node, vec![]);
|
|
}
|
|
call_list.get_mut(&node).unwrap().push(e_enc.clone());
|
|
}
|
|
}
|
|
|
|
let call_futures = call_list.drain().map(|(node, entries)| async move {
|
|
let rpc = TableRPC::<F>::Update(entries);
|
|
|
|
let resp = self.aux.rpc_client.call(node, rpc, TABLE_RPC_TIMEOUT).await?;
|
|
Ok::<_, Error>((node, resp))
|
|
});
|
|
let mut resps = call_futures.collect::<FuturesUnordered<_>>();
|
|
let mut errors = vec![];
|
|
|
|
while let Some(resp) = resps.next().await {
|
|
if let Err(e) = resp {
|
|
errors.push(e);
|
|
}
|
|
}
|
|
if errors.len() > self.aux.replication.max_write_errors() {
|
|
Err(Error::Message("Too many errors".into()))
|
|
} else {
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
pub async fn get(
|
|
self: &Arc<Self>,
|
|
partition_key: &F::P,
|
|
sort_key: &F::S,
|
|
) -> Result<Option<F::E>, Error> {
|
|
let hash = partition_key.hash();
|
|
let who = self.aux.replication.read_nodes(&hash, &self.aux.system);
|
|
//eprintln!("get who: {:?}", who);
|
|
|
|
let rpc = TableRPC::<F>::ReadEntry(partition_key.clone(), sort_key.clone());
|
|
let resps = self
|
|
.aux
|
|
.rpc_client
|
|
.try_call_many(
|
|
&who[..],
|
|
rpc,
|
|
RequestStrategy::with_quorum(self.aux.replication.read_quorum())
|
|
.with_timeout(TABLE_RPC_TIMEOUT)
|
|
.interrupt_after_quorum(true),
|
|
)
|
|
.await?;
|
|
|
|
let mut ret = None;
|
|
let mut not_all_same = false;
|
|
for resp in resps {
|
|
if let TableRPC::ReadEntryResponse(value) = resp {
|
|
if let Some(v_bytes) = value {
|
|
let v = self.data.decode_entry(v_bytes.as_slice())?;
|
|
ret = match ret {
|
|
None => Some(v),
|
|
Some(mut x) => {
|
|
if x != v {
|
|
not_all_same = true;
|
|
x.merge(&v);
|
|
}
|
|
Some(x)
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
return Err(Error::Message(format!("Invalid return value to read")));
|
|
}
|
|
}
|
|
if let Some(ret_entry) = &ret {
|
|
if not_all_same {
|
|
let self2 = self.clone();
|
|
let ent2 = ret_entry.clone();
|
|
self.aux.system
|
|
.background
|
|
.spawn_cancellable(async move { self2.repair_on_read(&who[..], ent2).await });
|
|
}
|
|
}
|
|
Ok(ret)
|
|
}
|
|
|
|
pub async fn get_range(
|
|
self: &Arc<Self>,
|
|
partition_key: &F::P,
|
|
begin_sort_key: Option<F::S>,
|
|
filter: Option<F::Filter>,
|
|
limit: usize,
|
|
) -> Result<Vec<F::E>, Error> {
|
|
let hash = partition_key.hash();
|
|
let who = self.aux.replication.read_nodes(&hash, &self.aux.system);
|
|
|
|
let rpc = TableRPC::<F>::ReadRange(partition_key.clone(), begin_sort_key, filter, limit);
|
|
|
|
let resps = self
|
|
.aux.rpc_client
|
|
.try_call_many(
|
|
&who[..],
|
|
rpc,
|
|
RequestStrategy::with_quorum(self.aux.replication.read_quorum())
|
|
.with_timeout(TABLE_RPC_TIMEOUT)
|
|
.interrupt_after_quorum(true),
|
|
)
|
|
.await?;
|
|
|
|
let mut ret = BTreeMap::new();
|
|
let mut to_repair = BTreeMap::new();
|
|
for resp in resps {
|
|
if let TableRPC::Update(entries) = resp {
|
|
for entry_bytes in entries.iter() {
|
|
let entry = self.data.decode_entry(entry_bytes.as_slice())?;
|
|
let entry_key = self.data.tree_key(entry.partition_key(), entry.sort_key());
|
|
match ret.remove(&entry_key) {
|
|
None => {
|
|
ret.insert(entry_key, Some(entry));
|
|
}
|
|
Some(Some(mut prev)) => {
|
|
let must_repair = prev != entry;
|
|
prev.merge(&entry);
|
|
if must_repair {
|
|
to_repair.insert(entry_key.clone(), Some(prev.clone()));
|
|
}
|
|
ret.insert(entry_key, Some(prev));
|
|
}
|
|
Some(None) => unreachable!(),
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if !to_repair.is_empty() {
|
|
let self2 = self.clone();
|
|
self.aux.system.background.spawn_cancellable(async move {
|
|
for (_, v) in to_repair.iter_mut() {
|
|
self2.repair_on_read(&who[..], v.take().unwrap()).await?;
|
|
}
|
|
Ok(())
|
|
});
|
|
}
|
|
let ret_vec = ret
|
|
.iter_mut()
|
|
.take(limit)
|
|
.map(|(_k, v)| v.take().unwrap())
|
|
.collect::<Vec<_>>();
|
|
Ok(ret_vec)
|
|
}
|
|
|
|
// =============== UTILITY FUNCTION FOR CLIENT OPERATIONS ===============
|
|
|
|
async fn repair_on_read(&self, who: &[UUID], what: F::E) -> Result<(), Error> {
|
|
let what_enc = Arc::new(ByteBuf::from(rmp_to_vec_all_named(&what)?));
|
|
self.aux.rpc_client
|
|
.try_call_many(
|
|
&who[..],
|
|
TableRPC::<F>::Update(vec![what_enc]),
|
|
RequestStrategy::with_quorum(who.len()).with_timeout(TABLE_RPC_TIMEOUT),
|
|
)
|
|
.await?;
|
|
Ok(())
|
|
}
|
|
|
|
// =============== HANDLERS FOR RPC OPERATIONS (SERVER SIDE) ==============
|
|
|
|
fn register_handler(self: Arc<Self>, rpc_server: &mut RpcServer, path: String) {
|
|
let self2 = self.clone();
|
|
rpc_server.add_handler::<TableRPC<F>, _, _>(path, move |msg, _addr| {
|
|
let self2 = self2.clone();
|
|
async move { self2.handle(&msg).await }
|
|
});
|
|
|
|
let self2 = self.clone();
|
|
self.aux.rpc_client
|
|
.set_local_handler(self.aux.system.id, move |msg| {
|
|
let self2 = self2.clone();
|
|
async move { self2.handle(&msg).await }
|
|
});
|
|
}
|
|
|
|
async fn handle(self: &Arc<Self>, msg: &TableRPC<F>) -> Result<TableRPC<F>, Error> {
|
|
match msg {
|
|
TableRPC::ReadEntry(key, sort_key) => {
|
|
let value = self.data.read_entry(key, sort_key)?;
|
|
Ok(TableRPC::ReadEntryResponse(value))
|
|
}
|
|
TableRPC::ReadRange(key, begin_sort_key, filter, limit) => {
|
|
let values = self.data.read_range(key, begin_sort_key, filter, *limit)?;
|
|
Ok(TableRPC::Update(values))
|
|
}
|
|
TableRPC::Update(pairs) => {
|
|
self.data.update_many(pairs)?;
|
|
Ok(TableRPC::Ok)
|
|
}
|
|
TableRPC::SyncRPC(rpc) => {
|
|
let response = self.syncer
|
|
.handle_rpc(rpc, self.aux.system.background.stop_signal.clone())
|
|
.await?;
|
|
Ok(TableRPC::SyncRPC(response))
|
|
}
|
|
_ => Err(Error::BadRPC(format!("Unexpected table RPC"))),
|
|
}
|
|
}
|
|
}
|