2020-04-09 21:45:07 +00:00
|
|
|
use std::path::PathBuf;
|
2020-04-17 13:36:16 +00:00
|
|
|
use std::sync::Arc;
|
|
|
|
use std::time::Duration;
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-17 16:51:29 +00:00
|
|
|
use arc_swap::ArcSwapOption;
|
2020-04-17 17:16:08 +00:00
|
|
|
use futures::future::*;
|
2020-04-17 13:36:16 +00:00
|
|
|
use futures::stream::*;
|
2020-04-09 21:45:07 +00:00
|
|
|
use tokio::fs;
|
|
|
|
use tokio::prelude::*;
|
2020-04-17 13:36:16 +00:00
|
|
|
use tokio::sync::{watch, Mutex};
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-17 13:36:16 +00:00
|
|
|
use crate::data;
|
2020-04-10 20:01:48 +00:00
|
|
|
use crate::data::*;
|
2020-04-09 21:45:07 +00:00
|
|
|
use crate::error::Error;
|
2020-04-17 13:36:16 +00:00
|
|
|
use crate::membership::System;
|
2020-04-09 21:45:07 +00:00
|
|
|
use crate::proto::*;
|
2020-04-17 13:36:16 +00:00
|
|
|
use crate::rpc_client::*;
|
2020-04-17 15:09:57 +00:00
|
|
|
use crate::server::Garage;
|
2020-04-11 21:00:26 +00:00
|
|
|
|
2020-04-17 17:16:08 +00:00
|
|
|
const NEED_BLOCK_QUERY_TIMEOUT: Duration = Duration::from_secs(5);
|
2020-04-17 18:58:10 +00:00
|
|
|
const RESYNC_RETRY_TIMEOUT: Duration = Duration::from_secs(10);
|
2020-04-17 17:16:08 +00:00
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
pub struct BlockManager {
|
|
|
|
pub data_dir: PathBuf,
|
|
|
|
pub rc: sled::Tree,
|
2020-04-17 13:36:16 +00:00
|
|
|
pub resync_queue: sled::Tree,
|
2020-04-11 21:00:26 +00:00
|
|
|
pub lock: Mutex<()>,
|
2020-04-17 13:36:16 +00:00
|
|
|
pub system: Arc<System>,
|
2020-04-17 15:09:57 +00:00
|
|
|
pub garage: ArcSwapOption<Garage>,
|
2020-04-09 21:45:07 +00:00
|
|
|
}
|
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
impl BlockManager {
|
2020-04-17 15:09:57 +00:00
|
|
|
pub fn new(db: &sled::Db, data_dir: PathBuf, system: Arc<System>) -> Arc<Self> {
|
2020-04-11 21:53:32 +00:00
|
|
|
let rc = db
|
|
|
|
.open_tree("block_local_rc")
|
2020-04-11 21:00:26 +00:00
|
|
|
.expect("Unable to open block_local_rc tree");
|
|
|
|
rc.set_merge_operator(rc_merge);
|
2020-04-17 13:36:16 +00:00
|
|
|
|
|
|
|
let resync_queue = db
|
|
|
|
.open_tree("block_local_resync_queue")
|
|
|
|
.expect("Unable to open block_local_resync_queue tree");
|
|
|
|
|
2020-04-17 15:09:57 +00:00
|
|
|
Arc::new(Self {
|
2020-04-11 21:00:26 +00:00
|
|
|
rc,
|
2020-04-17 13:36:16 +00:00
|
|
|
resync_queue,
|
2020-04-11 21:00:26 +00:00
|
|
|
data_dir,
|
|
|
|
lock: Mutex::new(()),
|
2020-04-17 13:36:16 +00:00
|
|
|
system,
|
2020-04-17 15:09:57 +00:00
|
|
|
garage: ArcSwapOption::from(None),
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
pub async fn spawn_background_worker(self: Arc<Self>) {
|
2020-04-17 18:58:10 +00:00
|
|
|
// Launch 2 simultaneous workers for background resync loop preprocessing
|
|
|
|
for _i in 0..2usize {
|
|
|
|
let bm2 = self.clone();
|
|
|
|
self.system
|
|
|
|
.background
|
|
|
|
.spawn_worker(move |must_exit| bm2.resync_loop(must_exit))
|
|
|
|
.await;
|
|
|
|
}
|
2020-04-11 21:00:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pub async fn write_block(&self, hash: &Hash, data: &[u8]) -> Result<Message, Error> {
|
|
|
|
let _lock = self.lock.lock().await;
|
|
|
|
|
|
|
|
let mut path = self.block_dir(hash);
|
|
|
|
fs::create_dir_all(&path).await?;
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
path.push(hex::encode(hash));
|
|
|
|
if fs::metadata(&path).await.is_ok() {
|
|
|
|
return Ok(Message::Ok);
|
|
|
|
}
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
let mut f = fs::File::create(path).await?;
|
|
|
|
f.write_all(data).await?;
|
|
|
|
drop(f);
|
|
|
|
|
|
|
|
Ok(Message::Ok)
|
2020-04-09 21:45:07 +00:00
|
|
|
}
|
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
pub async fn read_block(&self, hash: &Hash) -> Result<Message, Error> {
|
2020-04-17 17:20:17 +00:00
|
|
|
let path = self.block_path(hash);
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-17 16:38:11 +00:00
|
|
|
let mut f = match fs::File::open(&path).await {
|
|
|
|
Ok(f) => f,
|
|
|
|
Err(e) => {
|
|
|
|
// Not found but maybe we should have had it ??
|
|
|
|
self.put_to_resync(hash, DEFAULT_TIMEOUT.as_millis() as u64)?;
|
|
|
|
return Err(Into::into(e));
|
|
|
|
}
|
|
|
|
};
|
2020-04-11 21:00:26 +00:00
|
|
|
let mut data = vec![];
|
|
|
|
f.read_to_end(&mut data).await?;
|
2020-04-17 13:36:16 +00:00
|
|
|
drop(f);
|
|
|
|
|
|
|
|
if data::hash(&data[..]) != *hash {
|
|
|
|
let _lock = self.lock.lock().await;
|
|
|
|
eprintln!("Block {:?} is corrupted. Deleting and resyncing.", hash);
|
|
|
|
fs::remove_file(path).await?;
|
2020-04-17 15:09:57 +00:00
|
|
|
self.put_to_resync(&hash, 0)?;
|
2020-04-17 13:36:16 +00:00
|
|
|
return Err(Error::CorruptData(hash.clone()));
|
|
|
|
}
|
2020-04-11 21:00:26 +00:00
|
|
|
|
|
|
|
Ok(Message::PutBlock(PutBlockMessage {
|
|
|
|
hash: hash.clone(),
|
|
|
|
data,
|
|
|
|
}))
|
|
|
|
}
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-17 17:16:08 +00:00
|
|
|
pub async fn need_block(&self, hash: &Hash) -> Result<bool, Error> {
|
|
|
|
let needed = self
|
|
|
|
.rc
|
|
|
|
.get(hash.as_ref())?
|
|
|
|
.map(|x| u64_from_bytes(x.as_ref()) > 0)
|
|
|
|
.unwrap_or(false);
|
|
|
|
if needed {
|
2020-04-17 17:20:17 +00:00
|
|
|
let path = self.block_path(hash);
|
2020-04-17 17:16:08 +00:00
|
|
|
let exists = fs::metadata(&path).await.is_ok();
|
|
|
|
Ok(!exists)
|
|
|
|
} else {
|
|
|
|
Ok(false)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
fn block_dir(&self, hash: &Hash) -> PathBuf {
|
|
|
|
let mut path = self.data_dir.clone();
|
|
|
|
path.push(hex::encode(&hash.as_slice()[0..1]));
|
|
|
|
path.push(hex::encode(&hash.as_slice()[1..2]));
|
|
|
|
path
|
|
|
|
}
|
2020-04-17 17:20:17 +00:00
|
|
|
fn block_path(&self, hash: &Hash) -> PathBuf {
|
|
|
|
let mut path = self.block_dir(hash);
|
|
|
|
path.push(hex::encode(hash.as_ref()));
|
|
|
|
path
|
|
|
|
}
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-11 21:00:26 +00:00
|
|
|
pub fn block_incref(&self, hash: &Hash) -> Result<(), Error> {
|
2020-04-17 15:09:57 +00:00
|
|
|
let new_rc = self.rc.merge(&hash, vec![1])?;
|
|
|
|
if new_rc.map(|x| u64_from_bytes(&x[..]) == 0).unwrap_or(true) {
|
|
|
|
self.put_to_resync(&hash, BLOCK_RW_TIMEOUT.as_millis() as u64)?;
|
|
|
|
}
|
2020-04-11 21:00:26 +00:00
|
|
|
Ok(())
|
|
|
|
}
|
2020-04-09 21:45:07 +00:00
|
|
|
|
2020-04-17 13:36:16 +00:00
|
|
|
pub fn block_decref(&self, hash: &Hash) -> Result<(), Error> {
|
2020-04-17 15:09:57 +00:00
|
|
|
let new_rc = self.rc.merge(&hash, vec![0])?;
|
|
|
|
if new_rc.is_none() {
|
|
|
|
self.put_to_resync(&hash, 2 * BLOCK_RW_TIMEOUT.as_millis() as u64)?;
|
2020-04-17 13:36:16 +00:00
|
|
|
}
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2020-04-17 15:09:57 +00:00
|
|
|
fn put_to_resync(&self, hash: &Hash, delay_millis: u64) -> Result<(), Error> {
|
|
|
|
let when = now_msec() + delay_millis;
|
|
|
|
eprintln!("Put resync_queue: {} {:?}", when, hash);
|
|
|
|
let mut key = u64::to_be_bytes(when).to_vec();
|
|
|
|
key.extend(hash.as_ref());
|
|
|
|
self.resync_queue.insert(key, hash.as_ref())?;
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2020-04-17 13:36:16 +00:00
|
|
|
async fn resync_loop(self: Arc<Self>, must_exit: watch::Receiver<bool>) -> Result<(), Error> {
|
|
|
|
while !*must_exit.borrow() {
|
2020-04-17 18:58:10 +00:00
|
|
|
if let Some((time_bytes, hash_bytes)) = self.resync_queue.pop_min()? {
|
2020-04-17 15:09:57 +00:00
|
|
|
let time_msec = u64_from_bytes(&time_bytes[0..8]);
|
2020-04-17 16:51:29 +00:00
|
|
|
eprintln!(
|
|
|
|
"First in resync queue: {} (now = {})",
|
|
|
|
time_msec,
|
|
|
|
now_msec()
|
|
|
|
);
|
2020-04-17 15:09:57 +00:00
|
|
|
if now_msec() >= time_msec {
|
|
|
|
let mut hash = [0u8; 32];
|
|
|
|
hash.copy_from_slice(hash_bytes.as_ref());
|
|
|
|
let hash = Hash::from(hash);
|
|
|
|
|
2020-04-17 18:58:10 +00:00
|
|
|
if let Err(e) = self.resync_iter(&hash).await {
|
|
|
|
eprintln!("Failed to resync block {:?}, retrying later: {}", hash, e);
|
|
|
|
self.put_to_resync(&hash, RESYNC_RETRY_TIMEOUT.as_millis() as u64)?;
|
2020-04-17 13:36:16 +00:00
|
|
|
}
|
2020-04-17 15:09:57 +00:00
|
|
|
continue;
|
2020-04-17 18:58:10 +00:00
|
|
|
} else {
|
|
|
|
self.resync_queue.insert(time_bytes, hash_bytes)?;
|
2020-04-17 13:36:16 +00:00
|
|
|
}
|
2020-04-11 21:00:26 +00:00
|
|
|
}
|
2020-04-17 15:09:57 +00:00
|
|
|
tokio::time::delay_for(Duration::from_secs(1)).await;
|
2020-04-11 21:00:26 +00:00
|
|
|
}
|
2020-04-17 13:36:16 +00:00
|
|
|
Ok(())
|
2020-04-11 21:00:26 +00:00
|
|
|
}
|
2020-04-17 13:36:16 +00:00
|
|
|
|
|
|
|
async fn resync_iter(&self, hash: &Hash) -> Result<(), Error> {
|
2020-04-17 17:20:17 +00:00
|
|
|
let path = self.block_path(hash);
|
2020-04-17 13:36:16 +00:00
|
|
|
|
|
|
|
let exists = fs::metadata(&path).await.is_ok();
|
|
|
|
let needed = self
|
|
|
|
.rc
|
|
|
|
.get(hash.as_ref())?
|
|
|
|
.map(|x| u64_from_bytes(x.as_ref()) > 0)
|
|
|
|
.unwrap_or(false);
|
|
|
|
|
2020-04-17 16:51:29 +00:00
|
|
|
eprintln!(
|
|
|
|
"Resync block {:?}: exists {}, needed {}",
|
|
|
|
hash, exists, needed
|
|
|
|
);
|
2020-04-17 15:09:57 +00:00
|
|
|
|
2020-04-17 13:36:16 +00:00
|
|
|
if exists && !needed {
|
2020-04-17 15:09:57 +00:00
|
|
|
let garage = self.garage.load_full().unwrap();
|
2020-04-17 16:51:29 +00:00
|
|
|
let active_refs = garage
|
|
|
|
.block_ref_table
|
|
|
|
.get_range(&hash, &[0u8; 32].into(), Some(()), 1)
|
|
|
|
.await?;
|
2020-04-17 15:09:57 +00:00
|
|
|
let needed_by_others = !active_refs.is_empty();
|
|
|
|
if needed_by_others {
|
2020-04-17 17:16:08 +00:00
|
|
|
let ring = garage.system.ring.borrow().clone();
|
|
|
|
let who = ring.walk_ring(&hash, garage.system.config.data_replication_factor);
|
|
|
|
let msg = Message::NeedBlockQuery(hash.clone());
|
|
|
|
let who_needs_fut = who
|
|
|
|
.iter()
|
|
|
|
.map(|to| rpc_call(garage.system.clone(), to, &msg, NEED_BLOCK_QUERY_TIMEOUT));
|
|
|
|
let who_needs = join_all(who_needs_fut).await;
|
|
|
|
|
|
|
|
let mut need_nodes = vec![];
|
|
|
|
let mut errors = 0;
|
|
|
|
for (node, needed) in who.into_iter().zip(who_needs.iter()) {
|
|
|
|
match needed {
|
|
|
|
Ok(Message::NeedBlockReply(true)) => {
|
|
|
|
need_nodes.push(node);
|
|
|
|
}
|
|
|
|
Err(_) => {
|
|
|
|
errors += 1;
|
|
|
|
}
|
|
|
|
_ => (),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if errors > (garage.system.config.data_replication_factor - 1) / 2 {
|
|
|
|
return Err(Error::Message(format!(
|
|
|
|
"Should delete block, but not enough nodes confirm that they have it."
|
|
|
|
)));
|
|
|
|
}
|
|
|
|
|
|
|
|
if need_nodes.len() > 0 {
|
|
|
|
let put_block_message = self.read_block(hash).await?;
|
|
|
|
for resp in rpc_call_many(
|
|
|
|
garage.system.clone(),
|
|
|
|
&need_nodes[..],
|
|
|
|
put_block_message,
|
|
|
|
BLOCK_RW_TIMEOUT,
|
|
|
|
)
|
|
|
|
.await
|
|
|
|
{
|
|
|
|
resp?;
|
|
|
|
}
|
|
|
|
}
|
2020-04-17 15:09:57 +00:00
|
|
|
}
|
2020-04-17 13:36:16 +00:00
|
|
|
fs::remove_file(path).await?;
|
|
|
|
self.resync_queue.remove(&hash)?;
|
|
|
|
}
|
|
|
|
|
|
|
|
if needed && !exists {
|
|
|
|
// TODO find a way to not do this if they are sending it to us
|
2020-04-17 15:09:57 +00:00
|
|
|
// Let's suppose this isn't an issue for now with the BLOCK_RW_TIMEOUT delay
|
|
|
|
// between the RC being incremented and this part being called.
|
2020-04-17 13:36:16 +00:00
|
|
|
let block_data = rpc_get_block(&self.system, &hash).await?;
|
|
|
|
self.write_block(hash, &block_data[..]).await?;
|
|
|
|
}
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn u64_from_bytes(bytes: &[u8]) -> u64 {
|
|
|
|
assert!(bytes.len() == 8);
|
|
|
|
let mut x8 = [0u8; 8];
|
|
|
|
x8.copy_from_slice(bytes);
|
|
|
|
u64::from_be_bytes(x8)
|
2020-04-11 21:00:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
fn rc_merge(_key: &[u8], old: Option<&[u8]>, new: &[u8]) -> Option<Vec<u8>> {
|
2020-04-17 13:36:16 +00:00
|
|
|
let old = old.map(u64_from_bytes).unwrap_or(0);
|
2020-04-11 21:00:26 +00:00
|
|
|
assert!(new.len() == 1);
|
|
|
|
let new = match new[0] {
|
|
|
|
0 => {
|
|
|
|
if old > 0 {
|
|
|
|
old - 1
|
|
|
|
} else {
|
|
|
|
0
|
|
|
|
}
|
|
|
|
}
|
|
|
|
1 => old + 1,
|
|
|
|
_ => unreachable!(),
|
|
|
|
};
|
|
|
|
if new == 0 {
|
|
|
|
None
|
|
|
|
} else {
|
|
|
|
Some(u64::to_be_bytes(new).to_vec())
|
|
|
|
}
|
2020-04-09 21:45:07 +00:00
|
|
|
}
|
2020-04-17 13:36:16 +00:00
|
|
|
|
|
|
|
pub async fn rpc_get_block(system: &Arc<System>, hash: &Hash) -> Result<Vec<u8>, Error> {
|
2020-04-17 15:09:57 +00:00
|
|
|
let ring = system.ring.borrow().clone();
|
|
|
|
let who = ring.walk_ring(&hash, system.config.data_replication_factor);
|
2020-04-17 13:36:16 +00:00
|
|
|
let msg = Message::GetBlock(hash.clone());
|
|
|
|
let mut resp_stream = who
|
|
|
|
.iter()
|
|
|
|
.map(|to| rpc_call(system.clone(), to, &msg, BLOCK_RW_TIMEOUT))
|
|
|
|
.collect::<FuturesUnordered<_>>();
|
|
|
|
|
|
|
|
while let Some(resp) = resp_stream.next().await {
|
|
|
|
if let Ok(Message::PutBlock(msg)) = resp {
|
|
|
|
if data::hash(&msg.data[..]) == *hash {
|
|
|
|
return Ok(msg.data);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Err(Error::Message(format!(
|
|
|
|
"Unable to read block {:?}: no valid blocks returned",
|
|
|
|
hash
|
|
|
|
)))
|
|
|
|
}
|
|
|
|
|
|
|
|
pub async fn rpc_put_block(system: &Arc<System>, hash: Hash, data: Vec<u8>) -> Result<(), Error> {
|
2020-04-17 15:09:57 +00:00
|
|
|
let ring = system.ring.borrow().clone();
|
|
|
|
let who = ring.walk_ring(&hash, system.config.data_replication_factor);
|
2020-04-17 13:36:16 +00:00
|
|
|
rpc_try_call_many(
|
|
|
|
system.clone(),
|
|
|
|
&who[..],
|
|
|
|
Message::PutBlock(PutBlockMessage { hash, data }),
|
|
|
|
(system.config.data_replication_factor + 1) / 2,
|
|
|
|
BLOCK_RW_TIMEOUT,
|
|
|
|
)
|
|
|
|
.await?;
|
|
|
|
Ok(())
|
|
|
|
}
|