WIP implem storage

This commit is contained in:
Quentin 2024-04-04 11:28:15 +02:00
parent 9afbfeb427
commit f179479c30
Signed by: quentin
GPG key ID: E9602264D639FF68
2 changed files with 113 additions and 27 deletions

View file

@ -9,7 +9,7 @@ use aero_user::cryptoblob::{self, gen_key, open_deserialize, seal_serialize, Key
use aero_user::storage::{self, BlobRef, BlobVal, RowRef, RowVal, Selector, Store}; use aero_user::storage::{self, BlobRef, BlobVal, RowRef, RowVal, Selector, Store};
use crate::unique_ident::*; use crate::unique_ident::*;
use crate::davdag::DavDag; use crate::davdag::{DavDag, IndexEntry, Token, BlobId, SyncChange};
pub struct Calendar { pub struct Calendar {
pub(super) id: UniqueIdent, pub(super) id: UniqueIdent,
@ -21,7 +21,48 @@ impl Calendar {
creds: &Credentials, creds: &Credentials,
id: UniqueIdent, id: UniqueIdent,
) -> Result<Self> { ) -> Result<Self> {
todo!(); let bayou_path = format!("calendar/dag/{}", id);
let cal_path = format!("calendar/events/{}", id);
let mut davdag = Bayou::<DavDag>::new(creds, bayou_path).await?;
davdag.sync().await?;
let internal = RwLock::new(CalendarInternal {
id,
encryption_key: creds.keys.master.clone(),
storage: creds.storage.build().await?,
davdag,
cal_path,
});
Ok(Self { id, internal })
}
/// Sync data with backing store
pub async fn force_sync(&self) -> Result<()> {
self.internal.write().await.force_sync().await
}
/// Sync data with backing store only if changes are detected
/// or last sync is too old
pub async fn opportunistic_sync(&self) -> Result<()> {
self.internal.write().await.opportunistic_sync().await
}
pub async fn get(&self, blob_id: UniqueIdent, message_key: &Key) -> Result<Vec<u8>> {
self.internal.read().await.get(blob_id, message_key).await
}
pub async fn diff(&self, sync_token: Token) -> Result<(Token, Vec<SyncChange>)> {
self.internal.read().await.diff(sync_token).await
}
pub async fn put<'a>(&self, entry: IndexEntry, evt: &'a [u8]) -> Result<Token> {
self.internal.write().await.put(entry, evt).await
}
pub async fn delete(&self, blob_id: UniqueIdent) -> Result<Token> {
self.internal.write().await.delete(blob_id).await
} }
} }
@ -30,5 +71,35 @@ struct CalendarInternal {
cal_path: String, cal_path: String,
encryption_key: Key, encryption_key: Key,
storage: Store, storage: Store,
uid_index: Bayou<DavDag>, davdag: Bayou<DavDag>,
}
impl CalendarInternal {
async fn force_sync(&mut self) -> Result<()> {
self.davdag.sync().await?;
Ok(())
}
async fn opportunistic_sync(&mut self) -> Result<()> {
self.davdag.opportunistic_sync().await?;
Ok(())
}
async fn get(&self, blob_id: BlobId, message_key: &Key) -> Result<Vec<u8>> {
todo!()
}
async fn put<'a>(&mut self, entry: IndexEntry, evt: &'a [u8]) -> Result<Token> {
//@TODO write event to S3
//@TODO add entry into Bayou
todo!();
}
async fn delete(&mut self, blob_id: BlobId) -> Result<Token> {
todo!();
}
async fn diff(&self, sync_token: Token) -> Result<(Token, Vec<SyncChange>)> {
todo!();
}
} }

View file

@ -20,20 +20,31 @@ pub type IndexEntry = (BlobId, FileName, Etag);
#[derive(Clone, Default)] #[derive(Clone, Default)]
pub struct DavDag { pub struct DavDag {
/// Source of trust /// Source of trust
pub table: OrdMap<UniqueIdent, IndexEntry>, pub table: OrdMap<BlobId, IndexEntry>,
/// Indexes optimized for queries /// Indexes optimized for queries
pub idx_by_filename: OrdMap<FileName, UniqueIdent>, pub idx_by_filename: OrdMap<FileName, BlobId>,
// ------------ Below this line, data is ephemeral, ie. not checkpointed
/// Partial synchronization graph /// Partial synchronization graph
pub ancestors: OrdMap<UniqueIdent, OrdSet<UniqueIdent>>, pub ancestors: OrdMap<Token, OrdSet<Token>>,
/// All nodes /// All nodes
pub all_nodes: OrdSet<UniqueIdent>, pub all_nodes: OrdSet<Token>,
/// Head nodes /// Head nodes
pub heads: OrdSet<UniqueIdent>, pub heads: OrdSet<Token>,
/// Origin nodes /// Origin nodes
pub origins: OrdSet<UniqueIdent>, pub origins: OrdSet<Token>,
/// File change token by token
pub change: OrdMap<Token, SyncChange>,
}
#[derive(Clone, Debug)]
pub enum SyncChange {
Ok(FileName),
NotFound(FileName),
} }
#[derive(Clone, Serialize, Deserialize, Debug)] #[derive(Clone, Serialize, Deserialize, Debug)]
@ -66,8 +77,8 @@ impl DavDag {
DavDagOp::Put(self.sync_desc(), entry) DavDagOp::Put(self.sync_desc(), entry)
} }
pub fn op_delete(&self, ident: BlobId) -> DavDagOp { pub fn op_delete(&self, blob_id: BlobId) -> DavDagOp {
DavDagOp::Delete(self.sync_desc(), ident) DavDagOp::Delete(self.sync_desc(), blob_id)
} }
// HELPER functions // HELPER functions
@ -129,33 +140,41 @@ impl DavDag {
// INTERNAL functions // INTERNAL functions
/// Register a WebDAV item (put, copy, move) /// Register a WebDAV item (put, copy, move)
fn register(&mut self, entry: IndexEntry) { fn register(&mut self, sync_token: Option<Token>, entry: IndexEntry) {
let (blob_id, filename, _etag) = entry.clone(); let (blob_id, filename, _etag) = entry.clone();
// Insert item in the source of trust // Insert item in the source of trust
self.table.insert(blob_id, entry); self.table.insert(blob_id, entry);
// Update the cache // Update the cache
self.idx_by_filename.insert(filename, blob_id); self.idx_by_filename.insert(filename.to_string(), blob_id);
// Record the change in the ephemeral synchronization map
if let Some(sync_token) = sync_token {
self.change.insert(sync_token, SyncChange::Ok(filename));
}
} }
/// Unregister a WebDAV item (delete, move) /// Unregister a WebDAV item (delete, move)
fn unregister(&mut self, ident: &UniqueIdent) { fn unregister(&mut self, sync_token: Token, blob_id: &BlobId) {
// Query the source of truth to get the information we // Query the source of truth to get the information we
// need to clean the indexes // need to clean the indexes
let (_blob_id, filename, _etag) = match self.table.get(ident) { let (_blob_id, filename, _etag) = match self.table.get(blob_id) {
Some(v) => v, Some(v) => v,
// Element does not exist, return early // Element does not exist, return early
None => return, None => return,
}; };
self.idx_by_filename.remove(filename); self.idx_by_filename.remove(filename);
// Record the change in the ephemeral synchronization map
self.change.insert(sync_token, SyncChange::NotFound(filename.to_string()));
// Finally clear item from the source of trust // Finally clear item from the source of trust
self.table.remove(ident); self.table.remove(blob_id);
} }
/// When an event is processed, update the synchronization DAG /// When an event is processed, update the synchronization DAG
fn sync_dag(&mut self, sync_desc: &SyncDesc) -> bool { fn sync_dag(&mut self, sync_desc: &SyncDesc) {
let (parents, child) = sync_desc; let (parents, child) = sync_desc;
// --- Update ANCESTORS // --- Update ANCESTORS
@ -180,8 +199,6 @@ impl DavDag {
// --- Update ALL NODES // --- Update ALL NODES
self.all_nodes.insert(*child); self.all_nodes.insert(*child);
true
} }
} }
@ -193,14 +210,12 @@ impl BayouState for DavDag {
match op { match op {
DavDagOp::Put(sync_desc, entry) => { DavDagOp::Put(sync_desc, entry) => {
if new.sync_dag(sync_desc) { new.sync_dag(sync_desc);
new.register(entry.clone()); new.register(Some(sync_desc.1), entry.clone());
}
}, },
DavDagOp::Delete(sync_desc, blob_id) => { DavDagOp::Delete(sync_desc, blob_id) => {
if new.sync_dag(sync_desc) { new.sync_dag(sync_desc);
new.unregister(blob_id); new.unregister(sync_desc.1, blob_id);
}
}, },
DavDagOp::Merge(sync_desc) => { DavDagOp::Merge(sync_desc) => {
new.sync_dag(sync_desc); new.sync_dag(sync_desc);
@ -227,7 +242,7 @@ impl<'de> Deserialize<'de> for DavDag {
let mut davdag = DavDag::default(); let mut davdag = DavDag::default();
// Build the table + index // Build the table + index
val.items.into_iter().for_each(|entry| davdag.register(entry)); val.items.into_iter().for_each(|entry| davdag.register(None, entry));
// Initialize the synchronization DAG with its roots // Initialize the synchronization DAG with its roots
val.heads.into_iter().for_each(|ident| { val.heads.into_iter().for_each(|ident| {