forked from Deuxfleurs/garage
Alex
b44d3fc796
- [x] Design interface - [x] Implement Sled backend - [x] Re-implement the SledCountedTree hack ~~on Sled backend~~ on all backends (i.e. over the abstraction) - [x] Convert Garage code to use generic interface - [x] Proof-read converted Garage code - [ ] Test everything well - [x] Implement sqlite backend - [x] Implement LMDB backend - [ ] (Implement Persy backend?) - [ ] (Implement other backends? (like RocksDB, ...)) - [x] Implement backend choice in config file and garage server module - [x] Add CLI for converting between DB formats - Exploit the new interface to put more things in transactions - [x] `.updated()` trigger on Garage tables Fix #284 **Bugs** - [x] When exporting sqlite, trees iterate empty?? - [x] LMDB doesn't work **Known issues for various back-ends** - Sled: - Eats all my RAM and also all my disk space - `.len()` has to traverse the whole table - Is actually quite slow on some operations - And is actually pretty bad code... - Sqlite: - Requires a lock to be taken on all operations. The lock is also taken when iterating on a table with `.iter()`, and the lock isn't released until the iterator is dropped. This means that we must be VERY carefull to not do anything else inside a `.iter()` loop or else we will have a deadlock! Most such cases have been eliminated from the Garage codebase, but there might still be some that remain. If your Garage-over-Sqlite seems to hang/freeze, this is the reason. - (adapter uses a bunch of unsafe code) - Heed (LMDB): - Not suited for 32-bit machines as it has to map the whole DB in memory. - (adpater uses a tiny bit of unsafe code) **My recommendation:** avoid 32-bit machines and use LMDB as much as possible. **Converting databases** is actually quite easy. For example from Sled to LMDB: ```bash cd src/db cargo run --features cli --bin convert -- -i path/to/garage/meta/db -a sled -o path/to/garage/meta/db.lmdb -b lmdb ``` Then, just add this to your `config.toml`: ```toml db_engine = "lmdb" ``` Co-authored-by: Alex Auvolat <alex@adnab.me> Reviewed-on: Deuxfleurs/garage#322 Co-authored-by: Alex <alex@adnab.me> Co-committed-by: Alex <alex@adnab.me>
216 lines
4.9 KiB
Rust
216 lines
4.9 KiB
Rust
use serde::{Deserialize, Serialize};
|
|
use std::sync::Arc;
|
|
|
|
use garage_db as db;
|
|
|
|
use garage_util::background::BackgroundRunner;
|
|
use garage_util::data::*;
|
|
|
|
use garage_table::crdt::*;
|
|
use garage_table::replication::TableShardedReplication;
|
|
use garage_table::*;
|
|
|
|
use crate::s3::block_ref_table::*;
|
|
|
|
use garage_model_050::version_table as old;
|
|
|
|
/// A version of an object
|
|
#[derive(PartialEq, Clone, Debug, Serialize, Deserialize)]
|
|
pub struct Version {
|
|
/// UUID of the version, used as partition key
|
|
pub uuid: Uuid,
|
|
|
|
// Actual data: the blocks for this version
|
|
// In the case of a multipart upload, also store the etags
|
|
// of individual parts and check them when doing CompleteMultipartUpload
|
|
/// Is this version deleted
|
|
pub deleted: crdt::Bool,
|
|
/// list of blocks of data composing the version
|
|
pub blocks: crdt::Map<VersionBlockKey, VersionBlock>,
|
|
/// Etag of each part in case of a multipart upload, empty otherwise
|
|
pub parts_etags: crdt::Map<u64, String>,
|
|
|
|
// Back link to bucket+key so that we can figure if
|
|
// this was deleted later on
|
|
/// Bucket in which the related object is stored
|
|
pub bucket_id: Uuid,
|
|
/// Key in which the related object is stored
|
|
pub key: String,
|
|
}
|
|
|
|
impl Version {
|
|
pub fn new(uuid: Uuid, bucket_id: Uuid, key: String, deleted: bool) -> Self {
|
|
Self {
|
|
uuid,
|
|
deleted: deleted.into(),
|
|
blocks: crdt::Map::new(),
|
|
parts_etags: crdt::Map::new(),
|
|
bucket_id,
|
|
key,
|
|
}
|
|
}
|
|
|
|
pub fn has_part_number(&self, part_number: u64) -> bool {
|
|
let case1 = self
|
|
.parts_etags
|
|
.items()
|
|
.binary_search_by(|(k, _)| k.cmp(&part_number))
|
|
.is_ok();
|
|
let case2 = self
|
|
.blocks
|
|
.items()
|
|
.binary_search_by(|(k, _)| k.part_number.cmp(&part_number))
|
|
.is_ok();
|
|
case1 || case2
|
|
}
|
|
}
|
|
|
|
#[derive(PartialEq, Eq, Clone, Copy, Debug, Serialize, Deserialize)]
|
|
pub struct VersionBlockKey {
|
|
/// Number of the part
|
|
pub part_number: u64,
|
|
/// Offset of this sub-segment in its part
|
|
pub offset: u64,
|
|
}
|
|
|
|
impl Ord for VersionBlockKey {
|
|
fn cmp(&self, other: &Self) -> std::cmp::Ordering {
|
|
self.part_number
|
|
.cmp(&other.part_number)
|
|
.then(self.offset.cmp(&other.offset))
|
|
}
|
|
}
|
|
|
|
impl PartialOrd for VersionBlockKey {
|
|
fn partial_cmp(&self, other: &Self) -> Option<std::cmp::Ordering> {
|
|
Some(self.cmp(other))
|
|
}
|
|
}
|
|
|
|
/// Informations about a single block
|
|
#[derive(PartialEq, Eq, Ord, PartialOrd, Clone, Copy, Debug, Serialize, Deserialize)]
|
|
pub struct VersionBlock {
|
|
/// Blake2 sum of the block
|
|
pub hash: Hash,
|
|
/// Size of the block
|
|
pub size: u64,
|
|
}
|
|
|
|
impl AutoCrdt for VersionBlock {
|
|
const WARN_IF_DIFFERENT: bool = true;
|
|
}
|
|
|
|
impl Entry<Uuid, EmptyKey> for Version {
|
|
fn partition_key(&self) -> &Uuid {
|
|
&self.uuid
|
|
}
|
|
fn sort_key(&self) -> &EmptyKey {
|
|
&EmptyKey
|
|
}
|
|
fn is_tombstone(&self) -> bool {
|
|
self.deleted.get()
|
|
}
|
|
}
|
|
|
|
impl Crdt for Version {
|
|
fn merge(&mut self, other: &Self) {
|
|
self.deleted.merge(&other.deleted);
|
|
|
|
if self.deleted.get() {
|
|
self.blocks.clear();
|
|
self.parts_etags.clear();
|
|
} else {
|
|
self.blocks.merge(&other.blocks);
|
|
self.parts_etags.merge(&other.parts_etags);
|
|
}
|
|
}
|
|
}
|
|
|
|
pub struct VersionTable {
|
|
pub background: Arc<BackgroundRunner>,
|
|
pub block_ref_table: Arc<Table<BlockRefTable, TableShardedReplication>>,
|
|
}
|
|
|
|
impl TableSchema for VersionTable {
|
|
const TABLE_NAME: &'static str = "version";
|
|
|
|
type P = Uuid;
|
|
type S = EmptyKey;
|
|
type E = Version;
|
|
type Filter = DeletedFilter;
|
|
|
|
fn updated(
|
|
&self,
|
|
_tx: &mut db::Transaction,
|
|
old: Option<&Self::E>,
|
|
new: Option<&Self::E>,
|
|
) -> db::TxOpResult<()> {
|
|
let block_ref_table = self.block_ref_table.clone();
|
|
let old = old.cloned();
|
|
let new = new.cloned();
|
|
|
|
self.background.spawn(async move {
|
|
if let (Some(old_v), Some(new_v)) = (old, new) {
|
|
// Propagate deletion of version blocks
|
|
if new_v.deleted.get() && !old_v.deleted.get() {
|
|
let deleted_block_refs = old_v
|
|
.blocks
|
|
.items()
|
|
.iter()
|
|
.map(|(_k, vb)| BlockRef {
|
|
block: vb.hash,
|
|
version: old_v.uuid,
|
|
deleted: true.into(),
|
|
})
|
|
.collect::<Vec<_>>();
|
|
block_ref_table.insert_many(&deleted_block_refs[..]).await?;
|
|
}
|
|
}
|
|
Ok(())
|
|
});
|
|
|
|
Ok(())
|
|
}
|
|
|
|
fn matches_filter(entry: &Self::E, filter: &Self::Filter) -> bool {
|
|
filter.apply(entry.deleted.get())
|
|
}
|
|
|
|
fn try_migrate(bytes: &[u8]) -> Option<Self::E> {
|
|
let old = rmp_serde::decode::from_read_ref::<_, old::Version>(bytes).ok()?;
|
|
|
|
let blocks = old
|
|
.blocks
|
|
.items()
|
|
.iter()
|
|
.map(|(k, v)| {
|
|
(
|
|
VersionBlockKey {
|
|
part_number: k.part_number,
|
|
offset: k.offset,
|
|
},
|
|
VersionBlock {
|
|
hash: Hash::try_from(v.hash.as_slice()).unwrap(),
|
|
size: v.size,
|
|
},
|
|
)
|
|
})
|
|
.collect::<crdt::Map<_, _>>();
|
|
|
|
let parts_etags = old
|
|
.parts_etags
|
|
.items()
|
|
.iter()
|
|
.map(|(k, v)| (*k, v.clone()))
|
|
.collect::<crdt::Map<_, _>>();
|
|
|
|
Some(Version {
|
|
uuid: Hash::try_from(old.uuid.as_slice()).unwrap(),
|
|
deleted: crdt::Bool::new(old.deleted.get()),
|
|
blocks,
|
|
parts_etags,
|
|
bucket_id: blake2sum(old.bucket.as_bytes()),
|
|
key: old.key,
|
|
})
|
|
}
|
|
}
|