forked from Deuxfleurs/garage
Compare commits
1 commit
main
...
feat/lmdb-
Author | SHA1 | Date | |
---|---|---|---|
45425a5637 |
8 changed files with 62 additions and 40 deletions
|
@ -23,7 +23,7 @@ client = minio.Minio(
|
||||||
"GKyourapikey",
|
"GKyourapikey",
|
||||||
"abcd[...]1234",
|
"abcd[...]1234",
|
||||||
# Force the region, this is specific to garage
|
# Force the region, this is specific to garage
|
||||||
region="garage",
|
region="region",
|
||||||
)
|
)
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
|
@ -335,7 +335,6 @@ From the [official Mastodon documentation](https://docs.joinmastodon.org/admin/t
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ RAILS_ENV=production bin/tootctl media remove --days 3
|
$ RAILS_ENV=production bin/tootctl media remove --days 3
|
||||||
$ RAILS_ENV=production bin/tootctl media remove --days 15 --prune-profiles
|
|
||||||
$ RAILS_ENV=production bin/tootctl media remove-orphans
|
$ RAILS_ENV=production bin/tootctl media remove-orphans
|
||||||
$ RAILS_ENV=production bin/tootctl preview_cards remove --days 15
|
$ RAILS_ENV=production bin/tootctl preview_cards remove --days 15
|
||||||
```
|
```
|
||||||
|
@ -354,6 +353,8 @@ Imports: 1.7 KB
|
||||||
Settings: 0 Bytes
|
Settings: 0 Bytes
|
||||||
```
|
```
|
||||||
|
|
||||||
|
Unfortunately, [old avatars and headers cannot currently be cleaned up](https://github.com/mastodon/mastodon/issues/9567).
|
||||||
|
|
||||||
### Migrating your data
|
### Migrating your data
|
||||||
|
|
||||||
Data migration should be done with an efficient S3 client.
|
Data migration should be done with an efficient S3 client.
|
||||||
|
|
|
@ -50,20 +50,3 @@ locations. They use Garage themselves for the following tasks:
|
||||||
|
|
||||||
The Deuxfleurs Garage cluster is a multi-site cluster currently composed of
|
The Deuxfleurs Garage cluster is a multi-site cluster currently composed of
|
||||||
9 nodes in 3 physical locations.
|
9 nodes in 3 physical locations.
|
||||||
|
|
||||||
### Triplebit
|
|
||||||
|
|
||||||
[Triplebit](https://www.triplebit.org) is a non-profit hosting provider and
|
|
||||||
ISP focused on improving access to privacy-related services. They use
|
|
||||||
Garage themselves for the following tasks:
|
|
||||||
|
|
||||||
- Hosting of their homepage, [privacyguides.org](https://www.privacyguides.org/), and various other static sites
|
|
||||||
|
|
||||||
- As a Mastodon object storage backend for [mstdn.party](https://mstdn.party/) and [mstdn.plus](https://mstdn.plus/)
|
|
||||||
|
|
||||||
- As a PeerTube storage backend for [neat.tube](https://neat.tube/)
|
|
||||||
|
|
||||||
- As a [Matrix media backend](https://github.com/matrix-org/synapse-s3-storage-provider)
|
|
||||||
|
|
||||||
Triplebit's Garage cluster is a multi-site cluster currently composed of
|
|
||||||
10 nodes in 3 physical locations.
|
|
||||||
|
|
|
@ -76,9 +76,6 @@ spec:
|
||||||
- name: etc
|
- name: etc
|
||||||
mountPath: /etc/garage.toml
|
mountPath: /etc/garage.toml
|
||||||
subPath: garage.toml
|
subPath: garage.toml
|
||||||
{{- with .Values.extraVolumeMounts }}
|
|
||||||
{{- toYaml . | nindent 12 }}
|
|
||||||
{{- end }}
|
|
||||||
# TODO
|
# TODO
|
||||||
# livenessProbe:
|
# livenessProbe:
|
||||||
# httpGet:
|
# httpGet:
|
||||||
|
@ -113,9 +110,6 @@ spec:
|
||||||
- name: data
|
- name: data
|
||||||
emptyDir: {}
|
emptyDir: {}
|
||||||
{{- end }}
|
{{- end }}
|
||||||
{{- with .Values.extraVolumes }}
|
|
||||||
{{- toYaml . | nindent 8 }}
|
|
||||||
{{- end }}
|
|
||||||
{{- with .Values.nodeSelector }}
|
{{- with .Values.nodeSelector }}
|
||||||
nodeSelector:
|
nodeSelector:
|
||||||
{{- toYaml . | nindent 8 }}
|
{{- toYaml . | nindent 8 }}
|
||||||
|
|
|
@ -218,10 +218,6 @@ affinity: {}
|
||||||
|
|
||||||
environment: {}
|
environment: {}
|
||||||
|
|
||||||
extraVolumes: {}
|
|
||||||
|
|
||||||
extraVolumeMounts: {}
|
|
||||||
|
|
||||||
monitoring:
|
monitoring:
|
||||||
metrics:
|
metrics:
|
||||||
# If true, a service for monitoring is created with a prometheus.io/scrape annotation
|
# If true, a service for monitoring is created with a prometheus.io/scrape annotation
|
||||||
|
|
|
@ -15,6 +15,9 @@ path = "lib.rs"
|
||||||
err-derive.workspace = true
|
err-derive.workspace = true
|
||||||
hexdump.workspace = true
|
hexdump.workspace = true
|
||||||
tracing.workspace = true
|
tracing.workspace = true
|
||||||
|
opentelemetry.workspace = true
|
||||||
|
opentelemetry.workspace = true
|
||||||
|
xxhash-rust.workspace = true
|
||||||
|
|
||||||
heed = { workspace = true, optional = true }
|
heed = { workspace = true, optional = true }
|
||||||
rusqlite = { workspace = true, optional = true, features = ["backup"] }
|
rusqlite = { workspace = true, optional = true, features = ["backup"] }
|
||||||
|
|
|
@ -10,6 +10,8 @@ use std::sync::{Arc, RwLock};
|
||||||
use heed::types::ByteSlice;
|
use heed::types::ByteSlice;
|
||||||
use heed::{BytesDecode, Env, RoTxn, RwTxn, UntypedDatabase as Database};
|
use heed::{BytesDecode, Env, RoTxn, RwTxn, UntypedDatabase as Database};
|
||||||
|
|
||||||
|
use xxhash_rust::xxh3::xxh3_128;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
Db, Error, IDb, ITx, ITxFn, OnCommit, Result, TxError, TxFnResult, TxOpError, TxOpResult,
|
Db, Error, IDb, ITx, ITxFn, OnCommit, Result, TxError, TxFnResult, TxOpError, TxOpResult,
|
||||||
TxResult, TxValueIter, Value, ValueIter,
|
TxResult, TxValueIter, Value, ValueIter,
|
||||||
|
@ -58,6 +60,40 @@ impl LmdbDb {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn key_hash(key: &[u8]) -> [u8; 16] {
|
||||||
|
xxh3_128(key).to_ne_bytes()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn kv_to_value(key: &[u8], value: &[u8]) -> Vec<u8> {
|
||||||
|
[&key.len().to_ne_bytes(), key, value].concat()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn value_to_kv(value: &[u8]) -> (Vec<u8>, Vec<u8>) {
|
||||||
|
const USIZE_LEN: usize = std::mem::size_of::<usize>();
|
||||||
|
let klen = usize::from_ne_bytes(value[0..USIZE_LEN].try_into().unwrap());
|
||||||
|
(
|
||||||
|
value[USIZE_LEN..klen+USIZE_LEN].to_vec(),
|
||||||
|
value[USIZE_LEN+klen..].to_vec()
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn key_hash(key: &[u8]) -> [u8; 16] {
|
||||||
|
xxh3_128(key).to_ne_bytes()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn kv_to_value(key: &[u8], value: &[u8]) -> Vec<u8> {
|
||||||
|
[&key.len().to_ne_bytes(), key, value].concat()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn value_to_kv(value: &[u8]) -> (Vec<u8>, Vec<u8>) {
|
||||||
|
const USIZE_LEN: usize = std::mem::size_of::<usize>();
|
||||||
|
let klen = usize::from_ne_bytes(value[0..USIZE_LEN].try_into().unwrap());
|
||||||
|
(
|
||||||
|
value[USIZE_LEN..klen+USIZE_LEN].to_vec(),
|
||||||
|
value[USIZE_LEN+klen..].to_vec()
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
impl IDb for LmdbDb {
|
impl IDb for LmdbDb {
|
||||||
fn engine(&self) -> String {
|
fn engine(&self) -> String {
|
||||||
"LMDB (using Heed crate)".into()
|
"LMDB (using Heed crate)".into()
|
||||||
|
@ -119,10 +155,11 @@ impl IDb for LmdbDb {
|
||||||
let tree = self.get_tree(tree)?;
|
let tree = self.get_tree(tree)?;
|
||||||
|
|
||||||
let tx = self.db.read_txn()?;
|
let tx = self.db.read_txn()?;
|
||||||
let val = tree.get(&tx, key)?;
|
let kh = key_hash(key);
|
||||||
|
let val = tree.get(&tx, &kh)?;
|
||||||
match val {
|
match val {
|
||||||
None => Ok(None),
|
None => Ok(None),
|
||||||
Some(v) => Ok(Some(v.to_vec())),
|
Some(v) => Ok(Some(value_to_kv(v).1))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -135,7 +172,9 @@ impl IDb for LmdbDb {
|
||||||
fn insert(&self, tree: usize, key: &[u8], value: &[u8]) -> Result<()> {
|
fn insert(&self, tree: usize, key: &[u8], value: &[u8]) -> Result<()> {
|
||||||
let tree = self.get_tree(tree)?;
|
let tree = self.get_tree(tree)?;
|
||||||
let mut tx = self.db.write_txn()?;
|
let mut tx = self.db.write_txn()?;
|
||||||
tree.put(&mut tx, key, value)?;
|
let kh = key_hash(key);
|
||||||
|
let value = kv_to_value(key, value);
|
||||||
|
tree.put(&mut tx, &kh, &value)?;
|
||||||
tx.commit()?;
|
tx.commit()?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
@ -143,7 +182,8 @@ impl IDb for LmdbDb {
|
||||||
fn remove(&self, tree: usize, key: &[u8]) -> Result<()> {
|
fn remove(&self, tree: usize, key: &[u8]) -> Result<()> {
|
||||||
let tree = self.get_tree(tree)?;
|
let tree = self.get_tree(tree)?;
|
||||||
let mut tx = self.db.write_txn()?;
|
let mut tx = self.db.write_txn()?;
|
||||||
tree.delete(&mut tx, key)?;
|
let kh = key_hash(key);
|
||||||
|
tree.delete(&mut tx, &kh)?;
|
||||||
tx.commit()?;
|
tx.commit()?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
@ -242,8 +282,9 @@ impl<'a> LmdbTx<'a> {
|
||||||
impl<'a> ITx for LmdbTx<'a> {
|
impl<'a> ITx for LmdbTx<'a> {
|
||||||
fn get(&self, tree: usize, key: &[u8]) -> TxOpResult<Option<Value>> {
|
fn get(&self, tree: usize, key: &[u8]) -> TxOpResult<Option<Value>> {
|
||||||
let tree = self.get_tree(tree)?;
|
let tree = self.get_tree(tree)?;
|
||||||
match tree.get(&self.tx, key)? {
|
let kh = key_hash(key);
|
||||||
Some(v) => Ok(Some(v.to_vec())),
|
match tree.get(&self.tx, &kh)? {
|
||||||
|
Some(v) => Ok(Some(value_to_kv(v).1)),
|
||||||
None => Ok(None),
|
None => Ok(None),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -254,14 +295,18 @@ impl<'a> ITx for LmdbTx<'a> {
|
||||||
|
|
||||||
fn insert(&mut self, tree: usize, key: &[u8], value: &[u8]) -> TxOpResult<()> {
|
fn insert(&mut self, tree: usize, key: &[u8], value: &[u8]) -> TxOpResult<()> {
|
||||||
let tree = *self.get_tree(tree)?;
|
let tree = *self.get_tree(tree)?;
|
||||||
tree.put(&mut self.tx, key, value)?;
|
let kh = key_hash(key);
|
||||||
|
let value = kv_to_value(key, value);
|
||||||
|
tree.put(&mut self.tx, &kh, &value)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
fn remove(&mut self, tree: usize, key: &[u8]) -> TxOpResult<()> {
|
fn remove(&mut self, tree: usize, key: &[u8]) -> TxOpResult<()> {
|
||||||
let tree = *self.get_tree(tree)?;
|
let tree = *self.get_tree(tree)?;
|
||||||
tree.delete(&mut self.tx, key)?;
|
let kh = key_hash(key);
|
||||||
|
tree.delete(&mut self.tx, &kh)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn clear(&mut self, tree: usize) -> TxOpResult<()> {
|
fn clear(&mut self, tree: usize) -> TxOpResult<()> {
|
||||||
let tree = *self.get_tree(tree)?;
|
let tree = *self.get_tree(tree)?;
|
||||||
tree.clear(&mut self.tx)?;
|
tree.clear(&mut self.tx)?;
|
||||||
|
@ -370,7 +415,7 @@ where
|
||||||
match next {
|
match next {
|
||||||
None => None,
|
None => None,
|
||||||
Some(Err(e)) => Some(Err(e.into())),
|
Some(Err(e)) => Some(Err(e.into())),
|
||||||
Some(Ok((k, v))) => Some(Ok((k.to_vec(), v.to_vec()))),
|
Some(Ok((_k, v))) => Some(Ok(value_to_kv(v))),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -380,7 +425,7 @@ where
|
||||||
fn tx_iter_item<'a>(
|
fn tx_iter_item<'a>(
|
||||||
item: std::result::Result<(&'a [u8], &'a [u8]), heed::Error>,
|
item: std::result::Result<(&'a [u8], &'a [u8]), heed::Error>,
|
||||||
) -> TxOpResult<(Vec<u8>, Vec<u8>)> {
|
) -> TxOpResult<(Vec<u8>, Vec<u8>)> {
|
||||||
item.map(|(k, v)| (k.to_vec(), v.to_vec()))
|
item.map(|(_k, v)| value_to_kv(v))
|
||||||
.map_err(|e| TxOpError(Error::from(e)))
|
.map_err(|e| TxOpError(Error::from(e)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -141,7 +141,7 @@ impl Garage {
|
||||||
)?)
|
)?)
|
||||||
.ok()
|
.ok()
|
||||||
.and_then(|x| NetworkKey::from_slice(&x))
|
.and_then(|x| NetworkKey::from_slice(&x))
|
||||||
.ok_or_message("Invalid RPC secret key: expected 32 bytes of random hex, please check the documentation for requirements")?;
|
.ok_or_message("Invalid RPC secret key: expected 32 bits of entropy, please check the documentation for requirements")?;
|
||||||
|
|
||||||
let (replication_factor, consistency_mode) = parse_replication_mode(&config)?;
|
let (replication_factor, consistency_mode) = parse_replication_mode(&config)?;
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue