mirror of
https://git.deuxfleurs.fr/Deuxfleurs/garage.git
synced 2024-11-29 03:21:00 +00:00
block manager: use data paths from layout
This commit is contained in:
parent
6c420c0880
commit
887b3233f4
3 changed files with 105 additions and 90 deletions
|
@ -1,3 +1,5 @@
|
||||||
|
use std::path::PathBuf;
|
||||||
|
|
||||||
use bytes::Bytes;
|
use bytes::Bytes;
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use zstd::stream::{decode_all as zstd_decode, Encoder};
|
use zstd::stream::{decode_all as zstd_decode, Encoder};
|
||||||
|
@ -19,6 +21,13 @@ pub enum DataBlock {
|
||||||
Compressed(Bytes),
|
Compressed(Bytes),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub enum DataBlockPath {
|
||||||
|
/// Uncompressed data fail
|
||||||
|
Plain(PathBuf),
|
||||||
|
/// Compressed data fail
|
||||||
|
Compressed(PathBuf),
|
||||||
|
}
|
||||||
|
|
||||||
impl DataBlock {
|
impl DataBlock {
|
||||||
/// Query whether this block is compressed
|
/// Query whether this block is compressed
|
||||||
pub fn is_compressed(&self) -> bool {
|
pub fn is_compressed(&self) -> bool {
|
||||||
|
|
|
@ -190,32 +190,35 @@ impl DataLayout {
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn primary_data_dir(&self, hash: &Hash) -> PathBuf {
|
pub(crate) fn primary_block_dir(&self, hash: &Hash) -> PathBuf {
|
||||||
let ipart = self.partition_from(hash);
|
let ipart = self.partition_from(hash);
|
||||||
let idir = self.part_prim[ipart] as usize;
|
let idir = self.part_prim[ipart] as usize;
|
||||||
self.data_dir_from(hash, &self.data_dirs[idir].path)
|
self.block_dir_from(hash, &self.data_dirs[idir].path)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn secondary_data_dirs<'a>(&'a self, hash: &'a Hash) -> impl Iterator<Item=PathBuf> + 'a {
|
pub(crate) fn secondary_block_dirs<'a>(
|
||||||
let ipart = self.partition_from(hash);
|
&'a self,
|
||||||
self.part_sec[ipart]
|
hash: &'a Hash,
|
||||||
.iter()
|
) -> impl Iterator<Item = PathBuf> + 'a {
|
||||||
.map(move |idir| self.data_dir_from(hash, &self.data_dirs[*idir as usize].path))
|
let ipart = self.partition_from(hash);
|
||||||
|
self.part_sec[ipart]
|
||||||
|
.iter()
|
||||||
|
.map(move |idir| self.block_dir_from(hash, &self.data_dirs[*idir as usize].path))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn partition_from(&self, hash: &Hash) -> usize {
|
fn partition_from(&self, hash: &Hash) -> usize {
|
||||||
u16::from_be_bytes([
|
u16::from_be_bytes([
|
||||||
hash.as_slice()[DPART_BYTES.0],
|
hash.as_slice()[DPART_BYTES.0],
|
||||||
hash.as_slice()[DPART_BYTES.1]
|
hash.as_slice()[DPART_BYTES.1],
|
||||||
]) as usize % DRIVE_NPART
|
]) as usize % DRIVE_NPART
|
||||||
}
|
}
|
||||||
|
|
||||||
fn data_dir_from(&self, hash: &Hash, dir: &PathBuf) -> PathBuf {
|
fn block_dir_from(&self, hash: &Hash, dir: &PathBuf) -> PathBuf {
|
||||||
let mut path = dir.clone();
|
let mut path = dir.clone();
|
||||||
path.push(hex::encode(&hash.as_slice()[0..1]));
|
path.push(hex::encode(&hash.as_slice()[0..1]));
|
||||||
path.push(hex::encode(&hash.as_slice()[1..2]));
|
path.push(hex::encode(&hash.as_slice()[1..2]));
|
||||||
path
|
path
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl InitialFormat for DataLayout {
|
impl InitialFormat for DataLayout {
|
||||||
|
|
|
@ -543,21 +543,25 @@ impl BlockManager {
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn read_block_internal(&self, hash: &Hash) -> Result<DataBlock, Error> {
|
async fn read_block_internal(&self, hash: &Hash) -> Result<DataBlock, Error> {
|
||||||
let mut path = self.block_path(hash);
|
let block_path = match self.find_block(hash).await {
|
||||||
let compressed = match self.is_block_compressed(hash).await {
|
Some(p) => p,
|
||||||
Ok(c) => c,
|
None => {
|
||||||
Err(e) => {
|
|
||||||
// Not found but maybe we should have had it ??
|
// Not found but maybe we should have had it ??
|
||||||
self.resync
|
self.resync
|
||||||
.put_to_resync(hash, 2 * self.system.rpc.rpc_timeout())?;
|
.put_to_resync(hash, 2 * self.system.rpc.rpc_timeout())?;
|
||||||
return Err(Into::into(e));
|
return Err(Error::Message(format!(
|
||||||
|
"block {:?} not found on node",
|
||||||
|
hash
|
||||||
|
)));
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
if compressed {
|
|
||||||
path.set_extension("zst");
|
|
||||||
}
|
|
||||||
let mut f = fs::File::open(&path).await?;
|
|
||||||
|
|
||||||
|
let (path, compressed) = match &block_path {
|
||||||
|
DataBlockPath::Plain(p) => (p, false),
|
||||||
|
DataBlockPath::Compressed(p) => (p, true),
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut f = fs::File::open(&path).await?;
|
||||||
let mut data = vec![];
|
let mut data = vec![];
|
||||||
f.read_to_end(&mut data).await?;
|
f.read_to_end(&mut data).await?;
|
||||||
drop(f);
|
drop(f);
|
||||||
|
@ -571,11 +575,16 @@ impl BlockManager {
|
||||||
if data.verify(*hash).is_err() {
|
if data.verify(*hash).is_err() {
|
||||||
self.metrics.corruption_counter.add(1);
|
self.metrics.corruption_counter.add(1);
|
||||||
|
|
||||||
|
warn!(
|
||||||
|
"Block {:?} is corrupted. Renaming to .corrupted and resyncing.",
|
||||||
|
hash
|
||||||
|
);
|
||||||
self.lock_mutate(hash)
|
self.lock_mutate(hash)
|
||||||
.await
|
.await
|
||||||
.move_block_to_corrupted(hash, self)
|
.move_block_to_corrupted(&block_path)
|
||||||
.await?;
|
.await?;
|
||||||
self.resync.put_to_resync(hash, Duration::from_millis(0))?;
|
self.resync.put_to_resync(hash, Duration::from_millis(0))?;
|
||||||
|
|
||||||
return Err(Error::CorruptData(*hash));
|
return Err(Error::CorruptData(*hash));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -604,56 +613,51 @@ impl BlockManager {
|
||||||
.await
|
.await
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Utility: gives the path of the directory in which a block should be found
|
/// Utility: check if block is stored compressed.
|
||||||
fn block_dir(&self, hash: &Hash) -> PathBuf {
|
async fn find_block(&self, hash: &Hash) -> Option<DataBlockPath> {
|
||||||
self.data_layout.primary_data_dir(hash)
|
let dirs = Some(self.data_layout.primary_block_dir(hash))
|
||||||
}
|
.into_iter()
|
||||||
|
.chain(self.data_layout.secondary_block_dirs(hash));
|
||||||
|
let filename = hex::encode(hash.as_ref());
|
||||||
|
|
||||||
/// Utility: give the full path where a block should be found, minus extension if block is
|
for dir in dirs {
|
||||||
/// compressed
|
let mut path = dir;
|
||||||
fn block_path(&self, hash: &Hash) -> PathBuf {
|
path.push(&filename);
|
||||||
let mut path = self.block_dir(hash);
|
|
||||||
path.push(hex::encode(hash.as_ref()));
|
|
||||||
path
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Utility: check if block is stored compressed. Error if block is not stored
|
if self.compression_level.is_none() {
|
||||||
async fn is_block_compressed(&self, hash: &Hash) -> Result<bool, Error> {
|
// If compression is disabled on node - check for the raw block
|
||||||
let mut path = self.block_path(hash);
|
// first and then a compressed one (as compression may have been
|
||||||
|
// previously enabled).
|
||||||
// If compression is disabled on node - check for the raw block
|
|
||||||
// first and then a compressed one (as compression may have been
|
|
||||||
// previously enabled).
|
|
||||||
match self.compression_level {
|
|
||||||
None => {
|
|
||||||
if fs::metadata(&path).await.is_ok() {
|
if fs::metadata(&path).await.is_ok() {
|
||||||
return Ok(false);
|
return Some(DataBlockPath::Plain(path));
|
||||||
}
|
}
|
||||||
|
|
||||||
path.set_extension("zst");
|
path.set_extension("zst");
|
||||||
|
|
||||||
fs::metadata(&path).await.map(|_| true).map_err(Into::into)
|
|
||||||
}
|
|
||||||
_ => {
|
|
||||||
path.set_extension("zst");
|
|
||||||
|
|
||||||
if fs::metadata(&path).await.is_ok() {
|
if fs::metadata(&path).await.is_ok() {
|
||||||
return Ok(true);
|
return Some(DataBlockPath::Compressed(path));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
path.set_extension("zst");
|
||||||
|
if fs::metadata(&path).await.is_ok() {
|
||||||
|
return Some(DataBlockPath::Compressed(path));
|
||||||
}
|
}
|
||||||
|
|
||||||
path.set_extension("");
|
path.set_extension("");
|
||||||
|
if fs::metadata(&path).await.is_ok() {
|
||||||
fs::metadata(&path).await.map(|_| false).map_err(Into::into)
|
return Some(DataBlockPath::Plain(path));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
None
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn lock_mutate(&self, hash: &Hash) -> MutexGuard<'_, BlockManagerLocked> {
|
async fn lock_mutate(&self, hash: &Hash) -> MutexGuard<'_, BlockManagerLocked> {
|
||||||
let tracer = opentelemetry::global::tracer("garage");
|
let tracer = opentelemetry::global::tracer("garage");
|
||||||
self.mutation_lock[hash.as_slice()[0] as usize]
|
let ilock = u16::from_be_bytes([hash.as_slice()[0], hash.as_slice()[1]]) as usize
|
||||||
|
% self.mutation_lock.len();
|
||||||
|
self.mutation_lock[ilock]
|
||||||
.lock()
|
.lock()
|
||||||
.with_context(Context::current_with_span(
|
.with_context(Context::current_with_span(
|
||||||
tracer.start("Acquire mutation_lock"),
|
tracer.start(format!("Acquire mutation_lock #{}", ilock)),
|
||||||
))
|
))
|
||||||
.await
|
.await
|
||||||
}
|
}
|
||||||
|
@ -688,7 +692,7 @@ impl BlockManagerLocked {
|
||||||
hash: &Hash,
|
hash: &Hash,
|
||||||
mgr: &BlockManager,
|
mgr: &BlockManager,
|
||||||
) -> Result<BlockStatus, Error> {
|
) -> Result<BlockStatus, Error> {
|
||||||
let exists = mgr.is_block_compressed(hash).await.is_ok();
|
let exists = mgr.find_block(hash).await.is_some();
|
||||||
let needed = mgr.rc.get_block_rc(hash)?;
|
let needed = mgr.rc.get_block_rc(hash)?;
|
||||||
|
|
||||||
Ok(BlockStatus { exists, needed })
|
Ok(BlockStatus { exists, needed })
|
||||||
|
@ -703,21 +707,17 @@ impl BlockManagerLocked {
|
||||||
let compressed = data.is_compressed();
|
let compressed = data.is_compressed();
|
||||||
let data = data.inner_buffer();
|
let data = data.inner_buffer();
|
||||||
|
|
||||||
let mut path = mgr.block_dir(hash);
|
let mut path = mgr.data_layout.primary_block_dir(hash);
|
||||||
let directory = path.clone();
|
let directory = path.clone();
|
||||||
path.push(hex::encode(hash));
|
path.push(hex::encode(hash));
|
||||||
|
|
||||||
fs::create_dir_all(&directory).await?;
|
fs::create_dir_all(&directory).await?;
|
||||||
|
|
||||||
let to_delete = match (mgr.is_block_compressed(hash).await, compressed) {
|
let to_delete = match (mgr.find_block(hash).await, compressed) {
|
||||||
(Ok(true), _) => return Ok(()),
|
(Some(DataBlockPath::Compressed(_)), _) => return Ok(()),
|
||||||
(Ok(false), false) => return Ok(()),
|
(Some(DataBlockPath::Plain(_)), false) => return Ok(()),
|
||||||
(Ok(false), true) => {
|
(Some(DataBlockPath::Plain(plain_path)), true) => Some(plain_path),
|
||||||
let path_to_delete = path.clone();
|
(None, compressed) => {
|
||||||
path.set_extension("zst");
|
|
||||||
Some(path_to_delete)
|
|
||||||
}
|
|
||||||
(Err(_), compressed) => {
|
|
||||||
if compressed {
|
if compressed {
|
||||||
path.set_extension("zst");
|
path.set_extension("zst");
|
||||||
}
|
}
|
||||||
|
@ -766,19 +766,20 @@ impl BlockManagerLocked {
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn move_block_to_corrupted(&self, hash: &Hash, mgr: &BlockManager) -> Result<(), Error> {
|
async fn move_block_to_corrupted(&self, block_path: &DataBlockPath) -> Result<(), Error> {
|
||||||
warn!(
|
let (path, path2) = match block_path {
|
||||||
"Block {:?} is corrupted. Renaming to .corrupted and resyncing.",
|
DataBlockPath::Plain(p) => {
|
||||||
hash
|
let mut p2 = p.clone();
|
||||||
);
|
p2.set_extension("corrupted");
|
||||||
let mut path = mgr.block_path(hash);
|
(p, p2)
|
||||||
let mut path2 = path.clone();
|
}
|
||||||
if mgr.is_block_compressed(hash).await? {
|
DataBlockPath::Compressed(p) => {
|
||||||
path.set_extension("zst");
|
let mut p2 = p.clone();
|
||||||
path2.set_extension("zst.corrupted");
|
p2.set_extension("zst.corrupted");
|
||||||
} else {
|
(p, p2)
|
||||||
path2.set_extension("corrupted");
|
}
|
||||||
}
|
};
|
||||||
|
|
||||||
fs::rename(path, path2).await?;
|
fs::rename(path, path2).await?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
@ -787,12 +788,14 @@ impl BlockManagerLocked {
|
||||||
let BlockStatus { exists, needed } = self.check_block_status(hash, mgr).await?;
|
let BlockStatus { exists, needed } = self.check_block_status(hash, mgr).await?;
|
||||||
|
|
||||||
if exists && needed.is_deletable() {
|
if exists && needed.is_deletable() {
|
||||||
let mut path = mgr.block_path(hash);
|
let path_opt = match mgr.find_block(hash).await {
|
||||||
if mgr.is_block_compressed(hash).await? {
|
Some(DataBlockPath::Plain(p)) | Some(DataBlockPath::Compressed(p)) => Some(p),
|
||||||
path.set_extension("zst");
|
None => None,
|
||||||
|
};
|
||||||
|
if let Some(path) = path_opt {
|
||||||
|
fs::remove_file(path).await?;
|
||||||
|
mgr.metrics.delete_counter.add(1);
|
||||||
}
|
}
|
||||||
fs::remove_file(path).await?;
|
|
||||||
mgr.metrics.delete_counter.add(1);
|
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue