mirror of
https://git.deuxfleurs.fr/Deuxfleurs/garage.git
synced 2024-11-25 09:31:00 +00:00
block manager: skeleton for multi-hdd support
This commit is contained in:
parent
4b4f2000f4
commit
71c0188055
7 changed files with 280 additions and 93 deletions
57
src/block/layout.rs
Normal file
57
src/block/layout.rs
Normal file
|
@ -0,0 +1,57 @@
|
||||||
|
use std::path::PathBuf;
|
||||||
|
|
||||||
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
|
use garage_util::config::DataDirEnum;
|
||||||
|
use garage_util::data::Hash;
|
||||||
|
use garage_util::migrate::*;
|
||||||
|
|
||||||
|
pub const DRIVE_NPART: usize = 1024;
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug, Clone)]
|
||||||
|
pub(crate) struct DataLayout {
|
||||||
|
pub(crate) data_dirs: Vec<DataDir>,
|
||||||
|
pub(crate) partitions: Vec<Partition>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug, Clone)]
|
||||||
|
pub(crate) struct DataDir {
|
||||||
|
pub(crate) path: PathBuf,
|
||||||
|
pub(crate) state: DataDirState,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug, Clone)]
|
||||||
|
pub(crate) enum DataDirState {
|
||||||
|
Active { capacity: u64 },
|
||||||
|
ReadOnly,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Serialize, Deserialize, Debug, Clone)]
|
||||||
|
pub(crate) struct Partition {
|
||||||
|
pub(crate) prim: usize,
|
||||||
|
pub(crate) sec: Vec<usize>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DataLayout {
|
||||||
|
pub(crate) fn initialize(dirs: &DataDirEnum) -> Self {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) fn update(&mut self, dirs: &DataDirEnum) -> Self {
|
||||||
|
todo!()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub(crate) fn data_dir(&self, hash: &Hash) -> PathBuf {
|
||||||
|
todo!()
|
||||||
|
/*
|
||||||
|
let mut path = self.data_dir.clone();
|
||||||
|
path.push(hex::encode(&hash.as_slice()[0..1]));
|
||||||
|
path.push(hex::encode(&hash.as_slice()[1..2]));
|
||||||
|
path
|
||||||
|
*/
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl InitialFormat for DataLayout {
|
||||||
|
const VERSION_MARKER: &'static [u8] = b"G09bmdl";
|
||||||
|
}
|
|
@ -6,5 +6,6 @@ pub mod repair;
|
||||||
pub mod resync;
|
pub mod resync;
|
||||||
|
|
||||||
mod block;
|
mod block;
|
||||||
|
mod layout;
|
||||||
mod metrics;
|
mod metrics;
|
||||||
mod rc;
|
mod rc;
|
||||||
|
|
|
@ -25,10 +25,11 @@ use garage_rpc::rpc_helper::netapp::stream::{stream_asyncread, ByteStream};
|
||||||
use garage_db as db;
|
use garage_db as db;
|
||||||
|
|
||||||
use garage_util::background::{vars, BackgroundRunner};
|
use garage_util::background::{vars, BackgroundRunner};
|
||||||
|
use garage_util::config::DataDirEnum;
|
||||||
use garage_util::data::*;
|
use garage_util::data::*;
|
||||||
use garage_util::error::*;
|
use garage_util::error::*;
|
||||||
use garage_util::metrics::RecordDuration;
|
use garage_util::metrics::RecordDuration;
|
||||||
use garage_util::persister::PersisterShared;
|
use garage_util::persister::{Persister, PersisterShared};
|
||||||
use garage_util::time::msec_to_rfc3339;
|
use garage_util::time::msec_to_rfc3339;
|
||||||
|
|
||||||
use garage_rpc::rpc_helper::OrderTag;
|
use garage_rpc::rpc_helper::OrderTag;
|
||||||
|
@ -38,6 +39,7 @@ use garage_rpc::*;
|
||||||
use garage_table::replication::{TableReplication, TableShardedReplication};
|
use garage_table::replication::{TableReplication, TableShardedReplication};
|
||||||
|
|
||||||
use crate::block::*;
|
use crate::block::*;
|
||||||
|
use crate::layout::*;
|
||||||
use crate::metrics::*;
|
use crate::metrics::*;
|
||||||
use crate::rc::*;
|
use crate::rc::*;
|
||||||
use crate::repair::*;
|
use crate::repair::*;
|
||||||
|
@ -77,8 +79,11 @@ impl Rpc for BlockRpc {
|
||||||
pub struct BlockManager {
|
pub struct BlockManager {
|
||||||
/// Replication strategy, allowing to find on which node blocks should be located
|
/// Replication strategy, allowing to find on which node blocks should be located
|
||||||
pub replication: TableShardedReplication,
|
pub replication: TableShardedReplication,
|
||||||
/// Directory in which block are stored
|
|
||||||
pub data_dir: PathBuf,
|
/// Directory/ies in which block are stored
|
||||||
|
pub data_dir: DataDirEnum,
|
||||||
|
/// Data layout
|
||||||
|
pub(crate) data_layout: DataLayout,
|
||||||
|
|
||||||
data_fsync: bool,
|
data_fsync: bool,
|
||||||
compression_level: Option<i32>,
|
compression_level: Option<i32>,
|
||||||
|
@ -114,12 +119,22 @@ struct BlockManagerLocked();
|
||||||
impl BlockManager {
|
impl BlockManager {
|
||||||
pub fn new(
|
pub fn new(
|
||||||
db: &db::Db,
|
db: &db::Db,
|
||||||
data_dir: PathBuf,
|
data_dir: DataDirEnum,
|
||||||
data_fsync: bool,
|
data_fsync: bool,
|
||||||
compression_level: Option<i32>,
|
compression_level: Option<i32>,
|
||||||
replication: TableShardedReplication,
|
replication: TableShardedReplication,
|
||||||
system: Arc<System>,
|
system: Arc<System>,
|
||||||
) -> Arc<Self> {
|
) -> Arc<Self> {
|
||||||
|
let layout_persister: Persister<DataLayout> =
|
||||||
|
Persister::new(&system.metadata_dir, "data_layout");
|
||||||
|
let data_layout = match layout_persister.load() {
|
||||||
|
Ok(mut layout) => {
|
||||||
|
layout.update(&data_dir);
|
||||||
|
layout
|
||||||
|
}
|
||||||
|
Err(_) => DataLayout::initialize(&data_dir),
|
||||||
|
};
|
||||||
|
|
||||||
let rc = db
|
let rc = db
|
||||||
.open_tree("block_local_rc")
|
.open_tree("block_local_rc")
|
||||||
.expect("Unable to open block_local_rc tree");
|
.expect("Unable to open block_local_rc tree");
|
||||||
|
@ -143,6 +158,7 @@ impl BlockManager {
|
||||||
let block_manager = Arc::new(Self {
|
let block_manager = Arc::new(Self {
|
||||||
replication,
|
replication,
|
||||||
data_dir,
|
data_dir,
|
||||||
|
data_layout,
|
||||||
data_fsync,
|
data_fsync,
|
||||||
compression_level,
|
compression_level,
|
||||||
mutation_lock: [(); 256].map(|_| Mutex::new(BlockManagerLocked())),
|
mutation_lock: [(); 256].map(|_| Mutex::new(BlockManagerLocked())),
|
||||||
|
@ -586,10 +602,7 @@ impl BlockManager {
|
||||||
|
|
||||||
/// Utility: gives the path of the directory in which a block should be found
|
/// Utility: gives the path of the directory in which a block should be found
|
||||||
fn block_dir(&self, hash: &Hash) -> PathBuf {
|
fn block_dir(&self, hash: &Hash) -> PathBuf {
|
||||||
let mut path = self.data_dir.clone();
|
self.data_layout.data_dir(hash)
|
||||||
path.push(hex::encode(&hash.as_slice()[0..1]));
|
|
||||||
path.push(hex::encode(&hash.as_slice()[1..2]));
|
|
||||||
path
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Utility: give the full path where a block should be found, minus extension if block is
|
/// Utility: give the full path where a block should be found, minus extension if block is
|
||||||
|
|
|
@ -17,6 +17,7 @@ use garage_util::persister::PersisterShared;
|
||||||
use garage_util::time::*;
|
use garage_util::time::*;
|
||||||
use garage_util::tranquilizer::Tranquilizer;
|
use garage_util::tranquilizer::Tranquilizer;
|
||||||
|
|
||||||
|
use crate::layout::*;
|
||||||
use crate::manager::*;
|
use crate::manager::*;
|
||||||
|
|
||||||
// Full scrub every 25 days with a random element of 10 days mixed in below
|
// Full scrub every 25 days with a random element of 10 days mixed in below
|
||||||
|
@ -136,7 +137,7 @@ impl Worker for RepairWorker {
|
||||||
// Lists all blocks on disk and adds them to the resync queue.
|
// Lists all blocks on disk and adds them to the resync queue.
|
||||||
// This allows us to find blocks we are storing but don't actually need,
|
// This allows us to find blocks we are storing but don't actually need,
|
||||||
// so that we can offload them if necessary and then delete them locally.
|
// so that we can offload them if necessary and then delete them locally.
|
||||||
if let Some(hash) = bi.next().await? {
|
if let Some((_path, hash)) = bi.next().await? {
|
||||||
self.manager
|
self.manager
|
||||||
.resync
|
.resync
|
||||||
.put_to_resync(&hash, Duration::from_secs(0))?;
|
.put_to_resync(&hash, Duration::from_secs(0))?;
|
||||||
|
@ -376,7 +377,7 @@ impl Worker for ScrubWorker {
|
||||||
match &mut self.work {
|
match &mut self.work {
|
||||||
ScrubWorkerState::Running(bsi) => {
|
ScrubWorkerState::Running(bsi) => {
|
||||||
self.tranquilizer.reset();
|
self.tranquilizer.reset();
|
||||||
if let Some(hash) = bsi.next().await? {
|
if let Some((_path, hash)) = bsi.next().await? {
|
||||||
match self.manager.read_block(&hash).await {
|
match self.manager.read_block(&hash).await {
|
||||||
Err(Error::CorruptData(_)) => {
|
Err(Error::CorruptData(_)) => {
|
||||||
error!("Found corrupt data block during scrub: {:?}", hash);
|
error!("Found corrupt data block during scrub: {:?}", hash);
|
||||||
|
@ -447,100 +448,166 @@ impl Worker for ScrubWorker {
|
||||||
// UTILITY FOR ENUMERATING THE BLOCK STORE
|
// UTILITY FOR ENUMERATING THE BLOCK STORE
|
||||||
// ---- ---- ----
|
// ---- ---- ----
|
||||||
|
|
||||||
|
const PROGRESS_FP: u64 = 1_000_000_000;
|
||||||
|
|
||||||
struct BlockStoreIterator {
|
struct BlockStoreIterator {
|
||||||
path: Vec<ReadingDir>,
|
todo: Vec<BsiTodo>,
|
||||||
}
|
}
|
||||||
|
|
||||||
enum ReadingDir {
|
enum BsiTodo {
|
||||||
Pending(PathBuf),
|
Directory {
|
||||||
Read {
|
path: PathBuf,
|
||||||
subpaths: Vec<fs::DirEntry>,
|
progress_min: u64,
|
||||||
pos: usize,
|
progress_max: u64,
|
||||||
|
},
|
||||||
|
File {
|
||||||
|
path: PathBuf,
|
||||||
|
filename: String,
|
||||||
|
progress: u64,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
impl BlockStoreIterator {
|
impl BlockStoreIterator {
|
||||||
fn new(manager: &BlockManager) -> Self {
|
fn new(manager: &BlockManager) -> Self {
|
||||||
let root_dir = manager.data_dir.clone();
|
let min_cap = manager
|
||||||
Self {
|
.data_layout
|
||||||
path: vec![ReadingDir::Pending(root_dir)],
|
.data_dirs
|
||||||
|
.iter()
|
||||||
|
.filter_map(|x| match x.state {
|
||||||
|
DataDirState::Active { capacity } => Some(capacity),
|
||||||
|
_ => None,
|
||||||
|
})
|
||||||
|
.min()
|
||||||
|
.unwrap_or(0);
|
||||||
|
|
||||||
|
let sum_cap = manager
|
||||||
|
.data_layout
|
||||||
|
.data_dirs
|
||||||
|
.iter()
|
||||||
|
.map(|x| match x.state {
|
||||||
|
DataDirState::Active { capacity } => capacity,
|
||||||
|
_ => min_cap, // approximation
|
||||||
|
})
|
||||||
|
.sum::<u64>() as u128;
|
||||||
|
|
||||||
|
let mut cum_cap = 0;
|
||||||
|
let mut todo = vec![];
|
||||||
|
for dir in manager.data_layout.data_dirs.iter() {
|
||||||
|
let cap = match dir.state {
|
||||||
|
DataDirState::Active { capacity } => capacity,
|
||||||
|
_ => min_cap,
|
||||||
|
};
|
||||||
|
|
||||||
|
let progress_min = ((cum_cap as u128 * PROGRESS_FP as u128) / (sum_cap as u128)) as u64;
|
||||||
|
let progress_max =
|
||||||
|
(((cum_cap + cap) as u128 * PROGRESS_FP as u128) / (sum_cap as u128)) as u64;
|
||||||
|
cum_cap += cap;
|
||||||
|
|
||||||
|
todo.push(BsiTodo::Directory {
|
||||||
|
path: dir.path.clone(),
|
||||||
|
progress_min,
|
||||||
|
progress_max,
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
// entries are processed back-to-front (because of .pop()),
|
||||||
|
// so reverse entries to process them in increasing progress bounds
|
||||||
|
todo.reverse();
|
||||||
|
|
||||||
|
let ret = Self { todo };
|
||||||
|
debug_assert!(ret.progress_invariant());
|
||||||
|
|
||||||
|
ret
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns progress done, between 0 and 1
|
/// Returns progress done, between 0 and 1
|
||||||
fn progress(&self) -> f32 {
|
fn progress(&self) -> f32 {
|
||||||
if self.path.is_empty() {
|
self.todo
|
||||||
1.0
|
.last()
|
||||||
} else {
|
.map(|x| match x {
|
||||||
let mut ret = 0.0;
|
BsiTodo::Directory { progress_min, .. } => *progress_min,
|
||||||
let mut next_div = 1;
|
BsiTodo::File { progress, .. } => *progress,
|
||||||
for p in self.path.iter() {
|
})
|
||||||
match p {
|
.map(|x| x as f32 / PROGRESS_FP as f32)
|
||||||
ReadingDir::Pending(_) => break,
|
.unwrap_or(1.0)
|
||||||
ReadingDir::Read { subpaths, pos } => {
|
}
|
||||||
next_div *= subpaths.len();
|
|
||||||
ret += ((*pos - 1) as f32) / (next_div as f32);
|
async fn next(&mut self) -> Result<Option<(PathBuf, Hash)>, Error> {
|
||||||
|
loop {
|
||||||
|
match self.todo.pop() {
|
||||||
|
None => return Ok(None),
|
||||||
|
Some(BsiTodo::Directory {
|
||||||
|
path,
|
||||||
|
progress_min,
|
||||||
|
progress_max,
|
||||||
|
}) => {
|
||||||
|
let istart = self.todo.len();
|
||||||
|
|
||||||
|
let mut reader = fs::read_dir(&path).await?;
|
||||||
|
while let Some(ent) = reader.next_entry().await? {
|
||||||
|
let name = if let Ok(n) = ent.file_name().into_string() {
|
||||||
|
n
|
||||||
|
} else {
|
||||||
|
continue;
|
||||||
|
};
|
||||||
|
let ft = ent.file_type().await?;
|
||||||
|
if ft.is_dir() && hex::decode(&name).is_ok() {
|
||||||
|
self.todo.push(BsiTodo::Directory {
|
||||||
|
path: ent.path(),
|
||||||
|
progress_min: 0,
|
||||||
|
progress_max: 0,
|
||||||
|
});
|
||||||
|
} else if ft.is_file() {
|
||||||
|
self.todo.push(BsiTodo::File {
|
||||||
|
path: ent.path(),
|
||||||
|
filename: name,
|
||||||
|
progress: 0,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let count = self.todo.len() - istart;
|
||||||
|
for (i, ent) in self.todo[istart..].iter_mut().enumerate() {
|
||||||
|
let p1 = progress_min
|
||||||
|
+ ((progress_max - progress_min) * i as u64) / count as u64;
|
||||||
|
let p2 = progress_min
|
||||||
|
+ ((progress_max - progress_min) * (i + 1) as u64) / count as u64;
|
||||||
|
match ent {
|
||||||
|
BsiTodo::Directory {
|
||||||
|
progress_min,
|
||||||
|
progress_max,
|
||||||
|
..
|
||||||
|
} => {
|
||||||
|
*progress_min = p1;
|
||||||
|
*progress_max = p2;
|
||||||
|
}
|
||||||
|
BsiTodo::File { progress, .. } => {
|
||||||
|
*progress = p1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
self.todo[istart..].reverse();
|
||||||
|
debug_assert!(self.progress_invariant());
|
||||||
|
}
|
||||||
|
Some(BsiTodo::File { path, filename, .. }) => {
|
||||||
|
let filename = filename.strip_suffix(".zst").unwrap_or(&filename);
|
||||||
|
if filename.len() == 64 {
|
||||||
|
if let Ok(h) = hex::decode(filename) {
|
||||||
|
let mut hash = [0u8; 32];
|
||||||
|
hash.copy_from_slice(&h);
|
||||||
|
return Ok(Some((path, hash.into())));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
ret
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn next(&mut self) -> Result<Option<Hash>, Error> {
|
fn progress_invariant(&self) -> bool {
|
||||||
loop {
|
let iter = self.todo.iter().map(|x| match x {
|
||||||
let last_path = match self.path.last_mut() {
|
BsiTodo::Directory { progress_min, .. } => progress_min,
|
||||||
None => return Ok(None),
|
BsiTodo::File { progress, .. } => progress,
|
||||||
Some(lp) => lp,
|
});
|
||||||
};
|
let iter_1 = iter.clone().skip(1);
|
||||||
|
iter.zip(iter_1).all(|(prev, next)| prev >= next)
|
||||||
if let ReadingDir::Pending(path) = last_path {
|
|
||||||
let mut reader = fs::read_dir(&path).await?;
|
|
||||||
let mut subpaths = vec![];
|
|
||||||
while let Some(ent) = reader.next_entry().await? {
|
|
||||||
subpaths.push(ent);
|
|
||||||
}
|
|
||||||
*last_path = ReadingDir::Read { subpaths, pos: 0 };
|
|
||||||
}
|
|
||||||
|
|
||||||
let (subpaths, pos) = match *last_path {
|
|
||||||
ReadingDir::Read {
|
|
||||||
ref subpaths,
|
|
||||||
ref mut pos,
|
|
||||||
} => (subpaths, pos),
|
|
||||||
ReadingDir::Pending(_) => unreachable!(),
|
|
||||||
};
|
|
||||||
|
|
||||||
let data_dir_ent = match subpaths.get(*pos) {
|
|
||||||
None => {
|
|
||||||
self.path.pop();
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
Some(ent) => {
|
|
||||||
*pos += 1;
|
|
||||||
ent
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let name = data_dir_ent.file_name();
|
|
||||||
let name = if let Ok(n) = name.into_string() {
|
|
||||||
n
|
|
||||||
} else {
|
|
||||||
continue;
|
|
||||||
};
|
|
||||||
let ent_type = data_dir_ent.file_type().await?;
|
|
||||||
|
|
||||||
let name = name.strip_suffix(".zst").unwrap_or(&name);
|
|
||||||
if name.len() == 2 && hex::decode(name).is_ok() && ent_type.is_dir() {
|
|
||||||
let path = data_dir_ent.path();
|
|
||||||
self.path.push(ReadingDir::Pending(path));
|
|
||||||
} else if name.len() == 64 {
|
|
||||||
if let Ok(h) = hex::decode(name) {
|
|
||||||
let mut hash = [0u8; 32];
|
|
||||||
hash.copy_from_slice(&h);
|
|
||||||
return Ok(Some(hash.into()));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -92,8 +92,22 @@ impl Garage {
|
||||||
// Create meta dir and data dir if they don't exist already
|
// Create meta dir and data dir if they don't exist already
|
||||||
std::fs::create_dir_all(&config.metadata_dir)
|
std::fs::create_dir_all(&config.metadata_dir)
|
||||||
.ok_or_message("Unable to create Garage metadata directory")?;
|
.ok_or_message("Unable to create Garage metadata directory")?;
|
||||||
std::fs::create_dir_all(&config.data_dir)
|
match &config.data_dir {
|
||||||
.ok_or_message("Unable to create Garage data directory")?;
|
DataDirEnum::Single(data_dir) => {
|
||||||
|
std::fs::create_dir_all(data_dir).ok_or_message(format!(
|
||||||
|
"Unable to create Garage data directory: {}",
|
||||||
|
data_dir.to_string_lossy()
|
||||||
|
))?;
|
||||||
|
}
|
||||||
|
DataDirEnum::Multiple(data_dirs) => {
|
||||||
|
for dir in data_dirs {
|
||||||
|
std::fs::create_dir_all(&dir.path).ok_or_message(format!(
|
||||||
|
"Unable to create Garage data directory: {}",
|
||||||
|
dir.path.to_string_lossy()
|
||||||
|
))?;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
info!("Opening database...");
|
info!("Opening database...");
|
||||||
let mut db_path = config.metadata_dir.clone();
|
let mut db_path = config.metadata_dir.clone();
|
||||||
|
|
|
@ -22,9 +22,9 @@ use netapp::peering::fullmesh::FullMeshPeeringStrategy;
|
||||||
use netapp::util::parse_and_resolve_peer_addr_async;
|
use netapp::util::parse_and_resolve_peer_addr_async;
|
||||||
use netapp::{NetApp, NetworkKey, NodeID, NodeKey};
|
use netapp::{NetApp, NetworkKey, NodeID, NodeKey};
|
||||||
|
|
||||||
use garage_util::config::Config;
|
|
||||||
#[cfg(feature = "kubernetes-discovery")]
|
#[cfg(feature = "kubernetes-discovery")]
|
||||||
use garage_util::config::KubernetesDiscoveryConfig;
|
use garage_util::config::KubernetesDiscoveryConfig;
|
||||||
|
use garage_util::config::{Config, DataDirEnum};
|
||||||
use garage_util::data::*;
|
use garage_util::data::*;
|
||||||
use garage_util::error::*;
|
use garage_util::error::*;
|
||||||
use garage_util::persister::Persister;
|
use garage_util::persister::Persister;
|
||||||
|
@ -119,7 +119,7 @@ pub struct System {
|
||||||
/// Path to metadata directory
|
/// Path to metadata directory
|
||||||
pub metadata_dir: PathBuf,
|
pub metadata_dir: PathBuf,
|
||||||
/// Path to data directory
|
/// Path to data directory
|
||||||
pub data_dir: PathBuf,
|
pub data_dir: DataDirEnum,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||||
|
@ -890,7 +890,12 @@ impl NodeStatus {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn update_disk_usage(&mut self, meta_dir: &Path, data_dir: &Path, metrics: &SystemMetrics) {
|
fn update_disk_usage(
|
||||||
|
&mut self,
|
||||||
|
meta_dir: &Path,
|
||||||
|
data_dir: &DataDirEnum,
|
||||||
|
metrics: &SystemMetrics,
|
||||||
|
) {
|
||||||
use systemstat::{Platform, System};
|
use systemstat::{Platform, System};
|
||||||
let mounts = System::new().mounts().unwrap_or_default();
|
let mounts = System::new().mounts().unwrap_or_default();
|
||||||
|
|
||||||
|
@ -903,7 +908,17 @@ impl NodeStatus {
|
||||||
};
|
};
|
||||||
|
|
||||||
self.meta_disk_avail = mount_avail(meta_dir);
|
self.meta_disk_avail = mount_avail(meta_dir);
|
||||||
self.data_disk_avail = mount_avail(data_dir);
|
self.data_disk_avail = match data_dir {
|
||||||
|
DataDirEnum::Single(dir) => mount_avail(dir),
|
||||||
|
DataDirEnum::Multiple(dirs) => {
|
||||||
|
dirs.iter()
|
||||||
|
.map(|d| mount_avail(&d.path))
|
||||||
|
.fold(Some((0, 0)), |acc, cur| match (acc, cur) {
|
||||||
|
(Some((x, y)), Some((a, b))) => Some((x + a, y + b)),
|
||||||
|
_ => None,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
if let Some((avail, total)) = self.meta_disk_avail {
|
if let Some((avail, total)) = self.meta_disk_avail {
|
||||||
metrics
|
metrics
|
||||||
|
|
|
@ -13,7 +13,7 @@ pub struct Config {
|
||||||
/// Path where to store metadata. Should be fast, but low volume
|
/// Path where to store metadata. Should be fast, but low volume
|
||||||
pub metadata_dir: PathBuf,
|
pub metadata_dir: PathBuf,
|
||||||
/// Path where to store data. Can be slower, but need higher volume
|
/// Path where to store data. Can be slower, but need higher volume
|
||||||
pub data_dir: PathBuf,
|
pub data_dir: DataDirEnum,
|
||||||
|
|
||||||
/// Whether to fsync after all metadata transactions (disabled by default)
|
/// Whether to fsync after all metadata transactions (disabled by default)
|
||||||
#[serde(default)]
|
#[serde(default)]
|
||||||
|
@ -94,6 +94,26 @@ pub struct Config {
|
||||||
pub admin: AdminConfig,
|
pub admin: AdminConfig,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Value for data_dir: either a single directory or a list of dirs with attributes
|
||||||
|
#[derive(Deserialize, Debug, Clone)]
|
||||||
|
#[serde(untagged)]
|
||||||
|
pub enum DataDirEnum {
|
||||||
|
Single(PathBuf),
|
||||||
|
Multiple(Vec<DataDir>),
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Deserialize, Debug, Clone)]
|
||||||
|
pub struct DataDir {
|
||||||
|
/// Path to the data directory
|
||||||
|
pub path: PathBuf,
|
||||||
|
/// Capacity of the drive (required if read_only is false)
|
||||||
|
#[serde(default)]
|
||||||
|
pub capacity: Option<String>,
|
||||||
|
/// Whether this is a legacy read-only path (capacity should be None)
|
||||||
|
#[serde(default)]
|
||||||
|
pub read_only: bool,
|
||||||
|
}
|
||||||
|
|
||||||
/// Configuration for S3 api
|
/// Configuration for S3 api
|
||||||
#[derive(Deserialize, Debug, Clone)]
|
#[derive(Deserialize, Debug, Clone)]
|
||||||
pub struct S3ApiConfig {
|
pub struct S3ApiConfig {
|
||||||
|
|
Loading…
Reference in a new issue