datastore: clippy fixes
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
This commit is contained in:
parent
118e984996
commit
ca3f8757ba
@ -5,11 +5,11 @@ use anyhow::{bail, Error};
|
|||||||
use pbs_datastore::DataStore;
|
use pbs_datastore::DataStore;
|
||||||
|
|
||||||
fn run() -> Result<(), Error> {
|
fn run() -> Result<(), Error> {
|
||||||
let base: PathBuf = match std::env::args().skip(1).next() {
|
let base: PathBuf = match std::env::args().nth(1) {
|
||||||
Some(path) => path.into(),
|
Some(path) => path.into(),
|
||||||
None => bail!("no path passed!\n\nusage: ls-snapshots <path> [<max-depth>]"),
|
None => bail!("no path passed!\n\nusage: ls-snapshots <path> [<max-depth>]"),
|
||||||
};
|
};
|
||||||
let max_depth: Option<usize> = match std::env::args().skip(2).next() {
|
let max_depth: Option<usize> = match std::env::args().nth(2) {
|
||||||
Some(depth) => match depth.parse::<usize>() {
|
Some(depth) => match depth.parse::<usize>() {
|
||||||
Ok(depth) if depth < 8 => Some(depth),
|
Ok(depth) if depth < 8 => Some(depth),
|
||||||
Ok(_) => bail!("max-depth must be < 8"),
|
Ok(_) => bail!("max-depth must be < 8"),
|
||||||
|
@ -227,7 +227,7 @@ impl BackupGroup {
|
|||||||
/// Set the backup owner.
|
/// Set the backup owner.
|
||||||
pub fn set_owner(&self, auth_id: &Authid, force: bool) -> Result<(), Error> {
|
pub fn set_owner(&self, auth_id: &Authid, force: bool) -> Result<(), Error> {
|
||||||
self.store
|
self.store
|
||||||
.set_owner(&self.ns, &self.as_ref(), auth_id, force)
|
.set_owner(&self.ns, self.as_ref(), auth_id, force)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -572,7 +572,7 @@ impl From<&BackupDir> for pbs_api_types::BackupGroup {
|
|||||||
|
|
||||||
impl From<BackupDir> for pbs_api_types::BackupGroup {
|
impl From<BackupDir> for pbs_api_types::BackupGroup {
|
||||||
fn from(dir: BackupDir) -> pbs_api_types::BackupGroup {
|
fn from(dir: BackupDir) -> pbs_api_types::BackupGroup {
|
||||||
dir.dir.group.into()
|
dir.dir.group
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -394,7 +394,7 @@ impl DataStore {
|
|||||||
// construct ns before mkdir to enforce max-depth and name validity
|
// construct ns before mkdir to enforce max-depth and name validity
|
||||||
let ns = BackupNamespace::from_parent_ns(parent, name)?;
|
let ns = BackupNamespace::from_parent_ns(parent, name)?;
|
||||||
|
|
||||||
let mut ns_full_path = self.base_path().to_owned();
|
let mut ns_full_path = self.base_path();
|
||||||
ns_full_path.push(ns.path());
|
ns_full_path.push(ns.path());
|
||||||
|
|
||||||
std::fs::create_dir_all(ns_full_path)?;
|
std::fs::create_dir_all(ns_full_path)?;
|
||||||
@ -404,7 +404,7 @@ impl DataStore {
|
|||||||
|
|
||||||
/// Returns if the given namespace exists on the datastore
|
/// Returns if the given namespace exists on the datastore
|
||||||
pub fn namespace_exists(&self, ns: &BackupNamespace) -> bool {
|
pub fn namespace_exists(&self, ns: &BackupNamespace) -> bool {
|
||||||
let mut path = self.base_path().to_owned();
|
let mut path = self.base_path();
|
||||||
path.push(ns.path());
|
path.push(ns.path());
|
||||||
path.exists()
|
path.exists()
|
||||||
}
|
}
|
||||||
@ -704,7 +704,7 @@ impl DataStore {
|
|||||||
) -> Result<impl Iterator<Item = BackupNamespace> + 'static, Error> {
|
) -> Result<impl Iterator<Item = BackupNamespace> + 'static, Error> {
|
||||||
let this = Arc::clone(self);
|
let this = Arc::clone(self);
|
||||||
Ok(
|
Ok(
|
||||||
ListNamespaces::new(Arc::clone(&self), ns)?.filter_map(move |ns| match ns {
|
ListNamespaces::new(Arc::clone(self), ns)?.filter_map(move |ns| match ns {
|
||||||
Ok(ns) => Some(ns),
|
Ok(ns) => Some(ns),
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
log::error!("list groups error on datastore {} - {}", this.name(), err);
|
log::error!("list groups error on datastore {} - {}", this.name(), err);
|
||||||
@ -736,9 +736,9 @@ impl DataStore {
|
|||||||
) -> Result<impl Iterator<Item = BackupNamespace> + 'static, Error> {
|
) -> Result<impl Iterator<Item = BackupNamespace> + 'static, Error> {
|
||||||
let this = Arc::clone(self);
|
let this = Arc::clone(self);
|
||||||
Ok(if let Some(depth) = max_depth {
|
Ok(if let Some(depth) = max_depth {
|
||||||
ListNamespacesRecursive::new_max_depth(Arc::clone(&self), ns, depth)?
|
ListNamespacesRecursive::new_max_depth(Arc::clone(self), ns, depth)?
|
||||||
} else {
|
} else {
|
||||||
ListNamespacesRecursive::new(Arc::clone(&self), ns)?
|
ListNamespacesRecursive::new(Arc::clone(self), ns)?
|
||||||
}
|
}
|
||||||
.filter_map(move |ns| match ns {
|
.filter_map(move |ns| match ns {
|
||||||
Ok(ns) => Some(ns),
|
Ok(ns) => Some(ns),
|
||||||
@ -770,7 +770,7 @@ impl DataStore {
|
|||||||
) -> Result<impl Iterator<Item = BackupGroup> + 'static, Error> {
|
) -> Result<impl Iterator<Item = BackupGroup> + 'static, Error> {
|
||||||
let this = Arc::clone(self);
|
let this = Arc::clone(self);
|
||||||
Ok(
|
Ok(
|
||||||
ListGroups::new(Arc::clone(&self), ns)?.filter_map(move |group| match group {
|
ListGroups::new(Arc::clone(self), ns)?.filter_map(move |group| match group {
|
||||||
Ok(group) => Some(group),
|
Ok(group) => Some(group),
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
log::error!("list groups error on datastore {} - {}", this.name(), err);
|
log::error!("list groups error on datastore {} - {}", this.name(), err);
|
||||||
@ -865,11 +865,10 @@ impl DataStore {
|
|||||||
worker.fail_on_shutdown()?;
|
worker.fail_on_shutdown()?;
|
||||||
let digest = index.index_digest(pos).unwrap();
|
let digest = index.index_digest(pos).unwrap();
|
||||||
if !self.inner.chunk_store.cond_touch_chunk(digest, false)? {
|
if !self.inner.chunk_store.cond_touch_chunk(digest, false)? {
|
||||||
|
let hex = hex::encode(digest);
|
||||||
task_warn!(
|
task_warn!(
|
||||||
worker,
|
worker,
|
||||||
"warning: unable to access non-existent chunk {}, required by {:?}",
|
"warning: unable to access non-existent chunk {hex}, required by {file_name:?}"
|
||||||
hex::encode(digest),
|
|
||||||
file_name,
|
|
||||||
);
|
);
|
||||||
|
|
||||||
// touch any corresponding .bad files to keep them around, meaning if a chunk is
|
// touch any corresponding .bad files to keep them around, meaning if a chunk is
|
||||||
@ -1194,7 +1193,7 @@ impl DataStore {
|
|||||||
ns: BackupNamespace,
|
ns: BackupNamespace,
|
||||||
group: pbs_api_types::BackupGroup,
|
group: pbs_api_types::BackupGroup,
|
||||||
) -> BackupGroup {
|
) -> BackupGroup {
|
||||||
BackupGroup::new(Arc::clone(&self), ns, group)
|
BackupGroup::new(Arc::clone(self), ns, group)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Open a backup group from this datastore.
|
/// Open a backup group from this datastore.
|
||||||
|
@ -86,7 +86,7 @@ impl BackupManifest {
|
|||||||
pub fn new(snapshot: pbs_api_types::BackupDir) -> Self {
|
pub fn new(snapshot: pbs_api_types::BackupDir) -> Self {
|
||||||
Self {
|
Self {
|
||||||
backup_type: snapshot.group.ty,
|
backup_type: snapshot.group.ty,
|
||||||
backup_id: snapshot.group.id.into(),
|
backup_id: snapshot.group.id,
|
||||||
backup_time: snapshot.time,
|
backup_time: snapshot.time,
|
||||||
files: Vec::new(),
|
files: Vec::new(),
|
||||||
unprotected: json!({}),
|
unprotected: json!({}),
|
||||||
|
Loading…
Reference in New Issue
Block a user