reference the datastore in BackupGroup/Dir
And drop the base_path parameter on a first bunch of functions (more reordering will follow). Signed-off-by: Wolfgang Bumiller <w.bumiller@proxmox.com>
This commit is contained in:
parent
bb628c295a
commit
6da20161f0
@ -2,7 +2,7 @@ use std::path::PathBuf;
|
|||||||
|
|
||||||
use anyhow::{bail, Error};
|
use anyhow::{bail, Error};
|
||||||
|
|
||||||
use pbs_datastore::{ListGroups, ListSnapshots};
|
use pbs_datastore::DataStore;
|
||||||
|
|
||||||
fn run() -> Result<(), Error> {
|
fn run() -> Result<(), Error> {
|
||||||
let base: PathBuf = match std::env::args().skip(1).next() {
|
let base: PathBuf = match std::env::args().skip(1).next() {
|
||||||
@ -10,12 +10,13 @@ fn run() -> Result<(), Error> {
|
|||||||
None => bail!("no path passed"),
|
None => bail!("no path passed"),
|
||||||
};
|
};
|
||||||
|
|
||||||
for group in ListGroups::new(base.to_owned())? {
|
let store = unsafe { DataStore::open_path("", &base, None)? };
|
||||||
|
|
||||||
|
for group in store.iter_backup_groups()? {
|
||||||
let group = group?;
|
let group = group?;
|
||||||
println!("found group {}", group);
|
println!("found group {}", group);
|
||||||
|
|
||||||
let group_path = base.as_path().join(group.to_string());
|
for snapshot in group.iter_snapshots()? {
|
||||||
for snapshot in ListSnapshots::new(group, group_path)? {
|
|
||||||
println!("\t{}", snapshot?);
|
println!("\t{}", snapshot?);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,21 +1,35 @@
|
|||||||
|
use std::fmt;
|
||||||
use std::os::unix::io::RawFd;
|
use std::os::unix::io::RawFd;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::PathBuf;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
use anyhow::{bail, Error};
|
use anyhow::{bail, Error};
|
||||||
|
|
||||||
use pbs_api_types::{BackupType, GroupFilter, BACKUP_DATE_REGEX, BACKUP_FILE_REGEX};
|
use pbs_api_types::{BackupType, GroupFilter, BACKUP_DATE_REGEX, BACKUP_FILE_REGEX};
|
||||||
|
|
||||||
use super::manifest::MANIFEST_BLOB_NAME;
|
use crate::manifest::MANIFEST_BLOB_NAME;
|
||||||
|
use crate::DataStore;
|
||||||
|
|
||||||
/// BackupGroup is a directory containing a list of BackupDir
|
/// BackupGroup is a directory containing a list of BackupDir
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Clone)]
|
||||||
pub struct BackupGroup {
|
pub struct BackupGroup {
|
||||||
|
store: Arc<DataStore>,
|
||||||
|
|
||||||
group: pbs_api_types::BackupGroup,
|
group: pbs_api_types::BackupGroup,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl fmt::Debug for BackupGroup {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
||||||
|
f.debug_struct("BackupGroup")
|
||||||
|
.field("store", &self.store.name())
|
||||||
|
.field("group", &self.group)
|
||||||
|
.finish()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl BackupGroup {
|
impl BackupGroup {
|
||||||
pub(crate) fn new(group: pbs_api_types::BackupGroup) -> Self {
|
pub(crate) fn new(store: Arc<DataStore>, group: pbs_api_types::BackupGroup) -> Self {
|
||||||
Self { group }
|
Self { store, group }
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Access the underlying [`BackupGroup`](pbs_api_types::BackupGroup).
|
/// Access the underlying [`BackupGroup`](pbs_api_types::BackupGroup).
|
||||||
@ -32,13 +46,18 @@ impl BackupGroup {
|
|||||||
&self.group.id
|
&self.group.id
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn full_group_path(&self) -> PathBuf {
|
||||||
|
self.store.base_path().join(self.group.to_string())
|
||||||
|
}
|
||||||
|
|
||||||
pub fn relative_group_path(&self) -> PathBuf {
|
pub fn relative_group_path(&self) -> PathBuf {
|
||||||
self.group.to_string().into()
|
self.group.to_string().into()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn list_backups(&self, base_path: &Path) -> Result<Vec<BackupInfo>, Error> {
|
pub fn list_backups(&self) -> Result<Vec<BackupInfo>, Error> {
|
||||||
let mut list = vec![];
|
let mut list = vec![];
|
||||||
|
|
||||||
|
let base_path = self.store.base_path();
|
||||||
let mut path = base_path.to_owned();
|
let mut path = base_path.to_owned();
|
||||||
path.push(self.relative_group_path());
|
path.push(self.relative_group_path());
|
||||||
|
|
||||||
@ -54,7 +73,7 @@ impl BackupGroup {
|
|||||||
let backup_dir = self.backup_dir_with_rfc3339(backup_time)?;
|
let backup_dir = self.backup_dir_with_rfc3339(backup_time)?;
|
||||||
let files = list_backup_files(l2_fd, backup_time)?;
|
let files = list_backup_files(l2_fd, backup_time)?;
|
||||||
|
|
||||||
let protected = backup_dir.is_protected(base_path.to_owned());
|
let protected = backup_dir.is_protected();
|
||||||
|
|
||||||
list.push(BackupInfo {
|
list.push(BackupInfo {
|
||||||
backup_dir,
|
backup_dir,
|
||||||
@ -69,22 +88,18 @@ impl BackupGroup {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Finds the latest backup inside a backup group
|
/// Finds the latest backup inside a backup group
|
||||||
pub fn last_backup(
|
pub fn last_backup(&self, only_finished: bool) -> Result<Option<BackupInfo>, Error> {
|
||||||
&self,
|
let backups = self.list_backups()?;
|
||||||
base_path: &Path,
|
|
||||||
only_finished: bool,
|
|
||||||
) -> Result<Option<BackupInfo>, Error> {
|
|
||||||
let backups = self.list_backups(base_path)?;
|
|
||||||
Ok(backups
|
Ok(backups
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.filter(|item| !only_finished || item.is_finished())
|
.filter(|item| !only_finished || item.is_finished())
|
||||||
.max_by_key(|item| item.backup_dir.backup_time()))
|
.max_by_key(|item| item.backup_dir.backup_time()))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn last_successful_backup(&self, base_path: &Path) -> Result<Option<i64>, Error> {
|
pub fn last_successful_backup(&self) -> Result<Option<i64>, Error> {
|
||||||
let mut last = None;
|
let mut last = None;
|
||||||
|
|
||||||
let mut path = base_path.to_owned();
|
let mut path = self.store.base_path();
|
||||||
path.push(self.relative_group_path());
|
path.push(self.relative_group_path());
|
||||||
|
|
||||||
proxmox_sys::fs::scandir(
|
proxmox_sys::fs::scandir(
|
||||||
@ -149,6 +164,10 @@ impl BackupGroup {
|
|||||||
) -> Result<BackupDir, Error> {
|
) -> Result<BackupDir, Error> {
|
||||||
BackupDir::with_rfc3339(self.clone(), time_string.into())
|
BackupDir::with_rfc3339(self.clone(), time_string.into())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn iter_snapshots(&self) -> Result<crate::ListSnapshots, Error> {
|
||||||
|
crate::ListSnapshots::new(self.clone())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl AsRef<pbs_api_types::BackupGroup> for BackupGroup {
|
impl AsRef<pbs_api_types::BackupGroup> for BackupGroup {
|
||||||
@ -181,6 +200,7 @@ impl std::fmt::Display for BackupGroup {
|
|||||||
impl From<BackupDir> for BackupGroup {
|
impl From<BackupDir> for BackupGroup {
|
||||||
fn from(dir: BackupDir) -> BackupGroup {
|
fn from(dir: BackupDir) -> BackupGroup {
|
||||||
BackupGroup {
|
BackupGroup {
|
||||||
|
store: dir.store,
|
||||||
group: dir.dir.group,
|
group: dir.dir.group,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -189,6 +209,7 @@ impl From<BackupDir> for BackupGroup {
|
|||||||
impl From<&BackupDir> for BackupGroup {
|
impl From<&BackupDir> for BackupGroup {
|
||||||
fn from(dir: &BackupDir) -> BackupGroup {
|
fn from(dir: &BackupDir) -> BackupGroup {
|
||||||
BackupGroup {
|
BackupGroup {
|
||||||
|
store: Arc::clone(&dir.store),
|
||||||
group: dir.dir.group.clone(),
|
group: dir.dir.group.clone(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -197,18 +218,30 @@ impl From<&BackupDir> for BackupGroup {
|
|||||||
/// Uniquely identify a Backup (relative to data store)
|
/// Uniquely identify a Backup (relative to data store)
|
||||||
///
|
///
|
||||||
/// We also call this a backup snaphost.
|
/// We also call this a backup snaphost.
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Clone)]
|
||||||
pub struct BackupDir {
|
pub struct BackupDir {
|
||||||
|
store: Arc<DataStore>,
|
||||||
dir: pbs_api_types::BackupDir,
|
dir: pbs_api_types::BackupDir,
|
||||||
// backup_time as rfc3339
|
// backup_time as rfc3339
|
||||||
backup_time_string: String,
|
backup_time_string: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl fmt::Debug for BackupDir {
|
||||||
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
||||||
|
f.debug_struct("BackupDir")
|
||||||
|
.field("store", &self.store.name())
|
||||||
|
.field("dir", &self.dir)
|
||||||
|
.field("backup_time_string", &self.backup_time_string)
|
||||||
|
.finish()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl BackupDir {
|
impl BackupDir {
|
||||||
/// Temporarily used for tests.
|
/// Temporarily used for tests.
|
||||||
#[doc(hidden)]
|
#[doc(hidden)]
|
||||||
pub fn new_test(dir: pbs_api_types::BackupDir) -> Self {
|
pub fn new_test(dir: pbs_api_types::BackupDir) -> Self {
|
||||||
Self {
|
Self {
|
||||||
|
store: unsafe { DataStore::new_test() },
|
||||||
backup_time_string: Self::backup_time_to_string(dir.time).unwrap(),
|
backup_time_string: Self::backup_time_to_string(dir.time).unwrap(),
|
||||||
dir,
|
dir,
|
||||||
}
|
}
|
||||||
@ -217,6 +250,7 @@ impl BackupDir {
|
|||||||
pub(crate) fn with_group(group: BackupGroup, backup_time: i64) -> Result<Self, Error> {
|
pub(crate) fn with_group(group: BackupGroup, backup_time: i64) -> Result<Self, Error> {
|
||||||
let backup_time_string = Self::backup_time_to_string(backup_time)?;
|
let backup_time_string = Self::backup_time_to_string(backup_time)?;
|
||||||
Ok(Self {
|
Ok(Self {
|
||||||
|
store: group.store,
|
||||||
dir: (group.group, backup_time).into(),
|
dir: (group.group, backup_time).into(),
|
||||||
backup_time_string,
|
backup_time_string,
|
||||||
})
|
})
|
||||||
@ -228,6 +262,7 @@ impl BackupDir {
|
|||||||
) -> Result<Self, Error> {
|
) -> Result<Self, Error> {
|
||||||
let backup_time = proxmox_time::parse_rfc3339(&backup_time_string)?;
|
let backup_time = proxmox_time::parse_rfc3339(&backup_time_string)?;
|
||||||
Ok(Self {
|
Ok(Self {
|
||||||
|
store: group.store,
|
||||||
dir: (group.group, backup_time).into(),
|
dir: (group.group, backup_time).into(),
|
||||||
backup_time_string,
|
backup_time_string,
|
||||||
})
|
})
|
||||||
@ -257,21 +292,23 @@ impl BackupDir {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Returns the absolute path for backup_dir, using the cached formatted time string.
|
/// Returns the absolute path for backup_dir, using the cached formatted time string.
|
||||||
pub fn full_path(&self, mut base_path: PathBuf) -> PathBuf {
|
pub fn full_path(&self) -> PathBuf {
|
||||||
|
let mut base_path = self.store.base_path();
|
||||||
base_path.push(self.dir.group.ty.as_str());
|
base_path.push(self.dir.group.ty.as_str());
|
||||||
base_path.push(&self.dir.group.id);
|
base_path.push(&self.dir.group.id);
|
||||||
base_path.push(&self.backup_time_string);
|
base_path.push(&self.backup_time_string);
|
||||||
base_path
|
base_path
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn protected_file(&self, mut path: PathBuf) -> PathBuf {
|
pub fn protected_file(&self) -> PathBuf {
|
||||||
|
let mut path = self.store.base_path();
|
||||||
path.push(self.relative_path());
|
path.push(self.relative_path());
|
||||||
path.push(".protected");
|
path.push(".protected");
|
||||||
path
|
path
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn is_protected(&self, base_path: PathBuf) -> bool {
|
pub fn is_protected(&self) -> bool {
|
||||||
let path = self.protected_file(base_path);
|
let path = self.protected_file();
|
||||||
path.exists()
|
path.exists()
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -324,7 +361,7 @@ impl std::fmt::Display for BackupDir {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Detailed Backup Information, lists files inside a BackupDir
|
/// Detailed Backup Information, lists files inside a BackupDir
|
||||||
#[derive(Debug, Clone)]
|
#[derive(Clone, Debug)]
|
||||||
pub struct BackupInfo {
|
pub struct BackupInfo {
|
||||||
/// the backup directory
|
/// the backup directory
|
||||||
pub backup_dir: BackupDir,
|
pub backup_dir: BackupDir,
|
||||||
@ -335,12 +372,13 @@ pub struct BackupInfo {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl BackupInfo {
|
impl BackupInfo {
|
||||||
pub fn new(base_path: &Path, backup_dir: BackupDir) -> Result<BackupInfo, Error> {
|
pub fn new(backup_dir: BackupDir) -> Result<BackupInfo, Error> {
|
||||||
let mut path = base_path.to_owned();
|
let base_path = backup_dir.store.base_path();
|
||||||
|
let mut path = base_path.clone();
|
||||||
path.push(backup_dir.relative_path());
|
path.push(backup_dir.relative_path());
|
||||||
|
|
||||||
let files = list_backup_files(libc::AT_FDCWD, &path)?;
|
let files = list_backup_files(libc::AT_FDCWD, &path)?;
|
||||||
let protected = backup_dir.is_protected(base_path.to_owned());
|
let protected = backup_dir.is_protected();
|
||||||
|
|
||||||
Ok(BackupInfo {
|
Ok(BackupInfo {
|
||||||
backup_dir,
|
backup_dir,
|
||||||
|
@ -21,7 +21,7 @@ pub struct ChunkStore {
|
|||||||
pub(crate) base: PathBuf,
|
pub(crate) base: PathBuf,
|
||||||
chunk_dir: PathBuf,
|
chunk_dir: PathBuf,
|
||||||
mutex: Mutex<()>,
|
mutex: Mutex<()>,
|
||||||
locker: Arc<Mutex<ProcessLocker>>,
|
locker: Option<Arc<Mutex<ProcessLocker>>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: what about sysctl setting vm.vfs_cache_pressure (0 - 100) ?
|
// TODO: what about sysctl setting vm.vfs_cache_pressure (0 - 100) ?
|
||||||
@ -60,6 +60,17 @@ fn digest_to_prefix(digest: &[u8]) -> PathBuf {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl ChunkStore {
|
impl ChunkStore {
|
||||||
|
#[doc(hidden)]
|
||||||
|
pub unsafe fn panic_store() -> Self {
|
||||||
|
Self {
|
||||||
|
name: String::new(),
|
||||||
|
base: PathBuf::new(),
|
||||||
|
chunk_dir: PathBuf::new(),
|
||||||
|
mutex: Mutex::new(()),
|
||||||
|
locker: None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
fn chunk_dir<P: AsRef<Path>>(path: P) -> PathBuf {
|
fn chunk_dir<P: AsRef<Path>>(path: P) -> PathBuf {
|
||||||
let mut chunk_dir: PathBuf = PathBuf::from(path.as_ref());
|
let mut chunk_dir: PathBuf = PathBuf::from(path.as_ref());
|
||||||
chunk_dir.push(".chunks");
|
chunk_dir.push(".chunks");
|
||||||
@ -180,12 +191,15 @@ impl ChunkStore {
|
|||||||
name: name.to_owned(),
|
name: name.to_owned(),
|
||||||
base,
|
base,
|
||||||
chunk_dir,
|
chunk_dir,
|
||||||
locker,
|
locker: Some(locker),
|
||||||
mutex: Mutex::new(()),
|
mutex: Mutex::new(()),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn touch_chunk(&self, digest: &[u8; 32]) -> Result<(), Error> {
|
pub fn touch_chunk(&self, digest: &[u8; 32]) -> Result<(), Error> {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
self.cond_touch_chunk(digest, true)?;
|
self.cond_touch_chunk(digest, true)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@ -195,11 +209,17 @@ impl ChunkStore {
|
|||||||
digest: &[u8; 32],
|
digest: &[u8; 32],
|
||||||
fail_if_not_exist: bool,
|
fail_if_not_exist: bool,
|
||||||
) -> Result<bool, Error> {
|
) -> Result<bool, Error> {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
let (chunk_path, _digest_str) = self.chunk_path(digest);
|
let (chunk_path, _digest_str) = self.chunk_path(digest);
|
||||||
self.cond_touch_path(&chunk_path, fail_if_not_exist)
|
self.cond_touch_path(&chunk_path, fail_if_not_exist)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn cond_touch_path(&self, path: &Path, fail_if_not_exist: bool) -> Result<bool, Error> {
|
pub fn cond_touch_path(&self, path: &Path, fail_if_not_exist: bool) -> Result<bool, Error> {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
const UTIME_NOW: i64 = (1 << 30) - 1;
|
const UTIME_NOW: i64 = (1 << 30) - 1;
|
||||||
const UTIME_OMIT: i64 = (1 << 30) - 2;
|
const UTIME_OMIT: i64 = (1 << 30) - 2;
|
||||||
|
|
||||||
@ -239,6 +259,9 @@ impl ChunkStore {
|
|||||||
+ std::iter::FusedIterator,
|
+ std::iter::FusedIterator,
|
||||||
Error,
|
Error,
|
||||||
> {
|
> {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
use nix::dir::Dir;
|
use nix::dir::Dir;
|
||||||
use nix::fcntl::OFlag;
|
use nix::fcntl::OFlag;
|
||||||
use nix::sys::stat::Mode;
|
use nix::sys::stat::Mode;
|
||||||
@ -325,7 +348,8 @@ impl ChunkStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn oldest_writer(&self) -> Option<i64> {
|
pub fn oldest_writer(&self) -> Option<i64> {
|
||||||
ProcessLocker::oldest_shared_lock(self.locker.clone())
|
// unwrap: only `None` in unit tests
|
||||||
|
ProcessLocker::oldest_shared_lock(self.locker.clone().unwrap())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn sweep_unused_chunks(
|
pub fn sweep_unused_chunks(
|
||||||
@ -335,6 +359,9 @@ impl ChunkStore {
|
|||||||
status: &mut GarbageCollectionStatus,
|
status: &mut GarbageCollectionStatus,
|
||||||
worker: &dyn WorkerTaskContext,
|
worker: &dyn WorkerTaskContext,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
use nix::sys::stat::fstatat;
|
use nix::sys::stat::fstatat;
|
||||||
use nix::unistd::{unlinkat, UnlinkatFlags};
|
use nix::unistd::{unlinkat, UnlinkatFlags};
|
||||||
|
|
||||||
@ -426,6 +453,9 @@ impl ChunkStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn insert_chunk(&self, chunk: &DataBlob, digest: &[u8; 32]) -> Result<(bool, u64), Error> {
|
pub fn insert_chunk(&self, chunk: &DataBlob, digest: &[u8; 32]) -> Result<(bool, u64), Error> {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
//println!("DIGEST {}", hex::encode(digest));
|
//println!("DIGEST {}", hex::encode(digest));
|
||||||
|
|
||||||
let (chunk_path, digest_str) = self.chunk_path(digest);
|
let (chunk_path, digest_str) = self.chunk_path(digest);
|
||||||
@ -485,6 +515,9 @@ impl ChunkStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn chunk_path(&self, digest: &[u8; 32]) -> (PathBuf, String) {
|
pub fn chunk_path(&self, digest: &[u8; 32]) -> (PathBuf, String) {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
let mut chunk_path = self.chunk_dir.clone();
|
let mut chunk_path = self.chunk_dir.clone();
|
||||||
let prefix = digest_to_prefix(digest);
|
let prefix = digest_to_prefix(digest);
|
||||||
chunk_path.push(&prefix);
|
chunk_path.push(&prefix);
|
||||||
@ -494,6 +527,9 @@ impl ChunkStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn relative_path(&self, path: &Path) -> PathBuf {
|
pub fn relative_path(&self, path: &Path) -> PathBuf {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
let mut full_path = self.base.clone();
|
let mut full_path = self.base.clone();
|
||||||
full_path.push(path);
|
full_path.push(path);
|
||||||
full_path
|
full_path
|
||||||
@ -504,15 +540,20 @@ impl ChunkStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn base_path(&self) -> PathBuf {
|
pub fn base_path(&self) -> PathBuf {
|
||||||
|
// unwrap: only `None` in unit tests
|
||||||
|
assert!(self.locker.is_some());
|
||||||
|
|
||||||
self.base.clone()
|
self.base.clone()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn try_shared_lock(&self) -> Result<ProcessLockSharedGuard, Error> {
|
pub fn try_shared_lock(&self) -> Result<ProcessLockSharedGuard, Error> {
|
||||||
ProcessLocker::try_shared_lock(self.locker.clone())
|
// unwrap: only `None` in unit tests
|
||||||
|
ProcessLocker::try_shared_lock(self.locker.clone().unwrap())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn try_exclusive_lock(&self) -> Result<ProcessLockExclusiveGuard, Error> {
|
pub fn try_exclusive_lock(&self) -> Result<ProcessLockExclusiveGuard, Error> {
|
||||||
ProcessLocker::try_exclusive_lock(self.locker.clone())
|
// unwrap: only `None` in unit tests
|
||||||
|
ProcessLocker::try_exclusive_lock(self.locker.clone().unwrap())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -65,6 +65,22 @@ pub struct DataStoreImpl {
|
|||||||
last_update: i64,
|
last_update: i64,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl DataStoreImpl {
|
||||||
|
// This one just panics on everything
|
||||||
|
#[doc(hidden)]
|
||||||
|
pub unsafe fn new_test() -> Arc<Self> {
|
||||||
|
Arc::new(Self {
|
||||||
|
chunk_store: Arc::new(unsafe { ChunkStore::panic_store() }),
|
||||||
|
gc_mutex: Mutex::new(()),
|
||||||
|
last_gc_status: Mutex::new(GarbageCollectionStatus::default()),
|
||||||
|
verify_new: false,
|
||||||
|
chunk_order: ChunkOrder::None,
|
||||||
|
last_generation: 0,
|
||||||
|
last_update: 0,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub struct DataStore {
|
pub struct DataStore {
|
||||||
inner: Arc<DataStoreImpl>,
|
inner: Arc<DataStoreImpl>,
|
||||||
operation: Option<Operation>,
|
operation: Option<Operation>,
|
||||||
@ -98,6 +114,15 @@ impl Drop for DataStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl DataStore {
|
impl DataStore {
|
||||||
|
// This one just panics on everything
|
||||||
|
#[doc(hidden)]
|
||||||
|
pub unsafe fn new_test() -> Arc<Self> {
|
||||||
|
Arc::new(Self {
|
||||||
|
inner: unsafe { DataStoreImpl::new_test() },
|
||||||
|
operation: None,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
pub fn lookup_datastore(
|
pub fn lookup_datastore(
|
||||||
name: &str,
|
name: &str,
|
||||||
operation: Option<Operation>,
|
operation: Option<Operation>,
|
||||||
@ -108,7 +133,6 @@ impl DataStore {
|
|||||||
|
|
||||||
let (config, _digest) = pbs_config::datastore::config()?;
|
let (config, _digest) = pbs_config::datastore::config()?;
|
||||||
let config: DataStoreConfig = config.lookup("datastore", name)?;
|
let config: DataStoreConfig = config.lookup("datastore", name)?;
|
||||||
let path = PathBuf::from(&config.path);
|
|
||||||
|
|
||||||
if let Some(maintenance_mode) = config.get_maintenance_mode() {
|
if let Some(maintenance_mode) = config.get_maintenance_mode() {
|
||||||
if let Err(error) = maintenance_mode.check(operation) {
|
if let Err(error) = maintenance_mode.check(operation) {
|
||||||
@ -132,7 +156,8 @@ impl DataStore {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let datastore = DataStore::open_with_path(name, &path, config, generation, now)?;
|
let chunk_store = ChunkStore::open(name, &config.path)?;
|
||||||
|
let datastore = DataStore::with_store_and_config(chunk_store, config, generation, now)?;
|
||||||
|
|
||||||
let datastore = Arc::new(datastore);
|
let datastore = Arc::new(datastore);
|
||||||
map.insert(name.to_string(), datastore.clone());
|
map.insert(name.to_string(), datastore.clone());
|
||||||
@ -153,15 +178,43 @@ impl DataStore {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn open_with_path(
|
/// Open a raw database given a name and a path.
|
||||||
store_name: &str,
|
pub unsafe fn open_path(
|
||||||
path: &Path,
|
name: &str,
|
||||||
|
path: impl AsRef<Path>,
|
||||||
|
operation: Option<Operation>,
|
||||||
|
) -> Result<Arc<Self>, Error> {
|
||||||
|
let path = path
|
||||||
|
.as_ref()
|
||||||
|
.to_str()
|
||||||
|
.ok_or_else(|| format_err!("non-utf8 paths not supported"))?
|
||||||
|
.to_owned();
|
||||||
|
unsafe { Self::open_from_config(DataStoreConfig::new(name.to_owned(), path), operation) }
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Open a datastore given a raw configuration.
|
||||||
|
pub unsafe fn open_from_config(
|
||||||
|
config: DataStoreConfig,
|
||||||
|
operation: Option<Operation>,
|
||||||
|
) -> Result<Arc<Self>, Error> {
|
||||||
|
let name = config.name.clone();
|
||||||
|
|
||||||
|
let chunk_store = ChunkStore::open(&name, &config.path)?;
|
||||||
|
let inner = Arc::new(Self::with_store_and_config(chunk_store, config, 0, 0)?);
|
||||||
|
|
||||||
|
if let Some(operation) = operation {
|
||||||
|
update_active_operations(&name, operation, 1)?;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(Arc::new(Self { inner, operation }))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn with_store_and_config(
|
||||||
|
chunk_store: ChunkStore,
|
||||||
config: DataStoreConfig,
|
config: DataStoreConfig,
|
||||||
last_generation: usize,
|
last_generation: usize,
|
||||||
last_update: i64,
|
last_update: i64,
|
||||||
) -> Result<DataStoreImpl, Error> {
|
) -> Result<DataStoreImpl, Error> {
|
||||||
let chunk_store = ChunkStore::open(store_name, path)?;
|
|
||||||
|
|
||||||
let mut gc_status_path = chunk_store.base_path();
|
let mut gc_status_path = chunk_store.base_path();
|
||||||
gc_status_path.push(".gc-status");
|
gc_status_path.push(".gc-status");
|
||||||
|
|
||||||
@ -363,7 +416,7 @@ impl DataStore {
|
|||||||
/// Remove a complete backup group including all snapshots, returns true
|
/// Remove a complete backup group including all snapshots, returns true
|
||||||
/// if all snapshots were removed, and false if some were protected
|
/// if all snapshots were removed, and false if some were protected
|
||||||
pub fn remove_backup_group(
|
pub fn remove_backup_group(
|
||||||
&self,
|
self: &Arc<Self>,
|
||||||
backup_group: &pbs_api_types::BackupGroup,
|
backup_group: &pbs_api_types::BackupGroup,
|
||||||
) -> Result<bool, Error> {
|
) -> Result<bool, Error> {
|
||||||
let backup_group = self.backup_group(backup_group.clone());
|
let backup_group = self.backup_group(backup_group.clone());
|
||||||
@ -381,8 +434,8 @@ impl DataStore {
|
|||||||
let mut removed_all = true;
|
let mut removed_all = true;
|
||||||
|
|
||||||
// remove all individual backup dirs first to ensure nothing is using them
|
// remove all individual backup dirs first to ensure nothing is using them
|
||||||
for snap in backup_group.list_backups(&self.base_path())? {
|
for snap in backup_group.list_backups()? {
|
||||||
if snap.backup_dir.is_protected(self.base_path()) {
|
if snap.backup_dir.is_protected() {
|
||||||
removed_all = false;
|
removed_all = false;
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
@ -405,13 +458,13 @@ impl DataStore {
|
|||||||
|
|
||||||
/// Remove a backup directory including all content
|
/// Remove a backup directory including all content
|
||||||
pub fn remove_backup_dir(
|
pub fn remove_backup_dir(
|
||||||
&self,
|
self: &Arc<Self>,
|
||||||
backup_dir: &pbs_api_types::BackupDir,
|
backup_dir: &pbs_api_types::BackupDir,
|
||||||
force: bool,
|
force: bool,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let backup_dir = self.backup_dir(backup_dir.clone())?;
|
let backup_dir = self.backup_dir(backup_dir.clone())?;
|
||||||
|
|
||||||
let full_path = backup_dir.full_path(self.base_path());
|
let full_path = backup_dir.full_path();
|
||||||
|
|
||||||
let (_guard, _manifest_guard);
|
let (_guard, _manifest_guard);
|
||||||
if !force {
|
if !force {
|
||||||
@ -419,7 +472,7 @@ impl DataStore {
|
|||||||
_manifest_guard = self.lock_manifest(&backup_dir)?;
|
_manifest_guard = self.lock_manifest(&backup_dir)?;
|
||||||
}
|
}
|
||||||
|
|
||||||
if backup_dir.is_protected(self.base_path()) {
|
if backup_dir.is_protected() {
|
||||||
bail!("cannot remove protected snapshot");
|
bail!("cannot remove protected snapshot");
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -441,7 +494,7 @@ impl DataStore {
|
|||||||
///
|
///
|
||||||
/// Or None if there is no backup in the group (or the group dir does not exist).
|
/// Or None if there is no backup in the group (or the group dir does not exist).
|
||||||
pub fn last_successful_backup(
|
pub fn last_successful_backup(
|
||||||
&self,
|
self: &Arc<Self>,
|
||||||
backup_group: &pbs_api_types::BackupGroup,
|
backup_group: &pbs_api_types::BackupGroup,
|
||||||
) -> Result<Option<i64>, Error> {
|
) -> Result<Option<i64>, Error> {
|
||||||
let backup_group = self.backup_group(backup_group.clone());
|
let backup_group = self.backup_group(backup_group.clone());
|
||||||
@ -451,7 +504,7 @@ impl DataStore {
|
|||||||
group_path.push(backup_group.relative_group_path());
|
group_path.push(backup_group.relative_group_path());
|
||||||
|
|
||||||
if group_path.exists() {
|
if group_path.exists() {
|
||||||
backup_group.last_successful_backup(&base_path)
|
backup_group.last_successful_backup()
|
||||||
} else {
|
} else {
|
||||||
Ok(None)
|
Ok(None)
|
||||||
}
|
}
|
||||||
@ -584,20 +637,23 @@ impl DataStore {
|
|||||||
///
|
///
|
||||||
/// The iterated item is still a Result that can contain errors from rather unexptected FS or
|
/// The iterated item is still a Result that can contain errors from rather unexptected FS or
|
||||||
/// parsing errors.
|
/// parsing errors.
|
||||||
pub fn iter_backup_groups(&self) -> Result<ListGroups, Error> {
|
pub fn iter_backup_groups(self: &Arc<DataStore>) -> Result<ListGroups, Error> {
|
||||||
ListGroups::new(self.base_path())
|
ListGroups::new(Arc::clone(self))
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get a streaming iter over top-level backup groups of a datatstore, filtered by Ok results
|
/// Get a streaming iter over top-level backup groups of a datatstore, filtered by Ok results
|
||||||
///
|
///
|
||||||
/// The iterated item's result is already unwrapped, if it contained an error it will be
|
/// The iterated item's result is already unwrapped, if it contained an error it will be
|
||||||
/// logged. Can be useful in iterator chain commands
|
/// logged. Can be useful in iterator chain commands
|
||||||
pub fn iter_backup_groups_ok(&self) -> Result<impl Iterator<Item = BackupGroup> + '_, Error> {
|
pub fn iter_backup_groups_ok(
|
||||||
|
self: &Arc<DataStore>,
|
||||||
|
) -> Result<impl Iterator<Item = BackupGroup> + 'static, Error> {
|
||||||
|
let this = Arc::clone(self);
|
||||||
Ok(
|
Ok(
|
||||||
ListGroups::new(self.base_path())?.filter_map(move |group| match group {
|
ListGroups::new(Arc::clone(&self))?.filter_map(move |group| match group {
|
||||||
Ok(group) => Some(group),
|
Ok(group) => Some(group),
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
log::error!("list groups error on datastore {} - {}", self.name(), err);
|
log::error!("list groups error on datastore {} - {}", this.name(), err);
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
}),
|
}),
|
||||||
@ -607,8 +663,8 @@ impl DataStore {
|
|||||||
/// Get a in-memory vector for all top-level backup groups of a datatstore
|
/// Get a in-memory vector for all top-level backup groups of a datatstore
|
||||||
///
|
///
|
||||||
/// NOTE: using the iterator directly is most often more efficient w.r.t. memory usage
|
/// NOTE: using the iterator directly is most often more efficient w.r.t. memory usage
|
||||||
pub fn list_backup_groups(&self) -> Result<Vec<BackupGroup>, Error> {
|
pub fn list_backup_groups(self: &Arc<DataStore>) -> Result<Vec<BackupGroup>, Error> {
|
||||||
ListGroups::new(self.base_path())?.collect()
|
ListGroups::new(Arc::clone(self))?.collect()
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn list_images(&self) -> Result<Vec<PathBuf>, Error> {
|
pub fn list_images(&self) -> Result<Vec<PathBuf>, Error> {
|
||||||
@ -1022,11 +1078,11 @@ impl DataStore {
|
|||||||
|
|
||||||
/// Updates the protection status of the specified snapshot.
|
/// Updates the protection status of the specified snapshot.
|
||||||
pub fn update_protection(&self, backup_dir: &BackupDir, protection: bool) -> Result<(), Error> {
|
pub fn update_protection(&self, backup_dir: &BackupDir, protection: bool) -> Result<(), Error> {
|
||||||
let full_path = backup_dir.full_path(self.base_path());
|
let full_path = backup_dir.full_path();
|
||||||
|
|
||||||
let _guard = lock_dir_noblock(&full_path, "snapshot", "possibly running or in use")?;
|
let _guard = lock_dir_noblock(&full_path, "snapshot", "possibly running or in use")?;
|
||||||
|
|
||||||
let protected_path = backup_dir.protected_file(self.base_path());
|
let protected_path = backup_dir.protected_file();
|
||||||
if protection {
|
if protection {
|
||||||
std::fs::File::create(protected_path)
|
std::fs::File::create(protected_path)
|
||||||
.map_err(|err| format_err!("could not create protection file: {}", err))?;
|
.map_err(|err| format_err!("could not create protection file: {}", err))?;
|
||||||
@ -1093,12 +1149,12 @@ impl DataStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Open a backup group from this datastore.
|
/// Open a backup group from this datastore.
|
||||||
pub fn backup_group(&self, group: pbs_api_types::BackupGroup) -> BackupGroup {
|
pub fn backup_group(self: &Arc<Self>, group: pbs_api_types::BackupGroup) -> BackupGroup {
|
||||||
BackupGroup::new(group)
|
BackupGroup::new(Arc::clone(&self), group)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Open a backup group from this datastore.
|
/// Open a backup group from this datastore.
|
||||||
pub fn backup_group_from_parts<T>(&self, ty: BackupType, id: T) -> BackupGroup
|
pub fn backup_group_from_parts<T>(self: &Arc<Self>, ty: BackupType, id: T) -> BackupGroup
|
||||||
where
|
where
|
||||||
T: Into<String>,
|
T: Into<String>,
|
||||||
{
|
{
|
||||||
@ -1108,18 +1164,18 @@ impl DataStore {
|
|||||||
/// Open a backup group from this datastore by backup group path such as `vm/100`.
|
/// Open a backup group from this datastore by backup group path such as `vm/100`.
|
||||||
///
|
///
|
||||||
/// Convenience method for `store.backup_group(path.parse()?)`
|
/// Convenience method for `store.backup_group(path.parse()?)`
|
||||||
pub fn backup_group_from_path(&self, path: &str) -> Result<BackupGroup, Error> {
|
pub fn backup_group_from_path(self: &Arc<Self>, path: &str) -> Result<BackupGroup, Error> {
|
||||||
Ok(self.backup_group(path.parse()?))
|
Ok(self.backup_group(path.parse()?))
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Open a snapshot (backup directory) from this datastore.
|
/// Open a snapshot (backup directory) from this datastore.
|
||||||
pub fn backup_dir(&self, dir: pbs_api_types::BackupDir) -> Result<BackupDir, Error> {
|
pub fn backup_dir(self: &Arc<Self>, dir: pbs_api_types::BackupDir) -> Result<BackupDir, Error> {
|
||||||
BackupDir::with_group(self.backup_group(dir.group), dir.time)
|
BackupDir::with_group(self.backup_group(dir.group), dir.time)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Open a snapshot (backup directory) from this datastore.
|
/// Open a snapshot (backup directory) from this datastore.
|
||||||
pub fn backup_dir_from_parts<T>(
|
pub fn backup_dir_from_parts<T>(
|
||||||
&self,
|
self: &Arc<Self>,
|
||||||
ty: BackupType,
|
ty: BackupType,
|
||||||
id: T,
|
id: T,
|
||||||
time: i64,
|
time: i64,
|
||||||
@ -1132,7 +1188,7 @@ impl DataStore {
|
|||||||
|
|
||||||
/// Open a snapshot (backup directory) from this datastore with a cached rfc3339 time string.
|
/// Open a snapshot (backup directory) from this datastore with a cached rfc3339 time string.
|
||||||
pub fn backup_dir_with_rfc3339<T: Into<String>>(
|
pub fn backup_dir_with_rfc3339<T: Into<String>>(
|
||||||
&self,
|
self: &Arc<Self>,
|
||||||
group: BackupGroup,
|
group: BackupGroup,
|
||||||
time_string: T,
|
time_string: T,
|
||||||
) -> Result<BackupDir, Error> {
|
) -> Result<BackupDir, Error> {
|
||||||
@ -1140,7 +1196,7 @@ impl DataStore {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Open a snapshot (backup directory) from this datastore by a snapshot path.
|
/// Open a snapshot (backup directory) from this datastore by a snapshot path.
|
||||||
pub fn backup_dir_from_path(&self, path: &str) -> Result<BackupDir, Error> {
|
pub fn backup_dir_from_path(self: &Arc<Self>, path: &str) -> Result<BackupDir, Error> {
|
||||||
self.backup_dir(path.parse()?)
|
self.backup_dir(path.parse()?)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1152,9 +1208,9 @@ pub struct ListSnapshots {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl ListSnapshots {
|
impl ListSnapshots {
|
||||||
pub fn new(group: BackupGroup, group_path: PathBuf) -> Result<Self, Error> {
|
pub fn new(group: BackupGroup) -> Result<Self, Error> {
|
||||||
Ok(ListSnapshots {
|
Ok(ListSnapshots {
|
||||||
fd: proxmox_sys::fs::read_subdir(libc::AT_FDCWD, &group_path)?,
|
fd: proxmox_sys::fs::read_subdir(libc::AT_FDCWD, &group.full_group_path())?,
|
||||||
group,
|
group,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@ -1192,14 +1248,16 @@ impl Iterator for ListSnapshots {
|
|||||||
|
|
||||||
/// A iterator for a (single) level of Backup Groups
|
/// A iterator for a (single) level of Backup Groups
|
||||||
pub struct ListGroups {
|
pub struct ListGroups {
|
||||||
|
store: Arc<DataStore>,
|
||||||
type_fd: proxmox_sys::fs::ReadDir,
|
type_fd: proxmox_sys::fs::ReadDir,
|
||||||
id_state: Option<(BackupType, proxmox_sys::fs::ReadDir)>,
|
id_state: Option<(BackupType, proxmox_sys::fs::ReadDir)>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl ListGroups {
|
impl ListGroups {
|
||||||
pub fn new(base_path: PathBuf) -> Result<Self, Error> {
|
pub fn new(store: Arc<DataStore>) -> Result<Self, Error> {
|
||||||
Ok(ListGroups {
|
Ok(ListGroups {
|
||||||
type_fd: proxmox_sys::fs::read_subdir(libc::AT_FDCWD, &base_path)?,
|
type_fd: proxmox_sys::fs::read_subdir(libc::AT_FDCWD, &store.base_path())?,
|
||||||
|
store,
|
||||||
id_state: None,
|
id_state: None,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@ -1227,6 +1285,7 @@ impl Iterator for ListGroups {
|
|||||||
}
|
}
|
||||||
if BACKUP_ID_REGEX.is_match(name) {
|
if BACKUP_ID_REGEX.is_match(name) {
|
||||||
return Some(Ok(BackupGroup::new(
|
return Some(Ok(BackupGroup::new(
|
||||||
|
Arc::clone(&self.store),
|
||||||
(group_type, name.to_owned()).into(),
|
(group_type, name.to_owned()).into(),
|
||||||
)));
|
)));
|
||||||
}
|
}
|
||||||
|
@ -142,6 +142,8 @@
|
|||||||
//! * / = no interaction
|
//! * / = no interaction
|
||||||
//! * shared/exclusive from POV of 'starting' process
|
//! * shared/exclusive from POV of 'starting' process
|
||||||
|
|
||||||
|
#![deny(unsafe_op_in_unsafe_fn)]
|
||||||
|
|
||||||
// Note: .pcat1 => Proxmox Catalog Format version 1
|
// Note: .pcat1 => Proxmox Catalog Format version 1
|
||||||
pub const CATALOG_NAME: &str = "catalog.pcat1.didx";
|
pub const CATALOG_NAME: &str = "catalog.pcat1.didx";
|
||||||
|
|
||||||
|
@ -34,7 +34,7 @@ impl SnapshotReader {
|
|||||||
) -> Result<Self, Error> {
|
) -> Result<Self, Error> {
|
||||||
let snapshot = datastore.backup_dir(snapshot)?;
|
let snapshot = datastore.backup_dir(snapshot)?;
|
||||||
|
|
||||||
let snapshot_path = snapshot.full_path(datastore.base_path());
|
let snapshot_path = snapshot.full_path();
|
||||||
|
|
||||||
let locked_dir =
|
let locked_dir =
|
||||||
lock_dir_noblock_shared(&snapshot_path, "snapshot", "locked by another operation")?;
|
lock_dir_noblock_shared(&snapshot_path, "snapshot", "locked by another operation")?;
|
||||||
|
@ -4,6 +4,7 @@ use std::collections::HashSet;
|
|||||||
use std::ffi::OsStr;
|
use std::ffi::OsStr;
|
||||||
use std::os::unix::ffi::OsStrExt;
|
use std::os::unix::ffi::OsStrExt;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
use anyhow::{bail, format_err, Error};
|
use anyhow::{bail, format_err, Error};
|
||||||
use futures::*;
|
use futures::*;
|
||||||
@ -182,7 +183,7 @@ pub fn list_groups(
|
|||||||
return Ok(group_info);
|
return Ok(group_info);
|
||||||
}
|
}
|
||||||
|
|
||||||
let snapshots = match group.list_backups(&datastore.base_path()) {
|
let snapshots = match group.list_backups() {
|
||||||
Ok(snapshots) => snapshots,
|
Ok(snapshots) => snapshots,
|
||||||
Err(_) => return Ok(group_info),
|
Err(_) => return Ok(group_info),
|
||||||
};
|
};
|
||||||
@ -294,7 +295,7 @@ pub fn list_snapshot_files(
|
|||||||
PRIV_DATASTORE_AUDIT | PRIV_DATASTORE_READ,
|
PRIV_DATASTORE_AUDIT | PRIV_DATASTORE_READ,
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
let info = BackupInfo::new(&datastore.base_path(), snapshot)?;
|
let info = BackupInfo::new(snapshot)?;
|
||||||
|
|
||||||
let (_manifest, files) = get_all_snapshot_files(&datastore, &info)?;
|
let (_manifest, files) = get_all_snapshot_files(&datastore, &info)?;
|
||||||
|
|
||||||
@ -405,7 +406,7 @@ pub fn list_snapshots(
|
|||||||
group: group.into(),
|
group: group.into(),
|
||||||
time: info.backup_dir.backup_time(),
|
time: info.backup_dir.backup_time(),
|
||||||
};
|
};
|
||||||
let protected = info.backup_dir.is_protected(datastore.base_path());
|
let protected = info.backup_dir.is_protected();
|
||||||
|
|
||||||
match get_all_snapshot_files(&datastore, &info) {
|
match get_all_snapshot_files(&datastore, &info) {
|
||||||
Ok((manifest, files)) => {
|
Ok((manifest, files)) => {
|
||||||
@ -488,7 +489,7 @@ pub fn list_snapshots(
|
|||||||
return Ok(snapshots);
|
return Ok(snapshots);
|
||||||
}
|
}
|
||||||
|
|
||||||
let group_backups = group.list_backups(&datastore.base_path())?;
|
let group_backups = group.list_backups()?;
|
||||||
|
|
||||||
snapshots.extend(
|
snapshots.extend(
|
||||||
group_backups
|
group_backups
|
||||||
@ -500,7 +501,10 @@ pub fn list_snapshots(
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_snapshots_count(store: &DataStore, filter_owner: Option<&Authid>) -> Result<Counts, Error> {
|
fn get_snapshots_count(
|
||||||
|
store: &Arc<DataStore>,
|
||||||
|
filter_owner: Option<&Authid>,
|
||||||
|
) -> Result<Counts, Error> {
|
||||||
store
|
store
|
||||||
.iter_backup_groups_ok()?
|
.iter_backup_groups_ok()?
|
||||||
.filter(|group| {
|
.filter(|group| {
|
||||||
@ -519,7 +523,7 @@ fn get_snapshots_count(store: &DataStore, filter_owner: Option<&Authid>) -> Resu
|
|||||||
}
|
}
|
||||||
})
|
})
|
||||||
.try_fold(Counts::default(), |mut counts, group| {
|
.try_fold(Counts::default(), |mut counts, group| {
|
||||||
let snapshot_count = group.list_backups(&store.base_path())?.len() as u64;
|
let snapshot_count = group.list_backups()?.len() as u64;
|
||||||
|
|
||||||
// only include groups with snapshots, counting/displaying emtpy groups can confuse
|
// only include groups with snapshots, counting/displaying emtpy groups can confuse
|
||||||
if snapshot_count > 0 {
|
if snapshot_count > 0 {
|
||||||
@ -788,7 +792,7 @@ pub fn prune(
|
|||||||
|
|
||||||
let mut prune_result = Vec::new();
|
let mut prune_result = Vec::new();
|
||||||
|
|
||||||
let list = group.list_backups(&datastore.base_path())?;
|
let list = group.list_backups()?;
|
||||||
|
|
||||||
let mut prune_info = compute_prune_info(list, &prune_options)?;
|
let mut prune_info = compute_prune_info(list, &prune_options)?;
|
||||||
|
|
||||||
@ -1797,7 +1801,7 @@ pub fn get_protection(
|
|||||||
PRIV_DATASTORE_AUDIT,
|
PRIV_DATASTORE_AUDIT,
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
Ok(backup_dir.is_protected(datastore.base_path()))
|
Ok(backup_dir.is_protected())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[api(
|
#[api(
|
||||||
|
@ -133,9 +133,7 @@ fn upgrade_to_backup_protocol(
|
|||||||
}
|
}
|
||||||
|
|
||||||
let last_backup = {
|
let last_backup = {
|
||||||
let info = backup_group
|
let info = backup_group.last_backup(true).unwrap_or(None);
|
||||||
.last_backup(&datastore.base_path(), true)
|
|
||||||
.unwrap_or(None);
|
|
||||||
if let Some(info) = info {
|
if let Some(info) = info {
|
||||||
let (manifest, _) = datastore.load_manifest(&info.backup_dir)?;
|
let (manifest, _) = datastore.load_manifest(&info.backup_dir)?;
|
||||||
let verify = manifest.unprotected["verify_state"].clone();
|
let verify = manifest.unprotected["verify_state"].clone();
|
||||||
|
@ -123,7 +123,7 @@ fn upgrade_to_backup_reader_protocol(
|
|||||||
}
|
}
|
||||||
|
|
||||||
let _guard = lock_dir_noblock_shared(
|
let _guard = lock_dir_noblock_shared(
|
||||||
&backup_dir.full_path(datastore.base_path()),
|
&backup_dir.full_path(),
|
||||||
"snapshot",
|
"snapshot",
|
||||||
"locked by another operation",
|
"locked by another operation",
|
||||||
)?;
|
)?;
|
||||||
|
@ -458,7 +458,7 @@ fn backup_worker(
|
|||||||
progress.done_snapshots = 0;
|
progress.done_snapshots = 0;
|
||||||
progress.group_snapshots = 0;
|
progress.group_snapshots = 0;
|
||||||
|
|
||||||
let snapshot_list = group.list_backups(&datastore.base_path())?;
|
let snapshot_list = group.list_backups()?;
|
||||||
|
|
||||||
// filter out unfinished backups
|
// filter out unfinished backups
|
||||||
let mut snapshot_list: Vec<_> = snapshot_list
|
let mut snapshot_list: Vec<_> = snapshot_list
|
||||||
|
@ -450,7 +450,7 @@ pub fn verify_backup_group(
|
|||||||
filter: Option<&dyn Fn(&BackupManifest) -> bool>,
|
filter: Option<&dyn Fn(&BackupManifest) -> bool>,
|
||||||
) -> Result<Vec<String>, Error> {
|
) -> Result<Vec<String>, Error> {
|
||||||
let mut errors = Vec::new();
|
let mut errors = Vec::new();
|
||||||
let mut list = match group.list_backups(&verify_worker.datastore.base_path()) {
|
let mut list = match group.list_backups() {
|
||||||
Ok(list) => list,
|
Ok(list) => list,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
task_log!(
|
task_log!(
|
||||||
|
@ -44,7 +44,7 @@ pub fn prune_datastore(
|
|||||||
|
|
||||||
for group in datastore.iter_backup_groups()? {
|
for group in datastore.iter_backup_groups()? {
|
||||||
let group = group?;
|
let group = group?;
|
||||||
let list = group.list_backups(&datastore.base_path())?;
|
let list = group.list_backups()?;
|
||||||
|
|
||||||
if !has_privs && !datastore.owns_backup(group.as_ref(), &auth_id)? {
|
if !has_privs && !datastore.owns_backup(group.as_ref(), &auth_id)? {
|
||||||
continue;
|
continue;
|
||||||
|
@ -663,13 +663,13 @@ pub async fn pull_group(
|
|||||||
|
|
||||||
if params.remove_vanished {
|
if params.remove_vanished {
|
||||||
let group = params.store.backup_group(group.clone());
|
let group = params.store.backup_group(group.clone());
|
||||||
let local_list = group.list_backups(¶ms.store.base_path())?;
|
let local_list = group.list_backups()?;
|
||||||
for info in local_list {
|
for info in local_list {
|
||||||
let backup_time = info.backup_dir.backup_time();
|
let backup_time = info.backup_dir.backup_time();
|
||||||
if remote_snapshots.contains(&backup_time) {
|
if remote_snapshots.contains(&backup_time) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if info.backup_dir.is_protected(params.store.base_path()) {
|
if info.backup_dir.is_protected() {
|
||||||
task_log!(
|
task_log!(
|
||||||
worker,
|
worker,
|
||||||
"don't delete vanished snapshot {:?} (protected)",
|
"don't delete vanished snapshot {:?} (protected)",
|
||||||
|
Loading…
Reference in New Issue
Block a user