2020-07-29 11:29:13 +00:00
|
|
|
use std::collections::HashSet;
|
|
|
|
|
2020-06-24 11:11:45 +00:00
|
|
|
use anyhow::{bail, Error};
|
|
|
|
|
|
|
|
use crate::server::WorkerTask;
|
|
|
|
|
|
|
|
use super::{
|
|
|
|
DataStore, BackupGroup, BackupDir, BackupInfo, IndexFile,
|
|
|
|
ENCR_COMPR_BLOB_MAGIC_1_0, ENCRYPTED_BLOB_MAGIC_1_0,
|
|
|
|
FileInfo, ArchiveType, archive_type,
|
|
|
|
};
|
|
|
|
|
|
|
|
fn verify_blob(datastore: &DataStore, backup_dir: &BackupDir, info: &FileInfo) -> Result<(), Error> {
|
|
|
|
|
2020-07-28 08:23:16 +00:00
|
|
|
let blob = datastore.load_blob(backup_dir, &info.filename)?;
|
2020-06-24 11:11:45 +00:00
|
|
|
|
2020-07-29 11:29:13 +00:00
|
|
|
let raw_size = blob.raw_size();
|
2020-06-24 11:11:45 +00:00
|
|
|
if raw_size != info.size {
|
|
|
|
bail!("wrong size ({} != {})", info.size, raw_size);
|
|
|
|
}
|
|
|
|
|
2020-07-28 08:23:16 +00:00
|
|
|
let csum = openssl::sha::sha256(blob.raw_data());
|
2020-06-24 11:11:45 +00:00
|
|
|
if csum != info.csum {
|
|
|
|
bail!("wrong index checksum");
|
|
|
|
}
|
|
|
|
|
|
|
|
let magic = blob.magic();
|
|
|
|
|
|
|
|
if magic == &ENCR_COMPR_BLOB_MAGIC_1_0 || magic == &ENCRYPTED_BLOB_MAGIC_1_0 {
|
|
|
|
return Ok(());
|
|
|
|
}
|
|
|
|
|
|
|
|
blob.decode(None)?;
|
|
|
|
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2020-06-26 06:14:45 +00:00
|
|
|
fn verify_index_chunks(
|
|
|
|
datastore: &DataStore,
|
|
|
|
index: Box<dyn IndexFile>,
|
2020-07-29 11:29:13 +00:00
|
|
|
verified_chunks: &mut HashSet<[u8;32]>,
|
2020-06-26 06:14:45 +00:00
|
|
|
worker: &WorkerTask,
|
|
|
|
) -> Result<(), Error> {
|
|
|
|
|
2020-07-30 07:09:03 +00:00
|
|
|
let mut errors = 0;
|
2020-06-26 06:14:45 +00:00
|
|
|
for pos in 0..index.index_count() {
|
|
|
|
|
|
|
|
worker.fail_on_abort()?;
|
|
|
|
|
|
|
|
let info = index.chunk_info(pos).unwrap();
|
|
|
|
let size = info.range.end - info.range.start;
|
2020-07-29 11:29:13 +00:00
|
|
|
|
|
|
|
if !verified_chunks.contains(&info.digest) {
|
2020-07-30 07:09:03 +00:00
|
|
|
if let Err(err) = datastore.verify_stored_chunk(&info.digest, size) {
|
|
|
|
worker.log(format!("{}", err));
|
|
|
|
errors += 1;
|
|
|
|
} else {
|
|
|
|
verified_chunks.insert(info.digest);
|
|
|
|
}
|
2020-07-29 11:29:13 +00:00
|
|
|
}
|
2020-06-26 06:14:45 +00:00
|
|
|
}
|
|
|
|
|
2020-07-30 07:09:03 +00:00
|
|
|
if errors > 0 {
|
|
|
|
bail!("chunks could not be verified");
|
|
|
|
}
|
|
|
|
|
2020-06-26 06:14:45 +00:00
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
2020-07-29 11:29:13 +00:00
|
|
|
fn verify_fixed_index(
|
|
|
|
datastore: &DataStore,
|
|
|
|
backup_dir: &BackupDir,
|
|
|
|
info: &FileInfo,
|
|
|
|
verified_chunks: &mut HashSet<[u8;32]>,
|
|
|
|
worker: &WorkerTask,
|
|
|
|
) -> Result<(), Error> {
|
2020-06-24 11:11:45 +00:00
|
|
|
|
|
|
|
let mut path = backup_dir.relative_path();
|
|
|
|
path.push(&info.filename);
|
|
|
|
|
|
|
|
let index = datastore.open_fixed_reader(&path)?;
|
|
|
|
|
|
|
|
let (csum, size) = index.compute_csum();
|
|
|
|
if size != info.size {
|
|
|
|
bail!("wrong size ({} != {})", info.size, size);
|
|
|
|
}
|
|
|
|
|
|
|
|
if csum != info.csum {
|
|
|
|
bail!("wrong index checksum");
|
|
|
|
}
|
|
|
|
|
2020-07-29 11:29:13 +00:00
|
|
|
verify_index_chunks(datastore, Box::new(index), verified_chunks, worker)
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
|
2020-07-29 11:29:13 +00:00
|
|
|
fn verify_dynamic_index(
|
|
|
|
datastore: &DataStore,
|
|
|
|
backup_dir: &BackupDir,
|
|
|
|
info: &FileInfo,
|
|
|
|
verified_chunks: &mut HashSet<[u8;32]>,
|
|
|
|
worker: &WorkerTask,
|
|
|
|
) -> Result<(), Error> {
|
|
|
|
|
2020-06-24 11:11:45 +00:00
|
|
|
let mut path = backup_dir.relative_path();
|
|
|
|
path.push(&info.filename);
|
|
|
|
|
|
|
|
let index = datastore.open_dynamic_reader(&path)?;
|
|
|
|
|
|
|
|
let (csum, size) = index.compute_csum();
|
|
|
|
if size != info.size {
|
|
|
|
bail!("wrong size ({} != {})", info.size, size);
|
|
|
|
}
|
|
|
|
|
|
|
|
if csum != info.csum {
|
|
|
|
bail!("wrong index checksum");
|
|
|
|
}
|
|
|
|
|
2020-07-29 11:29:13 +00:00
|
|
|
verify_index_chunks(datastore, Box::new(index), verified_chunks, worker)
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/// Verify a single backup snapshot
|
|
|
|
///
|
|
|
|
/// This checks all archives inside a backup snapshot.
|
|
|
|
/// Errors are logged to the worker log.
|
|
|
|
///
|
2020-06-25 10:55:34 +00:00
|
|
|
/// Returns
|
|
|
|
/// - Ok(true) if verify is successful
|
|
|
|
/// - Ok(false) if there were verification errors
|
|
|
|
/// - Err(_) if task was aborted
|
2020-07-29 11:29:13 +00:00
|
|
|
pub fn verify_backup_dir(
|
|
|
|
datastore: &DataStore,
|
|
|
|
backup_dir: &BackupDir,
|
|
|
|
verified_chunks: &mut HashSet<[u8;32]>,
|
|
|
|
worker: &WorkerTask
|
|
|
|
) -> Result<bool, Error> {
|
2020-06-24 11:11:45 +00:00
|
|
|
|
|
|
|
let manifest = match datastore.load_manifest(&backup_dir) {
|
2020-07-08 07:19:24 +00:00
|
|
|
Ok((manifest, _crypt_mode, _)) => manifest,
|
2020-06-24 11:11:45 +00:00
|
|
|
Err(err) => {
|
|
|
|
worker.log(format!("verify {}:{} - manifest load error: {}", datastore.name(), backup_dir, err));
|
2020-06-25 10:55:34 +00:00
|
|
|
return Ok(false);
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
worker.log(format!("verify {}:{}", datastore.name(), backup_dir));
|
|
|
|
|
|
|
|
let mut error_count = 0;
|
|
|
|
|
|
|
|
for info in manifest.files() {
|
|
|
|
let result = proxmox::try_block!({
|
|
|
|
worker.log(format!(" check {}", info.filename));
|
|
|
|
match archive_type(&info.filename)? {
|
2020-07-29 11:29:13 +00:00
|
|
|
ArchiveType::FixedIndex => verify_fixed_index(&datastore, &backup_dir, info, verified_chunks, worker),
|
|
|
|
ArchiveType::DynamicIndex => verify_dynamic_index(&datastore, &backup_dir, info, verified_chunks, worker),
|
2020-06-24 11:11:45 +00:00
|
|
|
ArchiveType::Blob => verify_blob(&datastore, &backup_dir, info),
|
|
|
|
}
|
|
|
|
});
|
2020-06-25 10:55:34 +00:00
|
|
|
|
|
|
|
worker.fail_on_abort()?;
|
|
|
|
|
2020-06-24 11:11:45 +00:00
|
|
|
if let Err(err) = result {
|
|
|
|
worker.log(format!("verify {}:{}/{} failed: {}", datastore.name(), backup_dir, info.filename, err));
|
|
|
|
error_count += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-06-25 10:55:34 +00:00
|
|
|
Ok(error_count == 0)
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
|
2020-06-25 10:55:34 +00:00
|
|
|
/// Verify all backups inside a backup group
|
|
|
|
///
|
|
|
|
/// Errors are logged to the worker log.
|
|
|
|
///
|
|
|
|
/// Returns
|
|
|
|
/// - Ok(true) if verify is successful
|
|
|
|
/// - Ok(false) if there were verification errors
|
|
|
|
/// - Err(_) if task was aborted
|
|
|
|
pub fn verify_backup_group(datastore: &DataStore, group: &BackupGroup, worker: &WorkerTask) -> Result<bool, Error> {
|
2020-06-24 11:11:45 +00:00
|
|
|
|
|
|
|
let mut list = match group.list_backups(&datastore.base_path()) {
|
|
|
|
Ok(list) => list,
|
|
|
|
Err(err) => {
|
|
|
|
worker.log(format!("verify group {}:{} - unable to list backups: {}", datastore.name(), group, err));
|
2020-06-25 10:55:34 +00:00
|
|
|
return Ok(false);
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
worker.log(format!("verify group {}:{}", datastore.name(), group));
|
|
|
|
|
|
|
|
let mut error_count = 0;
|
|
|
|
|
2020-07-29 11:29:13 +00:00
|
|
|
let mut verified_chunks = HashSet::with_capacity(1024*16); // start with 16384 chunks (up to 65GB)
|
|
|
|
|
2020-06-24 11:11:45 +00:00
|
|
|
BackupInfo::sort_list(&mut list, false); // newest first
|
|
|
|
for info in list {
|
2020-07-29 11:29:13 +00:00
|
|
|
if !verify_backup_dir(datastore, &info.backup_dir, &mut verified_chunks, worker)? {
|
2020-06-24 11:11:45 +00:00
|
|
|
error_count += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-06-25 10:55:34 +00:00
|
|
|
Ok(error_count == 0)
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
|
2020-06-25 10:55:34 +00:00
|
|
|
/// Verify all backups inside a datastore
|
|
|
|
///
|
|
|
|
/// Errors are logged to the worker log.
|
|
|
|
///
|
|
|
|
/// Returns
|
|
|
|
/// - Ok(true) if verify is successful
|
|
|
|
/// - Ok(false) if there were verification errors
|
|
|
|
/// - Err(_) if task was aborted
|
|
|
|
pub fn verify_all_backups(datastore: &DataStore, worker: &WorkerTask) -> Result<bool, Error> {
|
2020-06-24 11:11:45 +00:00
|
|
|
|
|
|
|
let list = match BackupGroup::list_groups(&datastore.base_path()) {
|
|
|
|
Ok(list) => list,
|
|
|
|
Err(err) => {
|
|
|
|
worker.log(format!("verify datastore {} - unable to list backups: {}", datastore.name(), err));
|
2020-06-25 10:55:34 +00:00
|
|
|
return Ok(false);
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
worker.log(format!("verify datastore {}", datastore.name()));
|
|
|
|
|
|
|
|
let mut error_count = 0;
|
|
|
|
for group in list {
|
2020-06-25 10:55:34 +00:00
|
|
|
if !verify_backup_group(datastore, &group, worker)? {
|
2020-06-24 11:11:45 +00:00
|
|
|
error_count += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-06-25 10:55:34 +00:00
|
|
|
Ok(error_count == 0)
|
2020-06-24 11:11:45 +00:00
|
|
|
}
|