tree wide: typo fixes through codespell
Most, not all, found and fixes using `codespell -wci3 -L crate` Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
This commit is contained in:
parent
6a35698796
commit
d20137e5a9
|
@ -161,7 +161,7 @@ impl AcmeClient {
|
|||
let mut data = Vec::<u8>::new();
|
||||
self.write_to(&mut data)?;
|
||||
let account_path = self.account_path.as_ref().ok_or_else(|| {
|
||||
format_err!("no account path set, cannot save upated account information")
|
||||
format_err!("no account path set, cannot save updated account information")
|
||||
})?;
|
||||
crate::config::acme::make_acme_account_dir()?;
|
||||
replace_file(
|
||||
|
|
|
@ -588,7 +588,7 @@ fn get_snapshots_count(store: &Arc<DataStore>, owner: Option<&Authid>) -> Result
|
|||
};
|
||||
let snapshot_count = group.list_backups()?.len() as u64;
|
||||
|
||||
// only include groups with snapshots, counting/displaying emtpy groups can confuse
|
||||
// only include groups with snapshots, counting/displaying empty groups can confuse
|
||||
if snapshot_count > 0 {
|
||||
let type_count = match group.backup_type() {
|
||||
BackupType::Ct => counts.ct.get_or_insert(Default::default()),
|
||||
|
@ -647,12 +647,12 @@ pub fn status(
|
|||
false // allow at least counts, user can read groups anyway..
|
||||
} else {
|
||||
match user_info.any_privs_below(&auth_id, &["datastore", &store], NS_PRIVS_OK) {
|
||||
// avoid leaking existance info if users hasn't at least any priv. below
|
||||
// avoid leaking existence info if users hasn't at least any priv. below
|
||||
Ok(false) | Err(_) => return Err(http_err!(FORBIDDEN, "permission check failed")),
|
||||
_ => false,
|
||||
}
|
||||
};
|
||||
let datastore = datastore?; // only unwrap no to avoid leaking existance info
|
||||
let datastore = datastore?; // only unwrap no to avoid leaking existence info
|
||||
|
||||
let (counts, gc_status) = if verbose {
|
||||
let filter_owner = if store_privs & PRIV_DATASTORE_AUDIT != 0 {
|
||||
|
|
|
@ -133,7 +133,7 @@ pub fn list_namespaces(
|
|||
},
|
||||
"delete-groups": {
|
||||
type: bool,
|
||||
description: "If set, all groups will be destroyed in the whole hierachy below and\
|
||||
description: "If set, all groups will be destroyed in the whole hierarchy below and\
|
||||
including `ns`. If not set, only empty namespaces will be pruned.",
|
||||
optional: true,
|
||||
default: false,
|
||||
|
|
|
@ -121,7 +121,7 @@ pub fn update_webauthn_config(
|
|||
} else {
|
||||
let rp = webauthn
|
||||
.rp
|
||||
.ok_or_else(|| format_err!("missing proeprty: 'rp'"))?;
|
||||
.ok_or_else(|| format_err!("missing property: 'rp'"))?;
|
||||
let origin = webauthn.origin;
|
||||
let id = webauthn
|
||||
.id
|
||||
|
|
|
@ -317,7 +317,7 @@ fn upgrade_to_websocket(
|
|||
}
|
||||
|
||||
#[api]
|
||||
/// List Nodes (only for compatiblity)
|
||||
/// List Nodes (only for compatibility)
|
||||
fn list_nodes() -> Result<Value, Error> {
|
||||
Ok(json!([ { "node": proxmox_sys::nodename().to_string() } ]))
|
||||
}
|
||||
|
|
|
@ -824,7 +824,7 @@ fn restore_list_worker(
|
|||
if !media_file_chunk_map.is_empty() {
|
||||
task_log!(worker, "Phase 2: restore chunks to datastores");
|
||||
} else {
|
||||
task_log!(worker, "all chunks exist already, skipping phase 2...");
|
||||
task_log!(worker, "All chunks are already present, skip phase 2...");
|
||||
}
|
||||
|
||||
for (media_uuid, file_chunk_map) in media_file_chunk_map.iter_mut() {
|
||||
|
@ -1988,7 +1988,7 @@ pub fn fast_catalog_restore(
|
|||
if &media_uuid != catalog_uuid {
|
||||
task_log!(
|
||||
worker,
|
||||
"catalog uuid missmatch at pos {}",
|
||||
"catalog uuid mismatch at pos {}",
|
||||
current_file_number
|
||||
);
|
||||
continue;
|
||||
|
@ -1996,7 +1996,7 @@ pub fn fast_catalog_restore(
|
|||
if media_set_uuid != archive_header.media_set_uuid {
|
||||
task_log!(
|
||||
worker,
|
||||
"catalog media_set missmatch at pos {}",
|
||||
"catalog media_set mismatch at pos {}",
|
||||
current_file_number
|
||||
);
|
||||
continue;
|
||||
|
|
|
@ -90,7 +90,7 @@ pub fn can_access_any_namespace(
|
|||
})
|
||||
}
|
||||
|
||||
/// A priviledge aware iterator for all backup groups in all Namespaces below an anchor namespace,
|
||||
/// A privilege aware iterator for all backup groups in all Namespaces below an anchor namespace,
|
||||
/// most often that will be the `BackupNamespace::root()` one.
|
||||
///
|
||||
/// Is basically just a filter-iter for pbs_datastore::ListNamespacesRecursive including access and
|
||||
|
|
|
@ -1001,7 +1001,7 @@ async fn run_stat_generator() {
|
|||
async fn generate_host_stats() {
|
||||
match tokio::task::spawn_blocking(generate_host_stats_sync).await {
|
||||
Ok(()) => (),
|
||||
Err(err) => log::error!("generate_host_stats paniced: {}", err),
|
||||
Err(err) => log::error!("generate_host_stats panicked: {}", err),
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -154,7 +154,7 @@ pub fn complete_acme_plugin(_arg: &str, _param: &HashMap<String, String>) -> Vec
|
|||
pub fn complete_acme_plugin_type(_arg: &str, _param: &HashMap<String, String>) -> Vec<String> {
|
||||
vec![
|
||||
"dns".to_string(),
|
||||
//"http".to_string(), // makes currently not realyl sense to create or the like
|
||||
//"http".to_string(), // makes currently not really sense to create or the like
|
||||
]
|
||||
}
|
||||
|
||||
|
|
|
@ -40,11 +40,7 @@ pub fn do_garbage_collection_job(
|
|||
let status = worker.create_state(&result);
|
||||
|
||||
if let Err(err) = job.finish(status) {
|
||||
eprintln!(
|
||||
"could not finish job state for {}: {}",
|
||||
job.jobtype(),
|
||||
err
|
||||
);
|
||||
eprintln!("could not finish job state for {}: {}", job.jobtype(), err);
|
||||
}
|
||||
|
||||
if let Some(email) = email {
|
||||
|
|
|
@ -60,7 +60,7 @@ pub fn prune_datastore(
|
|||
&datastore,
|
||||
ns,
|
||||
max_depth,
|
||||
Some(PRIV_DATASTORE_MODIFY), // overides the owner check
|
||||
Some(PRIV_DATASTORE_MODIFY), // overrides the owner check
|
||||
Some(PRIV_DATASTORE_PRUNE), // additionally required if owner
|
||||
Some(&auth_id),
|
||||
)? {
|
||||
|
@ -188,11 +188,7 @@ pub fn do_prune_job(
|
|||
let status = worker.create_state(&result);
|
||||
|
||||
if let Err(err) = job.finish(status) {
|
||||
eprintln!(
|
||||
"could not finish job state for {}: {}",
|
||||
job.jobtype(),
|
||||
err
|
||||
);
|
||||
eprintln!("could not finish job state for {}: {}", job.jobtype(), err);
|
||||
}
|
||||
|
||||
result
|
||||
|
|
|
@ -75,11 +75,7 @@ pub fn do_verification_job(
|
|||
let status = worker.create_state(&job_result);
|
||||
|
||||
if let Err(err) = job.finish(status) {
|
||||
eprintln!(
|
||||
"could not finish job state for {}: {}",
|
||||
job.jobtype(),
|
||||
err
|
||||
);
|
||||
eprintln!("could not finish job state for {}: {}", job.jobtype(), err);
|
||||
}
|
||||
|
||||
if let Some(email) = email {
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
//! MediaSet Locking
|
||||
//!
|
||||
//! To add/remove media from a media set, or to modify catalogs we
|
||||
//! always do lock_media_set(). Also, we aquire this lock during
|
||||
//! always do lock_media_set(). Also, we acquire this lock during
|
||||
//! restore, to make sure it is not reused for backups.
|
||||
//!
|
||||
|
||||
|
|
|
@ -483,7 +483,7 @@ impl MediaCatalog {
|
|||
pub fn register_label(
|
||||
&mut self,
|
||||
uuid: &Uuid, // Media/MediaSet Uuid
|
||||
seq_nr: u64, // onyl used for media set labels
|
||||
seq_nr: u64, // only used for media set labels
|
||||
file_number: u64,
|
||||
) -> Result<(), Error> {
|
||||
self.check_register_label(file_number, uuid)?;
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
//! Media Pool
|
||||
//!
|
||||
//! A set of backup medias.
|
||||
//! A set of backup mediums.
|
||||
//!
|
||||
//! This struct manages backup media state during backup. The main
|
||||
//! purpose is to allocate media sets and assign new tapes to it.
|
||||
|
@ -392,7 +392,7 @@ impl MediaPool {
|
|||
|
||||
let uuid = media_id.label.uuid.clone();
|
||||
|
||||
MediaCatalog::overwrite(&self.state_path, &media_id, false)?; // overwite catalog
|
||||
MediaCatalog::overwrite(&self.state_path, &media_id, false)?; // overwrite catalog
|
||||
let clear_media_status = true; // remove Full status
|
||||
self.inventory.store(media_id, clear_media_status)?; // store persistently
|
||||
|
||||
|
|
|
@ -18,7 +18,7 @@ pub struct NewChunksIterator {
|
|||
impl NewChunksIterator {
|
||||
/// Creates the iterator, spawning a new thread
|
||||
///
|
||||
/// Make sure to join() the returnd thread handle.
|
||||
/// Make sure to join() the returned thread handle.
|
||||
pub fn spawn(
|
||||
datastore: Arc<DataStore>,
|
||||
snapshot_reader: Arc<Mutex<SnapshotReader>>,
|
||||
|
|
|
@ -34,7 +34,7 @@ struct ParsedTcRule {
|
|||
|
||||
/// Traffic control statistics
|
||||
pub struct TrafficStat {
|
||||
/// Total incomming traffic (bytes)
|
||||
/// Total incoming traffic (bytes)
|
||||
pub traffic_in: u64,
|
||||
/// Incoming data rate (bytes/second)
|
||||
pub rate_in: u64,
|
||||
|
|
Loading…
Reference in New Issue