Compare commits
6 Commits
Author | SHA1 | Date | |
---|---|---|---|
beaa683a52 | |||
33a88dafb9 | |||
224c65f8de | |||
f2b4b4b9fe | |||
ea9e559fc4 | |||
0cf14984cc |
@ -1,6 +1,6 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "proxmox-backup"
|
name = "proxmox-backup"
|
||||||
version = "0.8.8"
|
version = "0.8.9"
|
||||||
authors = ["Dietmar Maurer <dietmar@proxmox.com>"]
|
authors = ["Dietmar Maurer <dietmar@proxmox.com>"]
|
||||||
edition = "2018"
|
edition = "2018"
|
||||||
license = "AGPL-3"
|
license = "AGPL-3"
|
||||||
|
11
debian/changelog
vendored
11
debian/changelog
vendored
@ -1,3 +1,14 @@
|
|||||||
|
rust-proxmox-backup (0.8.9-1) unstable; urgency=medium
|
||||||
|
|
||||||
|
* improve termprocy (console) behavior on updating proxmox-backup-server and
|
||||||
|
other daemon restarts
|
||||||
|
|
||||||
|
* client: improve upload log output and speed calculation
|
||||||
|
|
||||||
|
* fix #2885: client upload: bail on duplicate backup targets
|
||||||
|
|
||||||
|
-- Proxmox Support Team <support@proxmox.com> Fri, 24 Jul 2020 11:24:07 +0200
|
||||||
|
|
||||||
rust-proxmox-backup (0.8.8-1) unstable; urgency=medium
|
rust-proxmox-backup (0.8.8-1) unstable; urgency=medium
|
||||||
|
|
||||||
* pxar: .pxarexclude: match behavior from absolute paths to the one described
|
* pxar: .pxarexclude: match behavior from absolute paths to the one described
|
||||||
|
@ -4,7 +4,7 @@ use std::os::unix::io::AsRawFd;
|
|||||||
use anyhow::{bail, format_err, Error};
|
use anyhow::{bail, format_err, Error};
|
||||||
use futures::{
|
use futures::{
|
||||||
future::{FutureExt, TryFutureExt},
|
future::{FutureExt, TryFutureExt},
|
||||||
try_join,
|
select,
|
||||||
};
|
};
|
||||||
use hyper::body::Body;
|
use hyper::body::Body;
|
||||||
use hyper::http::request::Parts;
|
use hyper::http::request::Parts;
|
||||||
@ -169,9 +169,10 @@ async fn termproxy(
|
|||||||
|
|
||||||
let mut cmd = tokio::process::Command::new("/usr/bin/termproxy");
|
let mut cmd = tokio::process::Command::new("/usr/bin/termproxy");
|
||||||
|
|
||||||
cmd.args(&arguments);
|
cmd.args(&arguments)
|
||||||
cmd.stdout(std::process::Stdio::piped());
|
.kill_on_drop(true)
|
||||||
cmd.stderr(std::process::Stdio::piped());
|
.stdout(std::process::Stdio::piped())
|
||||||
|
.stderr(std::process::Stdio::piped());
|
||||||
|
|
||||||
let mut child = cmd.spawn().expect("error executing termproxy");
|
let mut child = cmd.spawn().expect("error executing termproxy");
|
||||||
|
|
||||||
@ -184,7 +185,7 @@ async fn termproxy(
|
|||||||
while let Some(line) = reader.next_line().await? {
|
while let Some(line) = reader.next_line().await? {
|
||||||
worker_stdout.log(line);
|
worker_stdout.log(line);
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok::<(), Error>(())
|
||||||
};
|
};
|
||||||
|
|
||||||
let worker_stderr = worker.clone();
|
let worker_stderr = worker.clone();
|
||||||
@ -193,18 +194,24 @@ async fn termproxy(
|
|||||||
while let Some(line) = reader.next_line().await? {
|
while let Some(line) = reader.next_line().await? {
|
||||||
worker_stderr.warn(line);
|
worker_stderr.warn(line);
|
||||||
}
|
}
|
||||||
Ok(())
|
Ok::<(), Error>(())
|
||||||
};
|
};
|
||||||
|
|
||||||
let (exit_code, _, _) = try_join!(child, stdout_fut, stderr_fut)?;
|
select!{
|
||||||
if !exit_code.success() {
|
res = child.fuse() => {
|
||||||
match exit_code.code() {
|
let exit_code = res?;
|
||||||
Some(code) => bail!("termproxy exited with {}", code),
|
if !exit_code.success() {
|
||||||
None => bail!("termproxy exited by signal"),
|
match exit_code.code() {
|
||||||
}
|
Some(code) => bail!("termproxy exited with {}", code),
|
||||||
|
None => bail!("termproxy exited by signal"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
},
|
||||||
|
res = stdout_fut.fuse() => res,
|
||||||
|
res = stderr_fut.fuse() => res,
|
||||||
|
res = worker.abort_future().fuse() => res.map_err(Error::from),
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(())
|
|
||||||
},
|
},
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
@ -260,7 +267,7 @@ fn upgrade_to_websocket(
|
|||||||
|
|
||||||
let (ws, response) = WebSocket::new(parts.headers)?;
|
let (ws, response) = WebSocket::new(parts.headers)?;
|
||||||
|
|
||||||
tokio::spawn(async move {
|
crate::server::spawn_internal_task(async move {
|
||||||
let conn: Upgraded = match req_body.on_upgrade().map_err(Error::from).await {
|
let conn: Upgraded = match req_body.on_upgrade().map_err(Error::from).await {
|
||||||
Ok(upgraded) => upgraded,
|
Ok(upgraded) => upgraded,
|
||||||
_ => bail!("error"),
|
_ => bail!("error"),
|
||||||
|
@ -935,12 +935,18 @@ async fn create_backup(
|
|||||||
}
|
}
|
||||||
|
|
||||||
let mut upload_list = vec![];
|
let mut upload_list = vec![];
|
||||||
|
let mut target_set = HashSet::new();
|
||||||
|
|
||||||
for backupspec in backupspec_list {
|
for backupspec in backupspec_list {
|
||||||
let spec = parse_backup_specification(backupspec.as_str().unwrap())?;
|
let spec = parse_backup_specification(backupspec.as_str().unwrap())?;
|
||||||
let filename = &spec.config_string;
|
let filename = &spec.config_string;
|
||||||
let target = &spec.archive_name;
|
let target = &spec.archive_name;
|
||||||
|
|
||||||
|
if target_set.contains(target) {
|
||||||
|
bail!("got target twice: '{}'", target);
|
||||||
|
}
|
||||||
|
target_set.insert(target.to_string());
|
||||||
|
|
||||||
use std::os::unix::fs::FileTypeExt;
|
use std::os::unix::fs::FileTypeExt;
|
||||||
|
|
||||||
let metadata = std::fs::metadata(filename)
|
let metadata = std::fs::metadata(filename)
|
||||||
|
@ -264,9 +264,9 @@ impl BackupWriter {
|
|||||||
crate::tools::format::strip_server_file_expenstion(archive_name.clone())
|
crate::tools::format::strip_server_file_expenstion(archive_name.clone())
|
||||||
};
|
};
|
||||||
if archive_name != CATALOG_NAME {
|
if archive_name != CATALOG_NAME {
|
||||||
let speed: HumanByte = (uploaded / (duration.as_secs() as usize)).into();
|
let speed: HumanByte = ((uploaded * 1_000_000) / (duration.as_micros() as usize)).into();
|
||||||
let uploaded: HumanByte = uploaded.into();
|
let uploaded: HumanByte = uploaded.into();
|
||||||
println!("{}: had to upload {} from {} in {}s, avgerage speed {}/s).", archive, uploaded, vsize_h, duration.as_secs(), speed);
|
println!("{}: had to upload {} of {} in {:.2}s, avgerage speed {}/s).", archive, uploaded, vsize_h, duration.as_secs_f64(), speed);
|
||||||
} else {
|
} else {
|
||||||
println!("Uploaded backup catalog ({})", vsize_h);
|
println!("Uploaded backup catalog ({})", vsize_h);
|
||||||
}
|
}
|
||||||
|
@ -19,6 +19,7 @@ pub struct ServerState {
|
|||||||
pub shutdown_listeners: BroadcastData<()>,
|
pub shutdown_listeners: BroadcastData<()>,
|
||||||
pub last_worker_listeners: BroadcastData<()>,
|
pub last_worker_listeners: BroadcastData<()>,
|
||||||
pub worker_count: usize,
|
pub worker_count: usize,
|
||||||
|
pub task_count: usize,
|
||||||
pub reload_request: bool,
|
pub reload_request: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -28,6 +29,7 @@ lazy_static! {
|
|||||||
shutdown_listeners: BroadcastData::new(),
|
shutdown_listeners: BroadcastData::new(),
|
||||||
last_worker_listeners: BroadcastData::new(),
|
last_worker_listeners: BroadcastData::new(),
|
||||||
worker_count: 0,
|
worker_count: 0,
|
||||||
|
task_count: 0,
|
||||||
reload_request: false,
|
reload_request: false,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
@ -101,20 +103,40 @@ pub fn last_worker_future() -> impl Future<Output = Result<(), Error>> {
|
|||||||
}
|
}
|
||||||
|
|
||||||
pub fn set_worker_count(count: usize) {
|
pub fn set_worker_count(count: usize) {
|
||||||
let mut data = SERVER_STATE.lock().unwrap();
|
SERVER_STATE.lock().unwrap().worker_count = count;
|
||||||
data.worker_count = count;
|
|
||||||
|
|
||||||
if !(data.mode == ServerMode::Shutdown && data.worker_count == 0) { return; }
|
check_last_worker();
|
||||||
|
|
||||||
data.last_worker_listeners.notify_listeners(Ok(()));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
pub fn check_last_worker() {
|
pub fn check_last_worker() {
|
||||||
|
|
||||||
let mut data = SERVER_STATE.lock().unwrap();
|
let mut data = SERVER_STATE.lock().unwrap();
|
||||||
|
|
||||||
if !(data.mode == ServerMode::Shutdown && data.worker_count == 0) { return; }
|
if !(data.mode == ServerMode::Shutdown && data.worker_count == 0 && data.task_count == 0) { return; }
|
||||||
|
|
||||||
data.last_worker_listeners.notify_listeners(Ok(()));
|
data.last_worker_listeners.notify_listeners(Ok(()));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Spawns a tokio task that will be tracked for reload
|
||||||
|
/// and if it is finished, notify the last_worker_listener if we
|
||||||
|
/// are in shutdown mode
|
||||||
|
pub fn spawn_internal_task<T>(task: T)
|
||||||
|
where
|
||||||
|
T: Future + Send + 'static,
|
||||||
|
T::Output: Send + 'static,
|
||||||
|
{
|
||||||
|
let mut data = SERVER_STATE.lock().unwrap();
|
||||||
|
data.task_count += 1;
|
||||||
|
|
||||||
|
tokio::spawn(async move {
|
||||||
|
let _ = tokio::spawn(task).await; // ignore errors
|
||||||
|
|
||||||
|
{ // drop mutex
|
||||||
|
let mut data = SERVER_STATE.lock().unwrap();
|
||||||
|
if data.task_count > 0 {
|
||||||
|
data.task_count -= 1;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
check_last_worker();
|
||||||
|
});
|
||||||
|
}
|
||||||
|
Reference in New Issue
Block a user