add qemu-io crate, AioContext reactor helper
Signed-off-by: Wolfgang Bumiller <w.bumiller@proxmox.com>
This commit is contained in:
parent
2dcdd3b427
commit
8bf7342c92
|
@ -0,0 +1,29 @@
|
||||||
|
[package]
|
||||||
|
name = "qemu-io"
|
||||||
|
version = "0.1.0"
|
||||||
|
authors = [
|
||||||
|
"Wolfgang Bumiller <w.bumiller@proxmox.com>",
|
||||||
|
]
|
||||||
|
edition = "2018"
|
||||||
|
|
||||||
|
#[lib]
|
||||||
|
#crate-type = ['lib', 'cdylib']
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
failure = "0.1"
|
||||||
|
mio = "0.6"
|
||||||
|
|
||||||
|
# In this crate 'future' by default means standard-future.
|
||||||
|
# The 0.1-futures are exposed under the name 'futures_01'.
|
||||||
|
|
||||||
|
[dependencies.futures-preview]
|
||||||
|
version = "0.3.0-alpha.15"
|
||||||
|
features = ["compat", "io-compat"]
|
||||||
|
|
||||||
|
[dependencies.futures_01]
|
||||||
|
package = "futures"
|
||||||
|
version = "0.1"
|
||||||
|
|
||||||
|
[features]
|
||||||
|
default = ["standalone"]
|
||||||
|
standalone = []
|
|
@ -0,0 +1,11 @@
|
||||||
|
//! Provides a handle to an AioContext.
|
||||||
|
|
||||||
|
#[cfg(feature="standalone")]
|
||||||
|
mod standalone;
|
||||||
|
#[cfg(feature="standalone")]
|
||||||
|
pub use standalone::AioContext;
|
||||||
|
|
||||||
|
// TODO: Add the non-standalone variant to be linked with Qemu:
|
||||||
|
// The AioContext struct should provide a high-level version of `set_fd_handler` with the same
|
||||||
|
// interface the standalone version provides out of the box (transparently turning closures into
|
||||||
|
// `extern "C" fn(opaque: *const c_void)` calls.
|
|
@ -0,0 +1,188 @@
|
||||||
|
//! This implements the parts of qemu's AioContext interface we need for testing outside qemu.
|
||||||
|
|
||||||
|
use std::collections::HashMap;
|
||||||
|
use std::os::unix::io::RawFd;
|
||||||
|
use std::sync::{Arc, Mutex, RwLock};
|
||||||
|
use std::thread;
|
||||||
|
|
||||||
|
use failure::Error;
|
||||||
|
use mio::{Events, Poll, Token};
|
||||||
|
use mio::unix::EventedFd;
|
||||||
|
|
||||||
|
use crate::util::{AioCb, AioHandlerState};
|
||||||
|
|
||||||
|
/// This is a reference to a standalone `AioContextImpl` and allows instantiating a new context
|
||||||
|
/// with a polling thread.
|
||||||
|
#[derive(Clone)]
|
||||||
|
#[repr(transparent)]
|
||||||
|
pub struct AioContext(Arc<AioContextImpl>);
|
||||||
|
|
||||||
|
impl std::ops::Deref for AioContext {
|
||||||
|
type Target = AioContextImpl;
|
||||||
|
|
||||||
|
fn deref(&self) -> &Self::Target {
|
||||||
|
&*self.0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AioContext {
|
||||||
|
/// Create a new `AioContext` instance with an associated polling thread, which will live as
|
||||||
|
/// long as there are references to it.
|
||||||
|
pub fn new() -> Result<Self, Error> {
|
||||||
|
Ok(Self(AioContextImpl::new()?))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct AioContextImpl {
|
||||||
|
poll: Poll,
|
||||||
|
handlers: RwLock<HashMap<Token, AioHandlerState>>,
|
||||||
|
poll_thread: Mutex<Option<thread::JoinHandle<()>>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AioContextImpl {
|
||||||
|
pub fn new() -> Result<Arc<Self>, Error> {
|
||||||
|
let this = Arc::new(Self {
|
||||||
|
poll: Poll::new()?,
|
||||||
|
handlers: RwLock::new(HashMap::new()),
|
||||||
|
poll_thread: Mutex::new(None),
|
||||||
|
});
|
||||||
|
|
||||||
|
let this2 = Arc::clone(&this);
|
||||||
|
this.poll_thread.lock().unwrap().replace(thread::spawn(|| this2.main_loop()));
|
||||||
|
|
||||||
|
Ok(this)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Qemu's aio_set_fd_handler. We're skipping the `io_poll` parameter for this implementation
|
||||||
|
/// as we don't use it.
|
||||||
|
/// ```
|
||||||
|
/// void aio_set_fd_handler(AioContext *ctx,
|
||||||
|
/// int fd,
|
||||||
|
/// bool is_external,
|
||||||
|
/// IOHandler *io_read,
|
||||||
|
/// IOHandler *io_write,
|
||||||
|
/// AioPollFn *io_poll,
|
||||||
|
/// void *opaque);
|
||||||
|
/// ```
|
||||||
|
///
|
||||||
|
/// Since this does not have any ways of returning errors, wrong usage will cause a panic in
|
||||||
|
/// this test implementation.
|
||||||
|
pub fn set_fd_handler(
|
||||||
|
&self,
|
||||||
|
fd: RawFd,
|
||||||
|
io_read: Option<AioCb>,
|
||||||
|
io_write: Option<AioCb>,
|
||||||
|
// skipping io_poll,
|
||||||
|
//opaque: *const (),
|
||||||
|
) {
|
||||||
|
self.set_fd_handler_impl(fd, io_read, io_write, mio::PollOpt::level())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// This is going to be a proposed new api for Qemu's AioContext.
|
||||||
|
pub fn set_fd_handler_edge(
|
||||||
|
&self,
|
||||||
|
fd: RawFd,
|
||||||
|
io_read: Option<AioCb>,
|
||||||
|
io_write: Option<AioCb>,
|
||||||
|
// skipping io_poll,
|
||||||
|
//opaque: *const (),
|
||||||
|
) {
|
||||||
|
self.set_fd_handler_impl(fd, io_read, io_write, mio::PollOpt::edge())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn set_fd_handler_impl(
|
||||||
|
&self,
|
||||||
|
fd: RawFd,
|
||||||
|
io_read: Option<AioCb>,
|
||||||
|
io_write: Option<AioCb>,
|
||||||
|
// skipping io_poll,
|
||||||
|
//opaque: *const (),
|
||||||
|
poll_opt: mio::PollOpt,
|
||||||
|
) {
|
||||||
|
if io_read.is_none() && io_write.is_none() {
|
||||||
|
return self.remove_fd_handler(fd);
|
||||||
|
}
|
||||||
|
|
||||||
|
let handlers = AioHandlerState {
|
||||||
|
read: io_read,
|
||||||
|
write: io_write,
|
||||||
|
};
|
||||||
|
|
||||||
|
let mio_ready = handlers.mio_ready();
|
||||||
|
|
||||||
|
let token = Token(fd as usize);
|
||||||
|
|
||||||
|
use std::collections::hash_map::Entry;
|
||||||
|
match self.handlers.write().unwrap().entry(token) {
|
||||||
|
Entry::Vacant(entry) => {
|
||||||
|
self.poll.register(&EventedFd(&fd), token, mio_ready, poll_opt)
|
||||||
|
.expect("failed to register a new fd for polling");
|
||||||
|
entry.insert(handlers);
|
||||||
|
}
|
||||||
|
Entry::Occupied(mut entry) => {
|
||||||
|
self.poll.reregister(&EventedFd(&fd), token, mio_ready, poll_opt)
|
||||||
|
.expect("failed to update an existing poll fd");
|
||||||
|
entry.insert(handlers);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn remove_fd_handler(&self, fd: RawFd) {
|
||||||
|
let mut guard = self.handlers.write().unwrap();
|
||||||
|
self.poll.deregister(&EventedFd(&fd))
|
||||||
|
.expect("failed to remove an existing poll fd");
|
||||||
|
guard.remove(&Token(fd as usize));
|
||||||
|
}
|
||||||
|
|
||||||
|
/// We don't use qemu's aio_poll, so let's make this easy:
|
||||||
|
///
|
||||||
|
/// ```
|
||||||
|
/// bool aio_poll(AioContext *ctx, bool blocking);
|
||||||
|
/// ```
|
||||||
|
pub fn poll(&self) -> Result<(), Error> {
|
||||||
|
let timeout = Some(std::time::Duration::from_millis(100));
|
||||||
|
|
||||||
|
let mut events = Events::with_capacity(16);
|
||||||
|
|
||||||
|
if self.poll.poll(&mut events, timeout)? == 0 {
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
|
||||||
|
for event in events.iter() {
|
||||||
|
let token = event.token();
|
||||||
|
let ready = event.readiness();
|
||||||
|
// NOTE: We need to read-lock while fetching handlers, but handlers need a write-lock!!!
|
||||||
|
// because they need to be edge-triggered and therefore *update* this handler list!
|
||||||
|
//
|
||||||
|
// While we could instead do this here (or use edge triggering from mio), this would
|
||||||
|
// not properly simulate Qemu's AioContext, so we enforce this behavior here as well.
|
||||||
|
//
|
||||||
|
// This means we cannot just hold a read lock during the events.iter() iteration
|
||||||
|
// though.
|
||||||
|
let handler = self.handlers.read().unwrap().get(&token).map(|h| AioHandlerState {
|
||||||
|
// Those are Option<Arc>!
|
||||||
|
read: h.read.clone(),
|
||||||
|
write: h.write.clone(),
|
||||||
|
});
|
||||||
|
if let Some(handler) = handler {
|
||||||
|
if ready.is_readable() {
|
||||||
|
handler.read.as_ref().map(|func| func());
|
||||||
|
}
|
||||||
|
if ready.is_writable() {
|
||||||
|
handler.write.as_ref().map(|func| func());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn main_loop(mut self: Arc<Self>) {
|
||||||
|
while Arc::get_mut(&mut self).is_none() {
|
||||||
|
if let Err(err) = self.poll() {
|
||||||
|
dbg!("error AioContextImpl::poll(): {}", err);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,10 @@
|
||||||
|
// used for testing
|
||||||
|
|
||||||
|
mod util;
|
||||||
|
mod with_aio_context;
|
||||||
|
|
||||||
|
#[cfg(feature="standalone")]
|
||||||
|
mod aio_context;
|
||||||
|
|
||||||
|
pub use with_aio_context::WithAioContext;
|
||||||
|
pub use aio_context::AioContext;
|
|
@ -0,0 +1,39 @@
|
||||||
|
//! Some types used by both our internal testing AioContext implementation as well as our
|
||||||
|
//! WithAioContext wrapper.
|
||||||
|
|
||||||
|
/// An Aio Callback. Qemu's AioContext actually uses a void function taking an opaque pointer.
|
||||||
|
/// For simplicity we stick to closures for now.
|
||||||
|
pub type AioCb = std::sync::Arc<dyn Fn() + Send + Sync>;
|
||||||
|
|
||||||
|
/// This keeps track of our poll state (whether we wait to be notified for read or write
|
||||||
|
/// readiness.)
|
||||||
|
#[derive(Default)]
|
||||||
|
pub struct AioHandlerState {
|
||||||
|
pub read: Option<AioCb>,
|
||||||
|
pub write: Option<AioCb>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AioHandlerState {
|
||||||
|
/// Get an mio::Ready with readable set if `read` is `Some`, and writable
|
||||||
|
/// set if `write` is `Some`.
|
||||||
|
pub fn mio_ready(&self) -> mio::Ready {
|
||||||
|
use mio::Ready;
|
||||||
|
|
||||||
|
let mut ready = Ready::empty();
|
||||||
|
if self.read.is_some() {
|
||||||
|
ready |= Ready::readable();
|
||||||
|
}
|
||||||
|
|
||||||
|
if self.write.is_some() {
|
||||||
|
ready |= Ready::writable();
|
||||||
|
}
|
||||||
|
|
||||||
|
ready
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Shortcut
|
||||||
|
pub fn clear(&mut self) {
|
||||||
|
self.read = None;
|
||||||
|
self.write = None;
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,211 @@
|
||||||
|
//! This module provides `WithAioContext`, which is a helper to connect any raw I/O file descriptor
|
||||||
|
//! (`T: AsRawFd`) with an `AioContext`.
|
||||||
|
|
||||||
|
use std::io;
|
||||||
|
use std::os::unix::io::{AsRawFd, RawFd};
|
||||||
|
use std::pin::Pin;
|
||||||
|
use std::sync::{Arc, Mutex, MutexGuard};
|
||||||
|
use std::task::{Context, Poll};
|
||||||
|
|
||||||
|
use mio::Ready;
|
||||||
|
|
||||||
|
use crate::AioContext;
|
||||||
|
use crate::util::{AioCb, AioHandlerState};
|
||||||
|
|
||||||
|
/// This provides a basic mechanism to connect a type containing a file descriptor (i.e. it
|
||||||
|
/// implements `AsRawFd`) to an `AioContext`.
|
||||||
|
///
|
||||||
|
/// If the underlying type implements `Read` this wrapper also provides an `AsyncRead`
|
||||||
|
/// implementation. Likewise it'll provide `AsyncWrite` for types implementing `Write`.
|
||||||
|
/// For this to function properly, the underlying type needs to return `io::Error` of kind
|
||||||
|
/// `io::ErrorKind::WouldBlock` on blocking operations which should be retried when the file
|
||||||
|
/// descriptor becomes ready.
|
||||||
|
///
|
||||||
|
/// `WithAioContext` _owns_ the underlying object. This is because our Drop handler wants to
|
||||||
|
/// unregister the file descriptor, but systems like linux' epoll do that automatically when the fd
|
||||||
|
/// is closed, so we cannot have our file descriptor vanish before de-registering it, otherwise we
|
||||||
|
/// may be de-registering an already re-used number.
|
||||||
|
///
|
||||||
|
/// Implements `Deref<T>` so any methods of `T` still work on a `WithAioContext<T>`.
|
||||||
|
pub struct WithAioContext<T: AsRawFd> {
|
||||||
|
aio_context: AioContext,
|
||||||
|
fd: RawFd,
|
||||||
|
handlers: Arc<Mutex<AioHandlerState>>,
|
||||||
|
inner: Option<T>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: AsRawFd> std::ops::Deref for WithAioContext<T> {
|
||||||
|
type Target = T;
|
||||||
|
|
||||||
|
fn deref(&self) -> &Self::Target {
|
||||||
|
self.inner.as_ref().unwrap()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: AsRawFd> std::ops::DerefMut for WithAioContext<T> {
|
||||||
|
fn deref_mut(&mut self) -> &mut Self::Target {
|
||||||
|
self.inner.as_mut().unwrap()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: AsRawFd> WithAioContext<T> {
|
||||||
|
pub fn new(aio_context: AioContext, inner: T) -> Self {
|
||||||
|
Self {
|
||||||
|
aio_context,
|
||||||
|
fd: inner.as_raw_fd(),
|
||||||
|
handlers: Arc::new(Mutex::new(Default::default())),
|
||||||
|
inner: Some(inner),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Deregister from the `AioContext` and return the inner file handle.
|
||||||
|
pub fn into_inner(mut self) -> T {
|
||||||
|
let out = self.inner.take().unwrap();
|
||||||
|
std::mem::drop(self);
|
||||||
|
out
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Shortcut around the `unwrap()`. The `Option<>` around `inner` is only there because we have
|
||||||
|
/// a `Drop` implementation which prevents us to move-out the value in the `into_inner()`
|
||||||
|
/// method.
|
||||||
|
fn inner_mut(&mut self) -> &mut T {
|
||||||
|
self.inner.as_mut().unwrap()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Shortcut around the `unwrap()`, immutable variant:
|
||||||
|
//fn inner(&self) -> &T {
|
||||||
|
// self.inner.as_ref().unwrap()
|
||||||
|
//}
|
||||||
|
|
||||||
|
/// Shortcut to set_fd_handlers. For the "real" qemu interface we'll have to turn the closures
|
||||||
|
/// into raw function pointers here (they'll get an opaque pointer parameter).
|
||||||
|
fn commit_handlers(
|
||||||
|
aio_context: &AioContext,
|
||||||
|
fd: RawFd,
|
||||||
|
handlers: &mut MutexGuard<AioHandlerState>,
|
||||||
|
) {
|
||||||
|
aio_context.set_fd_handler(
|
||||||
|
fd,
|
||||||
|
handlers.read.as_ref().map(|x| (*x).clone()),
|
||||||
|
handlers.write.as_ref().map(|x| (*x).clone()),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a waker closure for a context for a specific ready state. When a file descriptor is
|
||||||
|
/// ready for reading or writing, we need to remove the corresponding handler from the
|
||||||
|
/// `AioContext` (make it an edge-trigger instead of a level trigger) before finally calling
|
||||||
|
/// `waker.wake_by_ref()` to queue the task for polling.
|
||||||
|
fn make_wake_fn(&self, cx: &mut Context, ready: Ready) -> AioCb {
|
||||||
|
let waker = cx.waker().clone();
|
||||||
|
|
||||||
|
// we don't want to be publicly clonable so clone manually here:
|
||||||
|
let aio_context = self.aio_context.clone();
|
||||||
|
let fd = self.fd;
|
||||||
|
let handlers = Arc::clone(&self.handlers);
|
||||||
|
Arc::new(move || {
|
||||||
|
let mut guard = handlers.lock().unwrap();
|
||||||
|
|
||||||
|
if ready.is_readable() {
|
||||||
|
guard.read = None;
|
||||||
|
}
|
||||||
|
|
||||||
|
if ready.is_writable() {
|
||||||
|
guard.write = None;
|
||||||
|
}
|
||||||
|
|
||||||
|
Self::commit_handlers(&aio_context, fd, &mut guard);
|
||||||
|
waker.wake_by_ref();
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Register our file descriptor with the `AioContext` for reading or writing.
|
||||||
|
/// This only affects the directions present in the provided `ready` value, and will leave the
|
||||||
|
/// other directions unchanged.
|
||||||
|
pub fn register(&self, cx: &mut Context, ready: Ready) {
|
||||||
|
let mut guard = self.handlers.lock().unwrap();
|
||||||
|
|
||||||
|
if ready.is_readable() {
|
||||||
|
guard.read = Some(self.make_wake_fn(cx, ready));
|
||||||
|
}
|
||||||
|
|
||||||
|
if ready.is_writable() {
|
||||||
|
guard.write = Some(self.make_wake_fn(cx, ready));
|
||||||
|
}
|
||||||
|
|
||||||
|
Self::commit_handlers(&self.aio_context, self.fd, &mut guard)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Helper to handle an `io::Result<T>`, turning `Result<T>` into `Poll<Result<T>>`, by
|
||||||
|
/// changing an `io::ErrorKind::WouldBlock` into `Poll::Pending` and taking care of registering
|
||||||
|
/// the file descriptor with the AioContext for the next wake-up.
|
||||||
|
/// `Ok` and errors other than the above will be passed through wrapped in `Poll::Ready`.
|
||||||
|
pub fn handle_aio_result<R>(
|
||||||
|
&self,
|
||||||
|
cx: &mut Context,
|
||||||
|
result: io::Result<R>,
|
||||||
|
ready: Ready,
|
||||||
|
) -> Poll<io::Result<R>> {
|
||||||
|
match result {
|
||||||
|
Ok(res) => Poll::Ready(Ok(res)),
|
||||||
|
Err(ref e) if e.kind() == io::ErrorKind::WouldBlock => {
|
||||||
|
self.register(cx, ready);
|
||||||
|
Poll::Pending
|
||||||
|
}
|
||||||
|
Err(err) => Poll::Ready(Err(err)),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T: AsRawFd> Drop for WithAioContext<T> {
|
||||||
|
fn drop(&mut self) {
|
||||||
|
let mut guard = self.handlers.lock().unwrap();
|
||||||
|
(*guard).clear();
|
||||||
|
if !guard.mio_ready().is_empty() {
|
||||||
|
Self::commit_handlers(&self.aio_context, self.fd, &mut guard);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T> futures::io::AsyncRead for WithAioContext<T>
|
||||||
|
where
|
||||||
|
T: AsRawFd + io::Read + Unpin,
|
||||||
|
{
|
||||||
|
fn poll_read(
|
||||||
|
mut self: Pin<&mut Self>,
|
||||||
|
cx: &mut Context,
|
||||||
|
buf: &mut [u8],
|
||||||
|
) -> Poll<io::Result<usize>> {
|
||||||
|
let res = self.inner_mut().read(buf);
|
||||||
|
self.handle_aio_result(cx, res, mio::Ready::readable())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<T> futures::io::AsyncWrite for WithAioContext<T>
|
||||||
|
where
|
||||||
|
T: AsRawFd + io::Write + Unpin,
|
||||||
|
{
|
||||||
|
fn poll_write(
|
||||||
|
mut self: Pin<&mut Self>,
|
||||||
|
cx: &mut Context,
|
||||||
|
buf: &[u8],
|
||||||
|
) -> Poll<io::Result<usize>> {
|
||||||
|
let result = self.inner_mut().write(buf);
|
||||||
|
self.handle_aio_result(cx, result, mio::Ready::writable())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn poll_flush(mut self: Pin<&mut Self>, cx: &mut Context) -> Poll<io::Result<()>> {
|
||||||
|
let result = self.inner_mut().flush();
|
||||||
|
self.handle_aio_result(cx, result, mio::Ready::writable())
|
||||||
|
}
|
||||||
|
|
||||||
|
// I'm not sure what they expect me to do here. The `close()` syscall has no async variant, so
|
||||||
|
// all I can do is `flush()` and then drop the inner stream...
|
||||||
|
//
|
||||||
|
// Using `.into_inner()` after this will cause a panic.
|
||||||
|
fn poll_close(mut self: Pin<&mut Self>, cx: &mut Context) -> Poll<io::Result<()>> {
|
||||||
|
let result = self.inner_mut().flush();
|
||||||
|
let _ = futures::ready!(self.handle_aio_result(cx, result, mio::Ready::writable()));
|
||||||
|
std::mem::drop(self.inner.take());
|
||||||
|
Poll::Ready(Ok(()))
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue