2019 is approaching. The rust team keeps their promise about asynchronous IO: async
is introduced as keywords, Pin, Future, Poll
and await
is introduced into standard library.
I have never used rust for asynchronous IO programming earlier, so I almost know nothing about it. However, I would use it for a project recently but couldn't find many documents that are remarkably helpful for newbie of rust asynchronous programming.
Eventually, I wrote several demo and implemented simple asynchronous IO based on mio
and coroutine
with the help of both of this blog (Tokio internals: Understanding Rust's asynchronous I/O framework from the bottom up) and source code of "new tokio" romio .
This is the final file server:
// examples/file-server.rs
#[macro_use]
extern crate log;
use asyncio::executor::{block_on, spawn, TcpListener, TcpStream};
use asyncio::fs_future::{read_to_string};
use failure::Error;
fn main() -> Result<(), Error> {
env_logger::init();
block_on(new_server())?
}
const CRLF: &[char] = &['\r', '\n'];
async fn new_server() -> Result<(), Error> {
let mut listener = TcpListener::bind(&"127.0.0.1:7878".parse()?)?;
info!("Listening on 127.0.0.1:7878");
while let Ok((stream, addr)) = listener.accept().await {
info!("connection from {}", addr);
spawn(handle_stream(stream))?;
}
Ok(())
}
async fn handle_stream(mut stream: TcpStream) -> Result<(), Error> {
stream.write_str("Please enter filename: ").await?;
let file_name_vec = stream.read().await?;
let file_name = String::from_utf8(file_name_vec)?.trim_matches(CRLF).to_owned();
let file_contents = read_to_string(file_name).await?;
stream.write_str(&file_contents).await?;
stream.close();
Ok(())
}
My purpose of writing this blog is to review and summarize, I will be happy if it can help someone who are interested in rust asynchronous programming. Given that the readability is the primary consideration when I wrote the code appearing in this blog, there may be some performance problem in code, please forgive me. If there are obvious problems in blog or code, you are welcome to point them up.
Most of the code appearing in this blog is collected in this repo (some code is too long, you had better clone and view it in editor), all examples work well at rustc-1.39`.
When executing examples/async-echo, set environment variable
RUST_LOG=info
for basic runtime information; setRUST_LOG=debug
for events polling information.
mio
is a tidy, low-level asynchronous IO library. Nowadays, almost all asynchronous IO libraries in rust ecosystem are based on mio
.
As sub modules like channel, timer
have been marked as deprecated since version-0.6.5, mio
provides only two core functions:
- basic encapsulation for OS asynchronous network IO
- custom events
The first core function corresponded to API in different OS respectively are:
- Linux(Android) => epoll
- Windows => iocp
- MacOS(iOS), FreeBSD => kqueue
- Fuchsia => <unknow>
mio
wraps different asynchronous network API in different OS into a common epoll-like asynchronous API, which supports both of udp
and tcp
.
besides
udp
andtcp
,mio
also provides some OS-specific API, likeuds
, we won't talk about them, you can find the usage in source code of mio.
This is a demo of asynchronous tcp IO:
// examples/tcp.rs
use mio::*;
use mio::net::{TcpListener, TcpStream};
use std::io::{Read, Write, self};
use failure::Error;
use std::time::{Duration, Instant};
const SERVER_ACCEPT: Token = Token(0);
const SERVER: Token = Token(1);
const CLIENT: Token = Token(2);
const SERVER_HELLO: &[u8] = b"PING";
const CLIENT_HELLO: &[u8] = b"PONG";
fn main() -> Result<(), Error> {
let addr = "127.0.0.1:13265".parse()?;
// Setup the server socket
let server = TcpListener::bind(&addr)?;
// Create a poll instance
let poll = Poll::new()?;
// Start listening for incoming connections
poll.register(&server, SERVER_ACCEPT, Ready::readable(),
PollOpt::edge())?;
// Setup the client socket
let mut client = TcpStream::connect(&addr)?;
let mut server_handler = None;
// Register the client
poll.register(&client, CLIENT, Ready::readable() | Ready::writable(),
PollOpt::edge())?;
// Create storage for events
let mut events = Events::with_capacity(1024);
let start = Instant::now();
let timeout = Duration::from_millis(10);
'top: loop {
poll.poll(&mut events, None)?;
for event in events.iter() {
if start.elapsed() >= timeout {
break 'top
}
match event.token() {
SERVER_ACCEPT => {
let (handler, addr) = server.accept()?;
println!("accept from addr: {}", &addr);
poll.register(&handler, SERVER, Ready::readable() | Ready::writable(), PollOpt::edge())?;
server_handler = Some(handler);
}
SERVER => {
if event.readiness().is_writable() {
if let Some(ref mut handler) = &mut server_handler {
match handler.write(SERVER_HELLO) {
Ok(_) => {
println!("server wrote");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
if event.readiness().is_readable() {
let mut hello = [0; 4];
if let Some(ref mut handler) = &mut server_handler {
match handler.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(CLIENT_HELLO, &hello);
println!("server received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
}
CLIENT => {
if event.readiness().is_writable() {
match client.write(CLIENT_HELLO) {
Ok(_) => {
println!("client wrote");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
if event.readiness().is_readable() {
let mut hello = [0; 4];
match client.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(SERVER_HELLO, &hello);
println!("client received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
_ => unreachable!(),
}
}
};
Ok(())
}
This demo is a little long, let's talk about its main loop first:
fn main() {
// ...
loop {
poll.poll(&mut events, None).unwrap();
// ...
}
}
We need call Poll::poll
in each loop, the first parameter events
is used for store events, we set it  with capacity 1024 here.
let mut events = Events::with_capacity(1024);
The type of second parameter timeout
is Option<Duration>
, method will return Ok(usize)
when some events occur or timeout if Some(duration) = timeout
.
The usize in Ok refers to the number of events, this value is deprecated and will be removed in 0.7.0.
Here we deliver timeout = None
,so when the method return without error, there must be some events. Let's traverse events:
match event.token() {
SERVER_ACCEPT => {
let (handler, addr) = server.accept()?;
println!("accept from addr: {}", &addr);
poll.register(&handler, SERVER, Ready::readable() | Ready::writable(), PollOpt::edge())?;
server_handler = Some(handler);
}
SERVER => {
if event.readiness().is_writable() {
if let Some(ref mut handler) = &mut server_handler {
match handler.write(SERVER_HELLO) {
Ok(_) => {
println!("server wrote");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
if event.readiness().is_readable() {
let mut hello = [0; 4];
if let Some(ref mut handler) = &mut server_handler {
match handler.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(CLIENT_HELLO, &hello);
println!("server received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
}
CLIENT => {
if event.readiness().is_writable() {
match client.write(CLIENT_HELLO) {
Ok(_) => {
println!("client wrote");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
if event.readiness().is_readable() {
let mut hello = [0; 4];
match client.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(SERVER_HELLO, &hello);
println!("client received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
_ => unreachable!(),
}
We need match token of each event, these tokens are just those we used to register. For example, we register server
using token SERVER_ACCEPT
.
const SERVER_ACCEPT: Token = Token(0);
...
// Start listening for incoming connections
poll.register(&server, SERVER_ACCEPT, Ready::readable(),
PollOpt::edge()).unwrap();
In this case, when we find event.token() == SERVER_ACCEPT
, we should think it's relevant to server
, so we try to accept a new TcpStream
and register it, using token SERVER
:
let (handler, addr) = server.accept()?;
println!("accept from addr: {}", &addr);
poll.register(&handler, SERVER, Ready::readable() | Ready::writable(), PollOpt::edge()).unwrap();
server_handler = Some(handler);
As the same, if we find event.token() == SERVER
,we should think it's relevant to handler
:
if event.readiness().is_writable() {
if let Some(ref mut handler) = &mut server_handler {
match handler.write(SERVER_HELLO) {
Ok(_) => {
println!("server wrote");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
if event.readiness().is_readable() {
let mut hello = [0; 4];
if let Some(ref mut handler) = &mut server_handler {
match handler.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(CLIENT_HELLO, &hello);
println!("server received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
}
In this case, we shoud response differently to different event.readiness()
, this is the third parameter of register, which named interest
. As its name, interest
means 'something you are interested', its type is Ready
. mio
support four kinds of Ready
, readable
, writable
, error
and hup
, you can union them.
We register handler
with Ready::readable() | Ready::writable()
, so event can be readable
or writable
or both, you can see it in control flow:
using
if event.readiness().is_writable() {
...
}
if event.readiness().is_readable() {
...
}
instead of
if event.readiness().is_writable() {
...
} else if event.readiness().is_readable() {
...
}
The code for dealing with event SERVER_ACCEPT
above is:
match event.token() {
SERVER_ACCEPT => {
let (handler, addr) = server.accept()?;
println!("accept from addr: {}", &addr);
poll.register(&handler, SERVER, Ready::readable() | Ready::writable(), PollOpt::edge()).unwrap();
server_handler = Some(handler);
}
The result server.accept()
returned is io::Result<(TcpStream, SocketAddr)>
. If we trust event
entirely, using try
is the right choise (if there should be a new TcpStream
is ready, server.accept()
returning Err
is unforeseen and unmanageable).
However, we should think event may be spurious, the possibility depends on OS and custom implement. There may not be a new TcpStream
is ready, in this case, server.accept()
will return WouldBlock Error
. We should regard WouldBlock Error
as a friendly warning: "there isn't a new TcpStream
is ready, please do it later again." So we should ignore it and continue the loop.
Like the code for dealing with event SERVER
match handler.write(SERVER_HELLO) {
Ok(_) => {
println!("server wrote");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
Now we can execute:
cargo run --example tcp
The terminal print some log:
client wrote
accept from addr: 127.0.0.1:53205
client wrote
server wrote
server received
...
We can see, in 10 milliseconds (let timeout = Duration::from_millis(10);
), server
and client
did dozens of writing and reading!
How should we do if we don't need dozens of writing and reading? In a pretty network environment, client
and server
is almost always writable, so Poll::poll
may return in dozens of microseconds.
In this case, we should change the forth parameter of register
:
poll.register(&server, SERVER_ACCEPT, Ready::readable(),
PollOpt::edge()).unwrap();
The type of PollOpt::edge()
is PollOpt
, means poll option. There are three kinds of poll options: level
, edge
and oneshot
, what's the difference of them?
For example, in this code:
if event.readiness().is_readable() {
let mut hello = [0; 4];
match client.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(SERVER_HELLO, &hello);
println!("client received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => continue,
err => {
err?;
}
}
}
When I receive a readable readiness, I read 4 bytes only. If there are 8 bytes in buffer:
- if I register this
TcpStream
withPollOpt::level()
, I MUST receive areadable readiness event
in next polling; - if I register this
TcpStream
withPollOpt::edge()
, I MAY cannot receive areadable readiness event
in next polling;
So, we can say that readiness using edge-triggered mode is a Draining readiness
, once a readiness event is received, the corresponding operation must be performed repeatedly until it returns WouldBlock
. We should alter code above into:
if event.readiness().is_readable() {
let mut hello = [0; 4];
loop {
match client.read_exact(&mut hello) {
Ok(_) => {
assert_eq!(SERVER_HELLO, &hello);
println!("client received");
}
Err(ref err) if err.kind() == io::ErrorKind::WouldBlock => break,
err => {
err?;
}
}
}
}
Then, what's the behavior of PollOpt::onshot()
? Let's talk about the first question of this section: if we want handler
to write only once, how should we do? The answer is: register it using PollOpt::oneshot()
let (handler, addr) = server.accept()?;
println!("accept from addr: {}", &addr);
poll.register(&handler, SERVER, Ready::readable() | Ready::writable(), PollOpt::oneshot())?;
server_handler = Some(handler);
In this case, you can only receive event SERVER
once, unless you re-register handler
using Poll::reregister
.
Poll::reregister
can using differentPollOpt
andinterest
from the last registering
There is still a problem in the code above: we using blocking IO macro println!
. We should avoid using blocking IO in the code for dealing events.
Given that FS(file-system) IO(including stdin, stdout, stderr
) is slow, we can send all FS-IO task to a worker thread.
use std::sync::mpsc::{Sender, Receiver, channel, SendError};
#[derive(Clone)]
pub struct Fs {
task_sender: Sender<Task>,
}
impl Fs {
pub fn new() -> Self {
let (sender, receiver) = channel();
std::thread::spawn(move || {
loop {
match receiver.recv() {
Ok(task) => {
match task {
Task::Println(ref string) => println!("{}", string),
Task::Exit => return
}
},
Err(_) => {
return;
}
}
}
});
Fs { task_sender: sender }
}
pub fn println(&self, string: String) {
self.task_sender.send(Task::Println(string)).unwrap()
}
}
pub enum Task {
Exit,
Println(String),
}
Now, we can replace all println!
with Fs::println
.
Implementing non-blocking println
is easy, because this function has no return. How should we do if we want other non-blocking FS-IO functions? For example, opening a file, then reading it to string, then printing the string, how should we do?
The easiest way is using callback, like this:
// src/fs.rs
use crossbeam_channel::{unbounded, Sender};
use failure::Error;
use std::fs::File;
use std::io::Read;
use std::thread;
#[derive(Clone)]
pub struct Fs {
task_sender: Sender<Task>,
}
pub struct FsHandler {
io_worker: thread::JoinHandle<Result<(), Error>>,
executor: thread::JoinHandle<Result<(), Error>>,
}
pub fn fs_async() -> (Fs, FsHandler) {
let (task_sender, task_receiver) = unbounded();
let (result_sender, result_receiver) = unbounded();
let io_worker = std::thread::spawn(move || {
while let Ok(task) = task_receiver.recv() {
match task {
Task::Println(ref string) => println!("{}", string),
Task::Open(path, callback, fs) => {
result_sender.send(TaskResult::Open(File::open(path)?, callback, fs))?
}
Task::ReadToString(mut file, callback, fs) => {
let mut value = String::new();
file.read_to_string(&mut value)?;
result_sender.send(TaskResult::ReadToString(value, callback, fs))?
}
Task::Exit => {
result_sender.send(TaskResult::Exit)?;
break;
}
}
}
Ok(())
});
let executor = std::thread::spawn(move || {
loop {
let result = result_receiver.recv()?;
match result {
TaskResult::ReadToString(value, callback, fs) => callback(value, fs)?,
TaskResult::Open(file, callback, fs) => callback(file, fs)?,
TaskResult::Exit => break,
};
}
Ok(())
});
(
Fs { task_sender },
FsHandler {
io_worker,
executor,
},
)
}
impl Fs {
pub fn println(&self, string: String) -> Result<(), Error> {
Ok(self.task_sender.send(Task::Println(string))?)
}
pub fn open<F>(&self, path: &str, callback: F) -> Result<(), Error>
where
F: FnOnce(File, Fs) -> Result<(), Error> + Sync + Send + 'static,
{
Ok(self.task_sender.send(Task::Open(
path.to_string(),
Box::new(callback),
self.clone(),
))?)
}
pub fn read_to_string<F>(&self, file: File, callback: F) -> Result<(), Error>
where
F: FnOnce(String, Fs) -> Result<(), Error> + Sync + Send + 'static,
{
Ok(self
.task_sender
.send(Task::ReadToString(file, Box::new(callback), self.clone()))?)
}
pub fn close(&self) -> Result<(), Error> {
Ok(self.task_sender.send(Task::Exit)?)
}
}
impl FsHandler {
pub fn join(self) -> Result<(), Error> {
self.io_worker.join().unwrap()?;
self.executor.join().unwrap()
}
}
type FileCallback = Box<dyn FnOnce(File, Fs) -> Result<(), Error> + Sync + Send>;
type StringCallback = Box<dyn FnOnce(String, Fs) -> Result<(), Error> + Sync + Send>;
pub enum Task {
Exit,
Println(String),
Open(String, FileCallback, Fs),
ReadToString(File, StringCallback, Fs),
}
pub enum TaskResult {
Exit,
Open(File, FileCallback, Fs),
ReadToString(String, StringCallback, Fs),
}
// examples/fs.rs
use asyncio::fs::fs_async;
use failure::Error;
const TEST_FILE_VALUE: &str = "Hello, World!\n";
fn main() -> Result<(), Error> {
let (fs, fs_handler) = fs_async();
fs.open("./examples/test.txt", |file, fs| {
fs.read_to_string(file, |value, fs| {
assert_eq!(TEST_FILE_VALUE, &value);
fs.println(value)?;
fs.close()
})
})?;
fs_handler.join()?;
Ok(())
}
running this example:
cargo run --example fs
This implementation work well, but the executor thread is still blocked by the io-worker thread (result_receiver.recv()
). Can we run a polling loop in executor thread to not be blocked by IO? (executor should execute (result_receiver.recv()
only when there are some results in result channel).
To implement a non-blocking executor, we can use custom events supported by mio
.
Altering the code above:
// src/fs_mio.rs
use crossbeam_channel::{unbounded, Sender, TryRecvError};
use failure::Error;
use mio::*;
use std::fs::File;
use std::io::Read;
use std::thread;
use std::time::Duration;
#[derive(Clone)]
pub struct Fs {
task_sender: Sender<Task>,
}
pub struct FsHandler {
io_worker: thread::JoinHandle<Result<(), Error>>,
executor: thread::JoinHandle<Result<(), Error>>,
}
const FS_TOKEN: Token = Token(0);
pub fn fs_async() -> (Fs, FsHandler) {
let (task_sender, task_receiver) = unbounded();
let (result_sender, result_receiver) = unbounded();
let poll = Poll::new().unwrap();
let (registration, set_readiness) = Registration::new2();
poll.register(
®istration,
FS_TOKEN,
Ready::readable(),
PollOpt::oneshot(),
)
.unwrap();
let io_worker = std::thread::spawn(move || {
while let Ok(task) = task_receiver.recv() {
match task {
Task::Println(ref string) => println!("{}", string),
Task::Open(path, callback, fs) => {
result_sender.send(TaskResult::Open(File::open(path)?, callback, fs))?;
set_readiness.set_readiness(Ready::readable())?;
}
Task::ReadToString(mut file, callback, fs) => {
let mut value = String::new();
file.read_to_string(&mut value)?;
result_sender.send(TaskResult::ReadToString(value, callback, fs))?;
set_readiness.set_readiness(Ready::readable())?;
}
Task::Exit => {
result_sender.send(TaskResult::Exit)?;
set_readiness.set_readiness(Ready::readable())?;
break;
}
}
}
Ok(())
});
let executor = thread::spawn(move || {
let mut events = Events::with_capacity(1024);
'outer: loop {
poll.poll(&mut events, Some(Duration::from_secs(1)))?;
for event in events.iter() {
match event.token() {
FS_TOKEN => {
loop {
match result_receiver.try_recv() {
Ok(result) => match result {
TaskResult::ReadToString(value, callback, fs) => {
callback(value, fs)?
}
TaskResult::Open(file, callback, fs) => callback(file, fs)?,
TaskResult::Exit => break 'outer,
},
Err(e) => match e {
TryRecvError::Empty => break,
TryRecvError::Disconnected => return Err(e.into()),
},
}
}
poll.reregister(
®istration,
FS_TOKEN,
Ready::readable(),
PollOpt::oneshot(),
)?;
}
_ => unreachable!(),
}
}
}
Ok(())
});
(
Fs { task_sender },
FsHandler {
io_worker,
executor,
},
)
}
impl Fs {
pub fn println(&self, string: String) -> Result<(), Error> {
Ok(self.task_sender.send(Task::Println(string))?)
}
pub fn open<F>(&self, path: &str, callback: F) -> Result<(), Error>
where
F: FnOnce(File, Fs) -> Result<(), Error> + Sync + Send + 'static,
{
Ok(self.task_sender.send(Task::Open(
path.to_string(),
Box::new(callback),
self.clone(),
))?)
}
pub fn read_to_string<F>(&self, file: File, callback: F) -> Result<(), Error>
where
F: FnOnce(String, Fs) -> Result<(), Error> + Sync + Send + 'static,
{
Ok(self
.task_sender
.send(Task::ReadToString(file, Box::new(callback), self.clone()))?)
}
pub fn close(&self) -> Result<(), Error> {
Ok(self.task_sender.send(Task::Exit)?)
}
}
impl FsHandler {
pub fn join(self) -> Result<(), Error> {
self.io_worker.join().unwrap()?;
self.executor.join().unwrap()
}
}
type FileCallback = Box<dyn FnOnce(File, Fs) -> Result<(), Error> + Sync + Send>;
type StringCallback = Box<dyn FnOnce(String, Fs) -> Result<(), Error> + Sync + Send>;
pub enum Task {
Exit,
Println(String),
Open(String, FileCallback, Fs),
ReadToString(File, StringCallback, Fs),
}
pub enum TaskResult {
Exit,
Open(File, FileCallback, Fs),
ReadToString(String, StringCallback, Fs),
}
// examples/fs-mio.rs
use asyncio::fs_mio::fs_async;
use failure::Error;
const TEST_FILE_VALUE: &str = "Hello, World!\n";
fn main() -> Result<(), Error> {
let (fs, fs_handler) = fs_async();
fs.open("./examples/test.txt", |file, fs| {
fs.read_to_string(file, |value, fs| {
assert_eq!(TEST_FILE_VALUE, &value);
fs.println(value)?;
fs.close()
})
})?;
fs_handler.join()?;
Ok(())
}
running this example:
cargo run --example fs-mio
We can see the difference between two implementations. On the one hand, executor will never be blocking by result_receiver.recv()
, instead, it will wait for Poll::poll
returning; on the other hand, io worker thread will execute set_readiness.set_readiness(Ready::readable())?
after executing result_sender.send
, to inform executor there are some events happens.
In this case, executor will never be blocked by io worker, because we can register all events in executor, and mio::Poll
will listen to all events (eg. combine fs-mio
with tcp
into a file server).
Before writing a file server, we should talk about the problems of callback.
Callback can make code confusing, you can realize it more clearly when handling error, like this:
use asyncio::fs_mio::fs_async;
use failure::Error;
const TEST_FILE_VALUE: &str = "Hello, World!";
fn main() -> Result<(), Error> {
let (fs, fs_handler) = fs_async();
fs.open("./examples/test.txt",
|file, fs| {
fs.read_to_string(file,
|value, fs| {
assert_eq!(TEST_FILE_VALUE, &value);
fs.println(value,
|err| {
...
}
);
fs.close()
},
|err| {
...
}
)
},
|err| {
...
}
)?;
fs_handler.join()?;
Ok(())
}
Moreover, there is a lifetime problem in rust when we use closure, which means, we have to clone a environment variable if we want to borrow it in closure (if it implements Clone
), otherwise we should deliver its reference as a parameter of closure (you should change signature of closure as you need, what a shit!).
Considering a variety of reasons, rust
eventully uses coroutine
as its asynchronous API abstraction.
coroutine
in this blog refers to stackless coroutine
based on rust generator
instead of green thread(stackful coroutine)
which is obsoleted earlier.
rust
introduced generator
at May of this year, however, it's still unstable and unsafe. Here is a typical Fibonacci sequence generator:
// examples/fab.rs
// cargo +nightly run --example fab
#![feature(generators, generator_trait)]
use std::ops::{Generator, GeneratorState};
use std::pin::Pin;
fn main() {
let mut gen = fab(5);
loop {
match unsafe { Pin::new_unchecked(&mut gen).resume() } {
GeneratorState::Yielded(value) => println!("yield {}", value),
GeneratorState::Complete(ret) => {
println!("return {}", ret);
break;
}
}
}
}
fn fab(mut n: u64) -> impl Generator<Yield = u64, Return = u64> {
move || {
let mut last = 0u64;
let mut current = 1;
yield last;
while n > 0 {
yield current;
let tmp = last;
last = current;
current = tmp + last;
n -= 1;
}
return last;
}
}
Because of the "interrupt behaviors" of generator
, we will naturally consider to combine it with mio
: assign a token
for each generator
, then poll, resume corresponding generator
when receive a event; register an awaking event and yield before each generator is going to block. Can we implement non-blocking IO in "synchronous code" on this way?
It seems to work well in theory, but there are still "two dark clouds"。
The first "dark cloud" is relevant to memory management of rust
.
If you write a generator
like this:
fn self_ref_generator() -> impl Generator<Yield=u64, Return=()> {
|| {
let x: u64 = 1;
let ref_x: &u64 = &x;
yield 0;
yield *ref_x;
}
}
rustc
will refuse to compile and tell you "borrow may still be in use when generator yields". You may be a little panic because rustc
doesn't tell you how to fix it. Going to google it, you will find it is relevant to implementation of generator
.
As memtioned earlier, generator
is stackless, which means rustc
doesn't reserve a complete "stack" for each generator, instead, only variables and values required by a specific "state" will be reserved.
This code is valid:
fn no_ref_generator() -> impl Generator<Yield=u64, Return=()> {
|| {
let x: u64 = 1;
let ref_x: &u64 = &x;
yield *ref_x;
yield 0;
}
}
Because rustc
knows the only variable or value need be reserved after the first "yield" is literal 0
. However, for the self_ref_generator
, rustc
should reserve both of variable x
and its reference ref_x
after the first "yield". In this case, generator should be compiled into a structure like this:
enum SomeGenerator<'a> {
...
SomeState {
x: u64
ref_x: &'a u64
}
...
}
This is the notorious "self-referential structs" in rust
, what will happen when you try to compile code like this?
struct A<'a> {
b: u64,
ref_b: Option<&'a u64>
}
impl<'a> A<'a> {
fn new() -> Self {
let mut a = A{b: 1, ref_b: None};
a.ref_b = Some(&a.b);
a
}
}
Of course, rustc
will refuse to compile it. It's reasonable, variable a
on stack will be copied and dropped and its field ref_b will be invalid when function new
returns. Lifetime rules of rust
helps you avoid this memory problem.
However, even if you write code like this:
use std::borrow::{BorrowMut};
struct A<'a> {
b: u64,
ref_b: Option<&'a u64>
}
impl<'a> A<'a> {
fn boxed() -> Box<Self> {
let mut a = Box::new(A{b: 1, ref_b: None});
let mut_ref: &mut A = a.borrow_mut();
mut_ref.ref_b = Some(&mut_ref.b);
a
}
}
rustc
still refuses to compile it. It's unreasonable,variable a
on heap will not be dropped after function new
returns, and its field ref_b
should be always valid. However, rustc
doesn't know, and you cannot prove it in the language that compiler can understand.
Moreover, you even cannot mutably borrow self-referential structs, like this:
struct A<'a> {
b: u64,
ref_b: Option<&'a u64>
}
impl<'a> A<'a> {
fn new() -> Self {
A{b: 1, ref_b: None}
}
fn mute(&mut self) {
}
}
fn main() {
let mut a = A::new();
a.ref_b = Some(&a.b);
a.mute();
}
rustc
still refuses to compile it. It's awful, because the signature of earlier Future::poll
is:
fn poll(&mut self) -> Poll<Self::Item, Self::Error>;
and the signature of Generator::resume
is still:
unsafe fn resume(&mut self) -> GeneratorState<Self::Yield, Self::Return>;
As a result, self-reference will lead to unable implementation of trait Generator
and trait Future
. In this case, we can use NonNull
to avoid compiler checking:
use std::ptr::NonNull;
struct A {
b: u64,
ref_b: NonNull<u64>
}
impl A {
fn new() -> Self {
A{b: 1, ref_b: NonNull::dangling()}
}
}
fn main() {
let mut a = A::new();
a.ref_b = NonNull::from(&a.b);
}
However, you should guarantee memory safety by yourself (self-referential structs MUST NOT be moved, and you MUST NOT deliver its mutable reference to mem::replace
or mem::swap
), it's not a nice solution.
Can we find some ways to guarantee its moving and mutably borrowing cannot be safe? rust
introduces Pin
to hold this job. Specifications of pin
can be found in this RFC, this blog will only introduce it simply.
rust
implement trait std::marker::Unpin
for almost all types by default. It's only a marker to indicate safely moving of a type. For types marked as Unpin
, Pin<&'a mut T>
and &'a mut T
have no difference, you can safely exchange them by Pin::new(&mut T)
and Pin::get_mut(this: Pin<&mut T>)
.
However, for types that cannot be moved safely, like the A
mentioned earlier, we should mark it as !Unpin
first, a safe way is to give it a field whose type is marked !Unpin
, for example, Pinned
.
#![feature(pin)]
use std::marker::{Pinned};
use std::ptr::NonNull;
struct A {
b: u64,
ref_b: NonNull<u64>,
_pin: Pinned,
}
impl A {
fn new() -> Self {
A {
b: 1,
ref_b: NonNull::dangling(),
_pin: Pinned,
}
}
}
fn main() {
let mut a = A::new();
let mut pinned = unsafe { Pin::new_unchecked(&mut a) };
let ref_b = NonNull::from(&pinned.b);
let mut_ref: Pin<&mut A> = pinned.as_mut();
unsafe {Pin::get_mut_unchecked(mut_ref).ref_b = ref_b};
let unmoved = pinned;
assert_eq!(unmoved.ref_b, NonNull::from(&unmoved.b));
}
For types marked as !Unpin
, Pin<&'a mut T>
and &'a mut T
cannot be safely exchanged, you can unsafely exchange them by Pin::new_unchecked
and Pin::get_mut_unchecked
. We can always guarantee the safety in the scope we construct it, so after calling two unsafe methods, we can guarantee:
- we can never get mutable reference safely:
Pin::get_mut_unchecked
is unsafe - we can never move it: because
Pin
only owns a mutable reference, andPin::get_mut_unchecked
is unsafe, so deliver the mutable reference intomem::replace
andmem::swap
is unsafe
Of course, if you don't want to construct Pin
in unsafe way or you want Pin
to own the ownership of the instance, you can use Box::pin
thus allocate instance on the heap.
struct A {
b: u64,
ref_b: NonNull<u64>,
_pin: Pinned,
}
impl A {
fn boxed() -> Pin<Box<Self>> {
let mut boxed = Box::pin(A {
b: 1,
ref_b: NonNull::dangling(),
_pin: Pinned,
});
let ref_b = NonNull::from(&boxed.b);
let mut_ref: Pin<&mut A> = boxed.as_mut();
unsafe { Pin::get_mut_unchecked(mut_ref).ref_b = ref_b };
boxed
}
}
fn main() {
let boxed = A::boxed();
let unmoved = boxed;
assert_eq!(unmoved.ref_b, NonNull::from(&unmoved.b));
}
After introducing of Pin
, the new Future
is defined as:
pub trait Future {
type Output;
fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output>;
}
The second "dark cloud" is relevant to API abstraction.
Now that rust
chooses coroutine
as API abstraction of non-blocking IO, what should be introduced into key words, what should be introduced into standard library and what should be implemented by community? Let developers call unsafe Generator::resume
is inapposite, using mio
as the only specified low-level non-blocking IO implementation is also unreasonable.
Up to now, rust
supports:
- key words
async
await
- standard libraries
std::future
trait Future
trait GenFuture
std::task
enum Poll<T>
struct Context
struct Waker
struct RawWaker
RawWaker
consists of an data ptr and a virtual table, you can construct one with SetReadiness
and implement fn wake(*const ())
by SetReadiness::set_readiness
. Then you should wrap your raw waker in Waker
and Context
.
Poll<T>
is defined as:
pub enum Poll<T> {
Ready(T),
Pending,
}
await
is the first suffix keyword in rust, which can only be used in async
block or function.
future.await
will be expanded into:
loop {
if let Poll::Ready(x) = std::future::poll_with_tls_context(unsafe{
Pin::new_unchecked(&mut future)
}) {
break x;
}
yield
}
std::future::poll_with_tls_context
will poll with "thread local context" via a thread-local variable TLS_CX
.
async
is used to wrap a Generator
into a GenFuture
. GenFuture
is defined as:
struct GenFuture<T: Generator<Yield = ()>>(T);
impl<T: Generator<Yield = ()>> !Unpin for GenFuture<T> {}
impl<T: Generator<Yield = ()>> Future for GenFuture<T> {
type Output = T::Return;
fn poll(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let gen = unsafe { Pin::map_unchecked_mut(self, |s| &mut s.0) };
set_task_context(cx, || match gen.resume() {
GeneratorState::Yielded(()) => Poll::Pending,
GeneratorState::Complete(x) => Poll::Ready(x),
})
}
}
pub fn from_generator<T: Generator<Yield = ()>>(x: T) -> impl Future<Output = T::Return> {
GenFuture(x)
}
We can see, GenFuture
will call set_task_context
before calling self.0.resume
, thus code in generator can get this Context
via TLS_CX
.
So, code like this:
async fn async_recv(string_channel: Receiver<String>) -> String {
string_channel.recv_future().await
}
will be expanded into:
fn async_recv(string_channel: Receiver<String>) -> impl Future<Output = T::Return> {
from_generator(move || {
let recv_future = string_channel.recv_future();
loop {
if let Poll::Ready(x) = std::future::poll_with_tls_context(unsafe{
Pin::new_unchecked(&mut recv_future)
}) {
break x;
}
yield
}
})
}
Mastering all the basic knowledge mentioned above, we can do some practices.
coroutine
doesn't mean "non-blocking", you can invoke blocking API in async blocks or functions. The key to non-blocking IO is, when GenFuture
is going to block (eg. an API returns io::ErrorKind::WouldBlock
), register a source task by local waker and sleep (yield
), lower-level non-blocking scheduler will awake this GenFuture
after task completed.
In src/executor.rs, I implement Executor
, block_on
, spawn
, TcpListener
and TcpStream
. Code is a little long, you had better clone and view it in editor.
Be careful to distinguish
Poll
(mio::Poll
) fromtask::Poll
and to distinguishnet::{TcpListener, TcpStream}
(mio::net::{TcpListener, TcpStream}
) fromTcpListener, TcpStream
Executor
is a struct containing mio::Poll
, main task waker and two Slab
s for managing Task
and Source
. I don't implement any special methods for it, its only duty is to be initialized as thread local variable EXECUTOR
and to be borrowed by other functions.
This function will block current thread, the only parameter is main_task: Future<Output=T>
; type of return value is T
. This function is generally called by main function.
block_on
borrows thread local variable EXECUTOR
, its main logic loop will call mio::Poll::poll
to wait for events. I classify all tokens (0 - MAX_RESOURCE_NUM(1 << 31)
) into three kinds:
-
main task token
receiving
Token
whose value isMAIN_TASK_TOKEN (1 << 31)
means main task need be awaked,main_task.poll
will be called,block_on
will returnOk(ret)
ifmain_task.poll
returnstask::Poll::Ready(ret)
. -
task token
odd
Token
means corresponding task (spawned by functionspawn
) need be awaked,task.poll
will be invoked,block_on
will returnErr(err)
iftask.poll
returnsErr(err)
. -
source token
even
Token
means corresponding source (registered by functionregister_source
) is completed,source.task_waker.waker
will be invoked to awake task which registered it.
Function to spawn tasks.
wrapper
for mio::net::TcpListener
, method accept
will return a Future
.
wrapper
for mio::net::TcpStream
, method read
and write
will both return a Future
.
Implemented executor
, we can write a simple echo server:
// examples/async-echo
#[macro_use]
extern crate log;
use asyncio::executor::{block_on, spawn, TcpListener};
use failure::Error;
fn main() -> Result<(), Error> {
env_logger::init();
block_on(
async {
let mut listener = TcpListener::bind(&"127.0.0.1:7878".parse()?)?;
info!("Listening on 127.0.0.1:7878");
while let Ok((mut stream, addr)) = listener.accept().await {
info!("connection from {}", addr);
spawn(
async move {
let client_hello = stream.read().await?;
let read_length = client_hello.len();
let write_length =
stream.write(client_hello).await?;
assert_eq!(read_length, write_length);
stream.close();
Ok(())
},
)?;
};
Ok(())
},
)?
}
run
RUST_LOG=info cargo run --example async-echo
You can test it using telnet
.
However, to run the example metioned at the beginning, we should implement a non-blocking FS-IO based on mio
and std::future
. Here is the final implementation: src/fs_future.rs.
Now, let's run:
RUST_LOG=info cargo run --example file-server
Test it using telnet
:
[~] telnet 127.0.0.1 7878
Trying 127.0.0.1...
Connected to localhost.
Escape character is '^]'.
Please enter filename: examples/test.txt
Hello, World!
Connection closed by foreign host.
You can look up source code by yourself if you are interested in it.