Initial impl for async-std support (#87)

* initial impl for async-std support
This commit is contained in:
Nikolay Kim 2021-12-28 23:58:01 +06:00 committed by GitHub
parent 7751e944f4
commit 56ed50c6e8
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
21 changed files with 946 additions and 51 deletions

View file

@ -56,6 +56,13 @@ jobs:
command: test command: test
args: --all --all-features --no-fail-fast -- --nocapture args: --all --all-features --no-fail-fast -- --nocapture
- name: Run async-std tests
timeout-minutes: 40
continue-on-error: true
run: |
cd ntex
cargo test --no-default-features --no-fail-fast --features="async-std,cookie,url,compress,openssl,rustls" --lib -- --test-threads 1
- name: Install tarpaulin - name: Install tarpaulin
if: matrix.version == '1.56.0' && (github.ref == 'refs/heads/master' || github.event_name == 'pull_request') if: matrix.version == '1.56.0' && (github.ref == 'refs/heads/master' || github.event_name == 'pull_request')
continue-on-error: true continue-on-error: true
@ -67,6 +74,8 @@ jobs:
continue-on-error: true continue-on-error: true
run: | run: |
cargo tarpaulin --out Xml --all --all-features cargo tarpaulin --out Xml --all --all-features
cd ntex
cargo tarpaulin --out Xml --output-dir=.. --no-default-features --features="async-std,cookie,url,compress,openssl,rustls" --lib
- name: Upload to Codecov - name: Upload to Codecov
if: matrix.version == '1.56.0' && (github.ref == 'refs/heads/master' || github.event_name == 'pull_request') if: matrix.version == '1.56.0' && (github.ref == 'refs/heads/master' || github.event_name == 'pull_request')

View file

@ -51,11 +51,12 @@
//! [struct docs]: struct.Bytes.html //! [struct docs]: struct.Bytes.html
#![deny( #![deny(
warnings, // warnings,
// missing_docs, // missing_docs,
// missing_debug_implementations, // missing_debug_implementations,
rust_2018_idioms rust_2018_idioms
)] )]
#![allow(clippy::return_self_not_must_use)]
#![doc(html_root_url = "https://docs.rs/ntex-bytes/")] #![doc(html_root_url = "https://docs.rs/ntex-bytes/")]
pub mod buf; pub mod buf;

View file

@ -1,5 +1,9 @@
# Changes # Changes
## [0.1.0-b.9] - 2021-12-29
* Add `async-std` support
## [0.1.0-b.8] - 2021-12-28 ## [0.1.0-b.8] - 2021-12-28
* Fix error handing for nested filters * Fix error handing for nested filters

View file

@ -1,6 +1,6 @@
[package] [package]
name = "ntex-io" name = "ntex-io"
version = "0.1.0-b.8" version = "0.1.0-b.9"
authors = ["ntex contributors <team@ntex.rs>"] authors = ["ntex contributors <team@ntex.rs>"]
description = "Utilities for encoding and decoding frames" description = "Utilities for encoding and decoding frames"
keywords = ["network", "framework", "async", "futures"] keywords = ["network", "framework", "async", "futures"]
@ -16,7 +16,7 @@ name = "ntex_io"
path = "src/lib.rs" path = "src/lib.rs"
[features] [features]
default = ["tokio"] default = ["tokio-traits"]
# tokio traits support # tokio traits support
tokio-traits = ["tok-io/net", "tok-io/rt"] tokio-traits = ["tok-io/net", "tok-io/rt"]
@ -24,6 +24,9 @@ tokio-traits = ["tok-io/net", "tok-io/rt"]
# tokio runtime support # tokio runtime support
tokio = ["tok-io/net", "tok-io/rt"] tokio = ["tok-io/net", "tok-io/rt"]
# async-std runtime support
async-std = ["async_std/unstable"]
[dependencies] [dependencies]
ntex-codec = "0.6.0" ntex-codec = "0.6.0"
ntex-bytes = "0.1.8" ntex-bytes = "0.1.8"
@ -36,6 +39,7 @@ log = "0.4"
pin-project-lite = "0.2" pin-project-lite = "0.2"
tok-io = { version = "1", package = "tokio", default-features = false, optional = true } tok-io = { version = "1", package = "tokio", default-features = false, optional = true }
async_std = { version = "1", package = "async-std", optional = true }
[dev-dependencies] [dev-dependencies]
ntex = "0.5.0-b.2" ntex = "0.5.0-b.2"

View file

@ -0,0 +1,38 @@
#![allow(dead_code)]
//! async net providers
use ntex_util::future::lazy;
use std::future::Future;
/// Spawn a future on the current thread. This does not create a new Arbiter
/// or Arbiter address, it is simply a helper for spawning futures on the current
/// thread.
///
/// # Panics
///
/// This function panics if ntex system is not running.
#[inline]
pub fn spawn<F>(f: F) -> async_std::task::JoinHandle<F::Output>
where
F: Future + 'static,
{
async_std::task::spawn_local(f)
}
/// Executes a future on the current thread. This does not create a new Arbiter
/// or Arbiter address, it is simply a helper for executing futures on the current
/// thread.
///
/// # Panics
///
/// This function panics if ntex system is not running.
#[inline]
pub fn spawn_fn<F, R>(f: F) -> async_std::task::JoinHandle<R::Output>
where
F: FnOnce() -> R + 'static,
R: Future + 'static,
{
spawn(async move {
let r = lazy(|_| f()).await;
r.await
})
}

View file

@ -1,3 +1,5 @@
//! Utilities for abstructing io streams
#![allow(clippy::return_self_not_must_use)]
use std::{ use std::{
any::Any, any::TypeId, fmt, future::Future, io as sio, io::Error as IoError, any::Any, any::TypeId, fmt, future::Future, io as sio, io::Error as IoError,
task::Context, task::Poll, task::Context, task::Poll,
@ -16,9 +18,11 @@ mod tasks;
mod time; mod time;
mod utils; mod utils;
#[cfg(feature = "async-std")]
mod asyncstd_rt;
#[cfg(any(feature = "tokio-traits", feature = "tokio"))] #[cfg(any(feature = "tokio-traits", feature = "tokio"))]
mod tokio_impl; mod tokio_impl;
#[cfg(any(feature = "tokio"))] #[cfg(feature = "tokio")]
mod tokio_rt; mod tokio_rt;
use ntex_bytes::BytesMut; use ntex_bytes::BytesMut;
@ -165,6 +169,9 @@ pub mod rt {
#[cfg(feature = "tokio")] #[cfg(feature = "tokio")]
pub use crate::tokio_rt::*; pub use crate::tokio_rt::*;
#[cfg(all(not(feature = "tokio"), feature = "async-std"))]
pub use crate::asyncstd_rt::*;
} }
#[cfg(test)] #[cfg(test)]

View file

@ -221,6 +221,7 @@ impl Future for WriteTask {
log::trace!( log::trace!(
"write task is closed with err during flush" "write task is closed with err during flush"
); );
this.state.close(None);
return Poll::Ready(()); return Poll::Ready(());
} }
_ => (), _ => (),
@ -259,10 +260,11 @@ impl Future for WriteTask {
} }
Poll::Pending => { Poll::Pending => {
*count += read_buf.filled().len() as u16; *count += read_buf.filled().len() as u16;
if *count > 8196 { if *count > 4096 {
log::trace!( log::trace!(
"write task is stopped, too much input" "write task is stopped, too much input"
); );
this.state.close(None);
return Poll::Ready(()); return Poll::Ready(());
} }
break; break;
@ -638,6 +640,7 @@ mod unixstream {
log::trace!( log::trace!(
"write task is closed with err during flush" "write task is closed with err during flush"
); );
this.state.close(None);
return Poll::Ready(()); return Poll::Ready(());
} }
_ => (), _ => (),
@ -677,10 +680,11 @@ mod unixstream {
} }
Poll::Pending => { Poll::Pending => {
*count += read_buf.filled().len() as u16; *count += read_buf.filled().len() as u16;
if *count > 8196 { if *count > 4096 {
log::trace!( log::trace!(
"write task is stopped, too much input" "write task is stopped, too much input"
); );
this.state.close(None);
return Poll::Ready(()); return Poll::Ready(());
} }
break; break;

View file

@ -1,5 +1,9 @@
# Changes # Changes
## [0.4.0-b.3] - 2021-12-28
* Add `async-std` support
## [0.4.0-b.1] - 2021-12-22 ## [0.4.0-b.1] - 2021-12-22
* Fix lifetimes for unix_connect/unix_connect_in * Fix lifetimes for unix_connect/unix_connect_in

View file

@ -1,6 +1,6 @@
[package] [package]
name = "ntex-rt" name = "ntex-rt"
version = "0.4.0-b.2" version = "0.4.0-b.3"
authors = ["ntex contributors <team@ntex.rs>"] authors = ["ntex contributors <team@ntex.rs>"]
description = "ntex runtime" description = "ntex runtime"
keywords = ["network", "framework", "async", "futures"] keywords = ["network", "framework", "async", "futures"]
@ -16,18 +16,23 @@ name = "ntex_rt"
path = "src/lib.rs" path = "src/lib.rs"
[features] [features]
default = ["tokio"] default = []
# tokio support # tokio support
tokio = ["tok-io", "ntex-io/tokio"] tokio = ["ntex-io/tokio", "tok-io"]
# async-std support
async-std = ["ntex-io/async-std", "async_std/unstable"]
[dependencies] [dependencies]
ntex-bytes = "0.1.8" ntex-bytes = "0.1.8"
ntex-io = "0.1.0-b.5" ntex-io = "0.1.0-b.9"
ntex-util = "0.1.3" ntex-util = "0.1.3"
async-oneshot = "0.5.0" async-oneshot = "0.5.0"
async-channel = "1.6.1" async-channel = "1.6.1"
derive_more = "0.99.14"
log = "0.4" log = "0.4"
pin-project-lite = "0.2" pin-project-lite = "0.2"
tok-io = { version = "1", package = "tokio", default-features = false, features = ["rt", "net", "signal"], optional = true } tok-io = { version = "1", package = "tokio", default-features = false, features = ["rt", "net", "signal"], optional = true }
async_std = { version = "1", package = "async-std", optional = true }

View file

@ -265,7 +265,7 @@ impl Future for ArbiterController {
return Poll::Ready(()); return Poll::Ready(());
} }
ArbiterCommand::Execute(fut) => { ArbiterCommand::Execute(fut) => {
tok_io::task::spawn(fut); crate::spawn(fut);
} }
ArbiterCommand::ExecuteFn(f) => { ArbiterCommand::ExecuteFn(f) => {
f.call_box(); f.call_box();

818
ntex-rt/src/asyncstd.rs Normal file
View file

@ -0,0 +1,818 @@
#![allow(dead_code)]
use std::future::Future;
use std::task::{Context, Poll};
use std::{any, cell::RefCell, io, net, net::SocketAddr, pin::Pin, rc::Rc};
use async_oneshot as oneshot;
use async_std::io::{Read, Write};
use ntex_bytes::{Buf, BufMut, BytesMut, PoolRef};
use ntex_io::{
types, Handle, Io, IoStream, ReadContext, ReadStatus, WriteContext, WriteStatus,
};
use ntex_util::{future::lazy, ready, time::sleep, time::Sleep};
use crate::{Runtime, Signal};
#[derive(Debug, Copy, Clone, derive_more::Display)]
pub struct JoinError;
impl std::error::Error for JoinError {}
#[derive(Clone)]
struct TcpStream(async_std::net::TcpStream);
#[cfg(unix)]
#[derive(Clone)]
struct UnixStream(async_std::os::unix::net::UnixStream);
/// Create new single-threaded async-std runtime.
pub fn create_runtime() -> Box<dyn Runtime> {
Box::new(AsyncStdRuntime::new().unwrap())
}
/// Opens a TCP connection to a remote host.
pub async fn tcp_connect(addr: SocketAddr) -> Result<Io, io::Error> {
let sock = async_std::net::TcpStream::connect(addr).await?;
sock.set_nodelay(true)?;
Ok(Io::new(TcpStream(sock)))
}
/// Opens a TCP connection to a remote host and use specified memory pool.
pub async fn tcp_connect_in(addr: SocketAddr, pool: PoolRef) -> Result<Io, io::Error> {
let sock = async_std::net::TcpStream::connect(addr).await?;
sock.set_nodelay(true)?;
Ok(Io::with_memory_pool(TcpStream(sock), pool))
}
#[cfg(unix)]
/// Opens a unix stream connection.
pub async fn unix_connect<P>(addr: P) -> Result<Io, io::Error>
where
P: AsRef<async_std::path::Path>,
{
let sock = async_std::os::unix::net::UnixStream::connect(addr).await?;
Ok(Io::new(UnixStream(sock)))
}
#[cfg(unix)]
/// Opens a unix stream connection and specified memory pool.
pub async fn unix_connect_in<P>(addr: P, pool: PoolRef) -> Result<Io, io::Error>
where
P: AsRef<async_std::path::Path>,
{
let sock = async_std::os::unix::net::UnixStream::connect(addr).await?;
Ok(Io::with_memory_pool(UnixStream(sock), pool))
}
/// Convert std TcpStream to async-std's TcpStream
pub fn from_tcp_stream(stream: net::TcpStream) -> Result<Io, io::Error> {
stream.set_nonblocking(true)?;
stream.set_nodelay(true)?;
Ok(Io::new(TcpStream(async_std::net::TcpStream::from(stream))))
}
#[cfg(unix)]
/// Convert std UnixStream to async-std's UnixStream
pub fn from_unix_stream(stream: std::os::unix::net::UnixStream) -> Result<Io, io::Error> {
stream.set_nonblocking(true)?;
Ok(Io::new(UnixStream(From::from(stream))))
}
/// Spawn a future on the current thread. This does not create a new Arbiter
/// or Arbiter address, it is simply a helper for spawning futures on the current
/// thread.
///
/// # Panics
///
/// This function panics if ntex system is not running.
#[inline]
pub fn spawn<F>(f: F) -> JoinHandle<F::Output>
where
F: Future + 'static,
{
JoinHandle {
fut: async_std::task::spawn_local(f),
}
}
/// Executes a future on the current thread. This does not create a new Arbiter
/// or Arbiter address, it is simply a helper for executing futures on the current
/// thread.
///
/// # Panics
///
/// This function panics if ntex system is not running.
#[inline]
pub fn spawn_fn<F, R>(f: F) -> JoinHandle<R::Output>
where
F: FnOnce() -> R + 'static,
R: Future + 'static,
{
spawn(async move {
let r = lazy(|_| f()).await;
r.await
})
}
/// Spawns a blocking task.
///
/// The task will be spawned onto a thread pool specifically dedicated
/// to blocking tasks. This is useful to prevent long-running synchronous
/// operations from blocking the main futures executor.
pub fn spawn_blocking<F, T>(f: F) -> JoinHandle<T>
where
F: FnOnce() -> T + Send + 'static,
T: Send + 'static,
{
JoinHandle {
fut: async_std::task::spawn_blocking(f),
}
}
pub struct JoinHandle<T> {
fut: async_std::task::JoinHandle<T>,
}
impl<T> Future for JoinHandle<T> {
type Output = Result<T, JoinError>;
fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
Poll::Ready(Ok(ready!(Pin::new(&mut self.fut).poll(cx))))
}
}
thread_local! {
static SRUN: RefCell<bool> = RefCell::new(false);
static SHANDLERS: Rc<RefCell<Vec<oneshot::Sender<Signal>>>> = Default::default();
}
/// Register signal handler.
///
/// Signals are handled by oneshots, you have to re-register
/// after each signal.
pub fn signal() -> Option<oneshot::Receiver<Signal>> {
if !SRUN.with(|v| *v.borrow()) {
spawn(Signals::new());
}
SHANDLERS.with(|handlers| {
let (tx, rx) = oneshot::oneshot();
handlers.borrow_mut().push(tx);
Some(rx)
})
}
/// Single-threaded async-std runtime.
#[derive(Debug)]
struct AsyncStdRuntime {}
impl AsyncStdRuntime {
/// Returns a new runtime initialized with default configuration values.
fn new() -> io::Result<Self> {
Ok(Self {})
}
}
impl Runtime for AsyncStdRuntime {
/// Spawn a future onto the single-threaded runtime.
fn spawn(&self, future: Pin<Box<dyn Future<Output = ()>>>) {
async_std::task::spawn_local(future);
}
/// Runs the provided future, blocking the current thread until the future
/// completes.
fn block_on(&self, f: Pin<Box<dyn Future<Output = ()>>>) {
// set ntex-util spawn fn
ntex_util::set_spawn_fn(|fut| {
async_std::task::spawn_local(fut);
});
async_std::task::block_on(f);
}
}
struct Signals {}
impl Signals {
pub(super) fn new() -> Signals {
Self {}
}
}
impl Future for Signals {
type Output = ();
fn poll(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll<Self::Output> {
Poll::Ready(())
}
}
impl IoStream for TcpStream {
fn start(self, read: ReadContext, write: WriteContext) -> Option<Box<dyn Handle>> {
spawn(ReadTask::new(self.clone(), read));
spawn(WriteTask::new(self.clone(), write));
Some(Box::new(self))
}
}
impl Handle for TcpStream {
fn query(&self, id: any::TypeId) -> Option<Box<dyn any::Any>> {
if id == any::TypeId::of::<types::PeerAddr>() {
if let Ok(addr) = self.0.peer_addr() {
return Some(Box::new(types::PeerAddr(addr)));
}
}
None
}
}
/// Read io task
struct ReadTask {
io: TcpStream,
state: ReadContext,
}
impl ReadTask {
/// Create new read io task
fn new(io: TcpStream, state: ReadContext) -> Self {
Self { io, state }
}
}
impl Future for ReadTask {
type Output = ();
fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let mut this = self.as_mut();
loop {
match ready!(this.state.poll_ready(cx)) {
ReadStatus::Ready => {
let pool = this.state.memory_pool();
let mut buf = this.state.get_read_buf();
let io = &mut this.io;
let (hw, lw) = pool.read_params().unpack();
// read data from socket
let mut new_bytes = 0;
let mut close = false;
let mut pending = false;
loop {
// make sure we've got room
let remaining = buf.remaining_mut();
if remaining < lw {
buf.reserve(hw - remaining);
}
match poll_read_buf(Pin::new(&mut io.0), cx, &mut buf) {
Poll::Pending => {
pending = true;
break;
}
Poll::Ready(Ok(n)) => {
if n == 0 {
log::trace!("async-std stream is disconnected");
close = true;
} else {
new_bytes += n;
if new_bytes <= hw {
continue;
}
}
break;
}
Poll::Ready(Err(err)) => {
log::trace!("read task failed on io {:?}", err);
let _ = this.state.release_read_buf(buf, new_bytes);
this.state.close(Some(err));
return Poll::Ready(());
}
}
}
if new_bytes == 0 && close {
this.state.close(None);
return Poll::Ready(());
}
this.state.release_read_buf(buf, new_bytes);
return if close {
this.state.close(None);
Poll::Ready(())
} else if pending {
Poll::Pending
} else {
continue;
};
}
ReadStatus::Terminate => {
log::trace!("read task is instructed to shutdown");
return Poll::Ready(());
}
}
}
}
}
#[derive(Debug)]
enum IoWriteState {
Processing(Option<Sleep>),
Shutdown(Sleep, Shutdown),
}
#[derive(Debug)]
enum Shutdown {
None,
Stopping(u16),
}
/// Write io task
struct WriteTask {
st: IoWriteState,
io: TcpStream,
state: WriteContext,
}
impl WriteTask {
/// Create new write io task
fn new(io: TcpStream, state: WriteContext) -> Self {
Self {
io,
state,
st: IoWriteState::Processing(None),
}
}
}
impl Future for WriteTask {
type Output = ();
fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let mut this = self.as_mut().get_mut();
match this.st {
IoWriteState::Processing(ref mut delay) => {
match this.state.poll_ready(cx) {
Poll::Ready(WriteStatus::Ready) => {
if let Some(delay) = delay {
if delay.poll_elapsed(cx).is_ready() {
this.state.close(Some(io::Error::new(
io::ErrorKind::TimedOut,
"Operation timedout",
)));
return Poll::Ready(());
}
}
// flush framed instance
match flush_io(&mut this.io.0, &this.state, cx) {
Poll::Pending | Poll::Ready(true) => Poll::Pending,
Poll::Ready(false) => Poll::Ready(()),
}
}
Poll::Ready(WriteStatus::Timeout(time)) => {
log::trace!("initiate timeout delay for {:?}", time);
if delay.is_none() {
*delay = Some(sleep(time));
}
self.poll(cx)
}
Poll::Ready(WriteStatus::Shutdown(time)) => {
log::trace!("write task is instructed to shutdown");
let timeout = if let Some(delay) = delay.take() {
delay
} else {
sleep(time)
};
this.st = IoWriteState::Shutdown(timeout, Shutdown::None);
self.poll(cx)
}
Poll::Ready(WriteStatus::Terminate) => {
log::trace!("write task is instructed to terminate");
let _ = Pin::new(&mut this.io.0).poll_close(cx);
this.state.close(None);
Poll::Ready(())
}
Poll::Pending => Poll::Pending,
}
}
IoWriteState::Shutdown(ref mut delay, ref mut st) => {
// close WRITE side and wait for disconnect on read side.
// use disconnect timeout, otherwise it could hang forever.
loop {
match st {
Shutdown::None => {
// flush write buffer
match flush_io(&mut this.io.0, &this.state, cx) {
Poll::Ready(true) => {
if let Err(_) =
this.io.0.shutdown(std::net::Shutdown::Write)
{
this.state.close(None);
return Poll::Ready(());
}
*st = Shutdown::Stopping(0);
continue;
}
Poll::Ready(false) => {
log::trace!(
"write task is closed with err during flush"
);
this.state.close(None);
return Poll::Ready(());
}
_ => (),
}
}
Shutdown::Stopping(ref mut count) => {
// read until 0 or err
let mut buf = [0u8; 512];
let io = &mut this.io;
loop {
match Pin::new(&mut io.0).poll_read(cx, &mut buf) {
Poll::Ready(Err(e)) => {
log::trace!("write task is stopped");
this.state.close(Some(e));
return Poll::Ready(());
}
Poll::Ready(Ok(0)) => {
log::trace!("async-std socket is disconnected");
this.state.close(None);
return Poll::Ready(());
}
Poll::Ready(Ok(n)) => {
*count += n as u16;
if *count > 4096 {
log::trace!(
"write task is stopped, too much input"
);
this.state.close(None);
return Poll::Ready(());
}
}
Poll::Pending => break,
}
}
}
}
// disconnect timeout
if delay.poll_elapsed(cx).is_pending() {
return Poll::Pending;
}
log::trace!("write task is stopped after delay");
this.state.close(None);
let _ = Pin::new(&mut this.io.0).poll_close(cx);
return Poll::Ready(());
}
}
}
}
}
/// Flush write buffer to underlying I/O stream.
pub(super) fn flush_io<T: Read + Write + Unpin>(
io: &mut T,
state: &WriteContext,
cx: &mut Context<'_>,
) -> Poll<bool> {
let mut buf = if let Some(buf) = state.get_write_buf() {
buf
} else {
return Poll::Ready(true);
};
let len = buf.len();
let pool = state.memory_pool();
if len != 0 {
// log::trace!("flushing framed transport: {:?}", buf.len());
let mut written = 0;
while written < len {
match Pin::new(&mut *io).poll_write(cx, &buf[written..]) {
Poll::Pending => break,
Poll::Ready(Ok(n)) => {
if n == 0 {
log::trace!("Disconnected during flush, written {}", written);
pool.release_write_buf(buf);
state.close(Some(io::Error::new(
io::ErrorKind::WriteZero,
"failed to write frame to transport",
)));
return Poll::Ready(false);
} else {
written += n
}
}
Poll::Ready(Err(e)) => {
log::trace!("Error during flush: {}", e);
pool.release_write_buf(buf);
state.close(Some(e));
return Poll::Ready(false);
}
}
}
log::trace!("flushed {} bytes", written);
// remove written data
let result = if written == len {
buf.clear();
if let Err(e) = state.release_write_buf(buf) {
state.close(Some(e));
return Poll::Ready(false);
}
Poll::Ready(true)
} else {
buf.advance(written);
if let Err(e) = state.release_write_buf(buf) {
state.close(Some(e));
return Poll::Ready(false);
}
Poll::Pending
};
// flush
match Pin::new(&mut *io).poll_flush(cx) {
Poll::Ready(Ok(_)) => result,
Poll::Pending => Poll::Pending,
Poll::Ready(Err(e)) => {
log::trace!("error during flush: {}", e);
state.close(Some(e));
Poll::Ready(false)
}
}
} else {
Poll::Ready(true)
}
}
pub fn poll_read_buf<T: Read>(
io: Pin<&mut T>,
cx: &mut Context<'_>,
buf: &mut BytesMut,
) -> Poll<io::Result<usize>> {
if !buf.has_remaining_mut() {
return Poll::Ready(Ok(0));
}
let dst = unsafe { &mut *(buf.chunk_mut() as *mut _ as *mut [u8]) };
let n = ready!(io.poll_read(cx, dst))?;
// Safety: This is guaranteed to be the number of initialized (and read)
// bytes due to the invariants provided by Read::poll_read() api
unsafe {
buf.advance_mut(n);
}
Poll::Ready(Ok(n))
}
#[cfg(unix)]
mod unixstream {
use super::*;
impl IoStream for UnixStream {
fn start(self, read: ReadContext, write: WriteContext) -> Option<Box<dyn Handle>> {
spawn(ReadTask::new(self.clone(), read));
spawn(WriteTask::new(self.clone(), write));
None
}
}
/// Read io task
struct ReadTask {
io: UnixStream,
state: ReadContext,
}
impl ReadTask {
/// Create new read io task
fn new(io: UnixStream, state: ReadContext) -> Self {
Self { io, state }
}
}
impl Future for ReadTask {
type Output = ();
fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let mut this = self.as_mut();
loop {
match ready!(this.state.poll_ready(cx)) {
ReadStatus::Ready => {
let pool = this.state.memory_pool();
let mut buf = this.state.get_read_buf();
let io = &mut this.io;
let (hw, lw) = pool.read_params().unpack();
// read data from socket
let mut new_bytes = 0;
let mut close = false;
let mut pending = false;
loop {
// make sure we've got room
let remaining = buf.remaining_mut();
if remaining < lw {
buf.reserve(hw - remaining);
}
match poll_read_buf(Pin::new(&mut io.0), cx, &mut buf) {
Poll::Pending => {
pending = true;
break;
}
Poll::Ready(Ok(n)) => {
if n == 0 {
log::trace!("async-std stream is disconnected");
close = true;
} else {
new_bytes += n;
if new_bytes <= hw {
continue;
}
}
break;
}
Poll::Ready(Err(err)) => {
log::trace!("read task failed on io {:?}", err);
let _ = this.state.release_read_buf(buf, new_bytes);
this.state.close(Some(err));
return Poll::Ready(());
}
}
}
if new_bytes == 0 && close {
this.state.close(None);
return Poll::Ready(());
}
this.state.release_read_buf(buf, new_bytes);
return if close {
this.state.close(None);
Poll::Ready(())
} else if pending {
Poll::Pending
} else {
continue;
};
}
ReadStatus::Terminate => {
log::trace!("read task is instructed to shutdown");
return Poll::Ready(());
}
}
}
}
}
/// Write io task
struct WriteTask {
st: IoWriteState,
io: UnixStream,
state: WriteContext,
}
impl WriteTask {
/// Create new write io task
fn new(io: UnixStream, state: WriteContext) -> Self {
Self {
io,
state,
st: IoWriteState::Processing(None),
}
}
}
impl Future for WriteTask {
type Output = ();
fn poll(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Self::Output> {
let mut this = self.as_mut().get_mut();
match this.st {
IoWriteState::Processing(ref mut delay) => {
match this.state.poll_ready(cx) {
Poll::Ready(WriteStatus::Ready) => {
if let Some(delay) = delay {
if delay.poll_elapsed(cx).is_ready() {
this.state.close(Some(io::Error::new(
io::ErrorKind::TimedOut,
"Operation timedout",
)));
return Poll::Ready(());
}
}
// flush framed instance
match flush_io(&mut this.io.0, &this.state, cx) {
Poll::Pending | Poll::Ready(true) => Poll::Pending,
Poll::Ready(false) => Poll::Ready(()),
}
}
Poll::Ready(WriteStatus::Timeout(time)) => {
log::trace!("initiate timeout delay for {:?}", time);
if delay.is_none() {
*delay = Some(sleep(time));
}
self.poll(cx)
}
Poll::Ready(WriteStatus::Shutdown(time)) => {
log::trace!("write task is instructed to shutdown");
let timeout = if let Some(delay) = delay.take() {
delay
} else {
sleep(time)
};
this.st = IoWriteState::Shutdown(timeout, Shutdown::None);
self.poll(cx)
}
Poll::Ready(WriteStatus::Terminate) => {
log::trace!("write task is instructed to terminate");
let _ = Pin::new(&mut this.io.0).poll_close(cx);
this.state.close(None);
Poll::Ready(())
}
Poll::Pending => Poll::Pending,
}
}
IoWriteState::Shutdown(ref mut delay, ref mut st) => {
// close WRITE side and wait for disconnect on read side.
// use disconnect timeout, otherwise it could hang forever.
loop {
match st {
Shutdown::None => {
// flush write buffer
match flush_io(&mut this.io.0, &this.state, cx) {
Poll::Ready(true) => {
if let Err(_) =
this.io.0.shutdown(std::net::Shutdown::Write)
{
this.state.close(None);
return Poll::Ready(());
}
*st = Shutdown::Stopping(0);
continue;
}
Poll::Ready(false) => {
log::trace!(
"write task is closed with err during flush"
);
this.state.close(None);
return Poll::Ready(());
}
_ => (),
}
}
Shutdown::Stopping(ref mut count) => {
// read until 0 or err
let mut buf = [0u8; 512];
let io = &mut this.io;
loop {
match Pin::new(&mut io.0).poll_read(cx, &mut buf) {
Poll::Ready(Err(e)) => {
log::trace!("write task is stopped");
this.state.close(Some(e));
return Poll::Ready(());
}
Poll::Ready(Ok(0)) => {
log::trace!(
"async-std unix socket is disconnected"
);
this.state.close(None);
return Poll::Ready(());
}
Poll::Ready(Ok(n)) => {
*count += n as u16;
if *count > 4096 {
log::trace!(
"write task is stopped, too much input"
);
this.state.close(None);
return Poll::Ready(());
}
}
Poll::Pending => break,
}
}
}
}
// disconnect timeout
if delay.poll_elapsed(cx).is_pending() {
return Poll::Pending;
}
log::trace!("write task is stopped after delay");
this.state.close(None);
let _ = Pin::new(&mut this.io.0).poll_close(cx);
return Poll::Ready(());
}
}
}
}
}
}

View file

@ -1,4 +1,5 @@
//! A runtime implementation that runs everything on the current thread. //! A runtime implementation that runs everything on the current thread.
#![allow(clippy::return_self_not_must_use)]
use std::{future::Future, pin::Pin}; use std::{future::Future, pin::Pin};
mod arbiter; mod arbiter;
@ -14,6 +15,11 @@ mod tokio;
#[cfg(feature = "tokio")] #[cfg(feature = "tokio")]
pub use self::tokio::*; pub use self::tokio::*;
#[cfg(feature = "async-std")]
mod asyncstd;
#[cfg(all(not(feature = "tokio"), feature = "async-std"))]
pub use self::asyncstd::*;
pub trait Runtime { pub trait Runtime {
/// Spawn a future onto the single-threaded runtime. /// Spawn a future onto the single-threaded runtime.
fn spawn(&self, future: Pin<Box<dyn Future<Output = ()>>>); fn spawn(&self, future: Pin<Box<dyn Future<Output = ()>>>);

View file

@ -17,55 +17,37 @@ pub fn create_runtime() -> Box<dyn Runtime> {
} }
/// Opens a TCP connection to a remote host. /// Opens a TCP connection to a remote host.
pub fn tcp_connect( pub async fn tcp_connect(addr: SocketAddr) -> Result<Io, io::Error> {
addr: SocketAddr, let sock = tok_io::net::TcpStream::connect(addr).await?;
) -> Pin<Box<dyn Future<Output = Result<Io, io::Error>>>> { sock.set_nodelay(true)?;
Box::pin(async move { Ok(Io::new(sock))
let sock = tok_io::net::TcpStream::connect(addr).await?;
sock.set_nodelay(true)?;
Ok(Io::new(sock))
})
} }
/// Opens a TCP connection to a remote host and use specified memory pool. /// Opens a TCP connection to a remote host and use specified memory pool.
pub fn tcp_connect_in( pub async fn tcp_connect_in(addr: SocketAddr, pool: PoolRef) -> Result<Io, io::Error> {
addr: SocketAddr, let sock = tok_io::net::TcpStream::connect(addr).await?;
pool: PoolRef, sock.set_nodelay(true)?;
) -> Pin<Box<dyn Future<Output = Result<Io, io::Error>>>> { Ok(Io::with_memory_pool(sock, pool))
Box::pin(async move {
let sock = tok_io::net::TcpStream::connect(addr).await?;
sock.set_nodelay(true)?;
Ok(Io::with_memory_pool(sock, pool))
})
} }
#[cfg(unix)] #[cfg(unix)]
/// Opens a unix stream connection. /// Opens a unix stream connection.
pub fn unix_connect<'a, P>( pub async fn unix_connect<'a, P>(addr: P) -> Result<Io, io::Error>
addr: P,
) -> Pin<Box<dyn Future<Output = Result<Io, io::Error>> + 'a>>
where where
P: AsRef<Path> + 'a, P: AsRef<Path> + 'a,
{ {
Box::pin(async move { let sock = tok_io::net::UnixStream::connect(addr).await?;
let sock = tok_io::net::UnixStream::connect(addr).await?; Ok(Io::new(sock))
Ok(Io::new(sock))
})
} }
#[cfg(unix)] #[cfg(unix)]
/// Opens a unix stream connection and specified memory pool. /// Opens a unix stream connection and specified memory pool.
pub fn unix_connect_in<'a, P>( pub async fn unix_connect_in<'a, P>(addr: P, pool: PoolRef) -> Result<Io, io::Error>
addr: P,
pool: PoolRef,
) -> Pin<Box<dyn Future<Output = Result<Io, io::Error>> + 'a>>
where where
P: AsRef<Path> + 'a, P: AsRef<Path> + 'a,
{ {
Box::pin(async move { let sock = tok_io::net::UnixStream::connect(addr).await?;
let sock = tok_io::net::UnixStream::connect(addr).await?; Ok(Io::with_memory_pool(sock, pool))
Ok(Io::with_memory_pool(sock, pool))
})
} }
/// Convert std TcpStream to tokio's TcpStream /// Convert std TcpStream to tokio's TcpStream

View file

@ -1,4 +1,5 @@
//! An implementations of SSL streams for ntex ecosystem //! An implementations of SSL streams for ntex ecosystem
#![allow(clippy::return_self_not_must_use)]
use std::sync::atomic::{AtomicUsize, Ordering}; use std::sync::atomic::{AtomicUsize, Ordering};
pub mod types; pub mod types;

View file

@ -1,5 +1,9 @@
# Changes # Changes
## [0.5.0-b.6] - 2021-12-29
* Add `async-std` support
## [0.5.0-b.5] - 2021-12-28 ## [0.5.0-b.5] - 2021-12-28
* http: proper send payload, if request payload is not consumed * http: proper send payload, if request payload is not consumed

View file

@ -1,6 +1,6 @@
[package] [package]
name = "ntex" name = "ntex"
version = "0.5.0-b.5" version = "0.5.0-b.6"
authors = ["ntex contributors <team@ntex.rs>"] authors = ["ntex contributors <team@ntex.rs>"]
description = "Framework for composable network services" description = "Framework for composable network services"
readme = "README.md" readme = "README.md"
@ -21,7 +21,7 @@ name = "ntex"
path = "src/lib.rs" path = "src/lib.rs"
[features] [features]
default = [] default = ["tokio"]
# openssl # openssl
openssl = ["tls-openssl", "ntex-tls/openssl"] openssl = ["tls-openssl", "ntex-tls/openssl"]
@ -38,6 +38,12 @@ cookie = ["coo-kie", "coo-kie/percent-encode"]
# url support # url support
url = ["url-pkg"] url = ["url-pkg"]
# tokio runtime
tokio = ["ntex-rt/tokio"]
# async-std runtime
async-std = ["ntex-rt/async-std"]
[dependencies] [dependencies]
ntex-codec = "0.6.0" ntex-codec = "0.6.0"
ntex-router = "0.5.1" ntex-router = "0.5.1"
@ -46,8 +52,8 @@ ntex-macros = "0.1.3"
ntex-util = "0.1.5" ntex-util = "0.1.5"
ntex-bytes = "0.1.8" ntex-bytes = "0.1.8"
ntex-tls = "0.1.0-b.6" ntex-tls = "0.1.0-b.6"
ntex-io = "0.1.0-b.8" ntex-rt = "0.4.0-b.3"
ntex-rt = { version = "0.4.0-b.2", default-features = false, features = ["tokio"] } ntex-io = { version = "0.1.0-b.9", features = ["tokio-traits"] }
base64 = "0.13" base64 = "0.13"
bitflags = "1.3" bitflags = "1.3"

View file

@ -175,7 +175,7 @@ impl<T: Address> TcpConnectorResponse<T> {
Either::Left(addr) => TcpConnectorResponse { Either::Left(addr) => TcpConnectorResponse {
req: Some(req), req: Some(req),
addrs: None, addrs: None,
stream: Some(tcp_connect_in(addr, pool)), stream: Some(Box::pin(tcp_connect_in(addr, pool))),
pool, pool,
port, port,
}, },
@ -230,7 +230,7 @@ impl<T: Address> Future for TcpConnectorResponse<T> {
// try to connect // try to connect
let addr = this.addrs.as_mut().unwrap().pop_front().unwrap(); let addr = this.addrs.as_mut().unwrap().pop_front().unwrap();
this.stream = Some(tcp_connect_in(addr, this.pool)); this.stream = Some(Box::pin(tcp_connect_in(addr, this.pool)));
} }
} }
} }

View file

@ -18,7 +18,8 @@
clippy::borrow_interior_mutable_const, clippy::borrow_interior_mutable_const,
clippy::needless_doctest_main, clippy::needless_doctest_main,
clippy::too_many_arguments, clippy::too_many_arguments,
clippy::new_without_default clippy::new_without_default,
clippy::return_self_not_must_use
)] )]
#[macro_use] #[macro_use]

View file

@ -56,7 +56,6 @@ async fn test_openssl_string() {
#[cfg(feature = "openssl")] #[cfg(feature = "openssl")]
#[ntex::test] #[ntex::test]
async fn test_openssl_read_before_error() { async fn test_openssl_read_before_error() {
env_logger::init();
use ntex::server::openssl; use ntex::server::openssl;
use tls_openssl::ssl::{SslConnector, SslMethod, SslVerifyMode}; use tls_openssl::ssl::{SslConnector, SslMethod, SslVerifyMode};

View file

@ -263,6 +263,7 @@ async fn test_connection_force_close() {
assert!(response.status().is_success()); assert!(response.status().is_success());
// req 2 // req 2
let client = Client::build().timeout(Seconds(10)).finish();
let req = client.post(srv.url("/")).force_close(); let req = client.post(srv.url("/")).force_close();
let response = req.send().await.unwrap(); let response = req.send().await.unwrap();
assert!(response.status().is_success()); assert!(response.status().is_success());

View file

@ -266,6 +266,7 @@ async fn test_listen_uds() {
let sys = ntex::rt::System::new("test"); let sys = ntex::rt::System::new("test");
let srv = sys.exec(|| { let srv = sys.exec(|| {
let _ = std::fs::remove_file("/tmp/uds-test2");
let lst = std::os::unix::net::UnixListener::bind("/tmp/uds-test2").unwrap(); let lst = std::os::unix::net::UnixListener::bind("/tmp/uds-test2").unwrap();
HttpServer::new(|| { HttpServer::new(|| {