forked from lx/netapp
WIP: associated stream #1
3 changed files with 124 additions and 71 deletions
|
@ -42,7 +42,7 @@ where
|
||||||
(self.clone(), None)
|
(self.clone(), None)
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn deserialize_msg(ser_self: Self::SerializableSelf, stream: AssociatedStream) -> Self {
|
async fn deserialize_msg(ser_self: Self::SerializableSelf, _stream: AssociatedStream) -> Self {
|
||||||
// TODO verify no stream
|
// TODO verify no stream
|
||||||
ser_self
|
ser_self
|
||||||
}
|
}
|
||||||
|
|
185
src/proto.rs
185
src/proto.rs
|
@ -7,7 +7,7 @@ use log::{trace, warn};
|
||||||
|
|
||||||
use futures::channel::mpsc::{unbounded, UnboundedReceiver, UnboundedSender};
|
use futures::channel::mpsc::{unbounded, UnboundedReceiver, UnboundedSender};
|
||||||
use futures::Stream;
|
use futures::Stream;
|
||||||
use futures::{AsyncReadExt, AsyncWriteExt, FutureExt, StreamExt};
|
use futures::{AsyncReadExt, AsyncWriteExt};
|
||||||
use kuska_handshake::async_std::BoxStreamWrite;
|
use kuska_handshake::async_std::BoxStreamWrite;
|
||||||
|
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc;
|
||||||
|
@ -53,7 +53,8 @@ pub const PRIO_SECONDARY: RequestPriority = 0x01;
|
||||||
|
|
||||||
pub(crate) type RequestID = u32;
|
pub(crate) type RequestID = u32;
|
||||||
type ChunkLength = u16;
|
type ChunkLength = u16;
|
||||||
const MAX_CHUNK_LENGTH: ChunkLength = 0x4000;
|
const MAX_CHUNK_LENGTH: ChunkLength = 0x3FF0;
|
||||||
|
const ERROR_MARKER: ChunkLength = 0x4000;
|
||||||
const CHUNK_HAS_CONTINUATION: ChunkLength = 0x8000;
|
const CHUNK_HAS_CONTINUATION: ChunkLength = 0x8000;
|
||||||
|
|
||||||
struct SendQueueItem {
|
struct SendQueueItem {
|
||||||
|
@ -99,8 +100,29 @@ impl From<Data> for DataReader {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
struct DataReaderItem {
|
||||||
|
/// a fixed size buffer containing some data, possibly padded with 0s
|
||||||
|
data: [u8; MAX_CHUNK_LENGTH as usize],
|
||||||
|
/// actuall lenght of data
|
||||||
|
len: usize,
|
||||||
|
/// whethere there may be more data comming from this stream. Can be used for some
|
||||||
|
/// optimization. It's an error to set it to false if there is more data, but it is correct
|
||||||
|
/// (albeit sub-optimal) to set it to true if there is nothing coming after
|
||||||
|
may_have_more: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl DataReaderItem {
|
||||||
|
fn empty_last() -> Self {
|
||||||
|
DataReaderItem {
|
||||||
|
data: [0; MAX_CHUNK_LENGTH as usize],
|
||||||
|
len: 0,
|
||||||
|
may_have_more: false,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl Stream for DataReader {
|
impl Stream for DataReader {
|
||||||
type Item = ([u8; MAX_CHUNK_LENGTH as usize], usize);
|
type Item = DataReaderItem;
|
||||||
|
|
||||||
fn poll_next(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> {
|
fn poll_next(self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll<Option<Self::Item>> {
|
||||||
match self.project() {
|
match self.project() {
|
||||||
|
@ -114,7 +136,11 @@ impl Stream for DataReader {
|
||||||
let mut body = [0; MAX_CHUNK_LENGTH as usize];
|
let mut body = [0; MAX_CHUNK_LENGTH as usize];
|
||||||
body[..len].copy_from_slice(&data[*pos..end]);
|
body[..len].copy_from_slice(&data[*pos..end]);
|
||||||
*pos = end;
|
*pos = end;
|
||||||
Poll::Ready(Some((body, len)))
|
Poll::Ready(Some(DataReaderItem {
|
||||||
|
data: body,
|
||||||
|
len,
|
||||||
|
may_have_more: end < data.len(),
|
||||||
|
}))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
DataReaderProj::Streaming {
|
DataReaderProj::Streaming {
|
||||||
|
@ -154,7 +180,11 @@ impl Stream for DataReader {
|
||||||
let len = buf.len();
|
let len = buf.len();
|
||||||
body[..len].copy_from_slice(buf);
|
body[..len].copy_from_slice(buf);
|
||||||
buf.clear();
|
buf.clear();
|
||||||
Poll::Ready(Some((body, len)))
|
Poll::Ready(Some(DataReaderItem {
|
||||||
|
data: body,
|
||||||
|
len,
|
||||||
|
may_have_more: !*eos,
|
||||||
|
}))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -181,6 +211,8 @@ impl SendQueue {
|
||||||
};
|
};
|
||||||
self.items[pos_prio].1.push_back(item);
|
self.items[pos_prio].1.push_back(item);
|
||||||
}
|
}
|
||||||
|
// used only in tests. They should probably be rewriten
|
||||||
|
#[allow(dead_code)]
|
||||||
fn pop(&mut self) -> Option<SendQueueItem> {
|
fn pop(&mut self) -> Option<SendQueueItem> {
|
||||||
match self.items.pop_front() {
|
match self.items.pop_front() {
|
||||||
None => None,
|
None => None,
|
||||||
|
@ -196,6 +228,54 @@ impl SendQueue {
|
||||||
fn is_empty(&self) -> bool {
|
fn is_empty(&self) -> bool {
|
||||||
self.items.iter().all(|(_k, v)| v.is_empty())
|
self.items.iter().all(|(_k, v)| v.is_empty())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// this is like an async fn, but hand implemented
|
||||||
|
fn next_ready(&mut self) -> SendQueuePollNextReady<'_> {
|
||||||
|
SendQueuePollNextReady { queue: self }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
struct SendQueuePollNextReady<'a> {
|
||||||
|
queue: &'a mut SendQueue,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'a> futures::Future for SendQueuePollNextReady<'a> {
|
||||||
|
type Output = (RequestID, DataReaderItem);
|
||||||
|
|
||||||
|
fn poll(mut self: Pin<&mut Self>, ctx: &mut Context<'_>) -> Poll<Self::Output> {
|
||||||
|
for i in 0..self.queue.items.len() {
|
||||||
|
let (_prio, items_at_prio) = &mut self.queue.items[i];
|
||||||
|
|
||||||
|
for _ in 0..items_at_prio.len() {
|
||||||
|
let mut item = items_at_prio.pop_front().unwrap();
|
||||||
|
|
||||||
|
match Pin::new(&mut item.data).poll_next(ctx) {
|
||||||
|
Poll::Pending => items_at_prio.push_back(item),
|
||||||
|
Poll::Ready(Some(data)) => {
|
||||||
|
let id = item.id;
|
||||||
|
if data.may_have_more {
|
||||||
|
self.queue.push(item);
|
||||||
|
} else {
|
||||||
|
if items_at_prio.is_empty() {
|
||||||
|
// this priority level is empty, remove it
|
||||||
|
self.queue.items.remove(i);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return Poll::Ready((id, data));
|
||||||
|
}
|
||||||
|
Poll::Ready(None) => {
|
||||||
|
if items_at_prio.is_empty() {
|
||||||
|
// this priority level is empty, remove it
|
||||||
|
self.queue.items.remove(i);
|
||||||
|
}
|
||||||
|
return Poll::Ready((item.id, DataReaderItem::empty_last()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// TODO what do we do if self.queue is empty? We won't get scheduled again.
|
||||||
|
Poll::Pending
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// The SendLoop trait, which is implemented both by the client and the server
|
/// The SendLoop trait, which is implemented both by the client and the server
|
||||||
|
@ -219,77 +299,42 @@ pub(crate) trait SendLoop: Sync {
|
||||||
let mut sending = SendQueue::new();
|
let mut sending = SendQueue::new();
|
||||||
let mut should_exit = false;
|
let mut should_exit = false;
|
||||||
while !should_exit || !sending.is_empty() {
|
while !should_exit || !sending.is_empty() {
|
||||||
if let Ok((id, prio, data)) = msg_recv.try_recv() {
|
let recv_fut = msg_recv.recv();
|
||||||
match &data {
|
futures::pin_mut!(recv_fut);
|
||||||
Data::Full(data) => {
|
let send_fut = sending.next_ready();
|
||||||
trace!("send_loop: got {}, {} bytes", id, data.len());
|
|
||||||
}
|
// recv_fut is cancellation-safe according to tokio doc,
|
||||||
Data::Streaming(_) => {
|
// send_fut is cancellation-safe as implemented above?
|
||||||
trace!("send_loop: got {}, unknown size", id);
|
use futures::future::Either;
|
||||||
}
|
match futures::future::select(recv_fut, send_fut).await {
|
||||||
|
Either::Left((sth, _send_fut)) => {
|
||||||
|
if let Some((id, prio, data)) = sth {
|
||||||
|
sending.push(SendQueueItem {
|
||||||
|
id,
|
||||||
|
prio,
|
||||||
|
data: data.into(),
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
should_exit = true;
|
||||||
|
};
|
||||||
}
|
}
|
||||||
sending.push(SendQueueItem {
|
Either::Right(((id, data), _recv_fut)) => {
|
||||||
id,
|
trace!("send_loop: sending bytes for {}", id);
|
||||||
prio,
|
|
||||||
data: data.into(),
|
|
||||||
});
|
|
||||||
} else if let Some(mut item) = sending.pop() {
|
|
||||||
trace!("send_loop: sending bytes for {}", item.id,);
|
|
||||||
|
|
||||||
let data = futures::select! {
|
let header_id = RequestID::to_be_bytes(id);
|
||||||
data = item.data.next().fuse() => data,
|
write.write_all(&header_id[..]).await?;
|
||||||
default => {
|
|
||||||
// nothing to send yet; re-schedule and find something else to do
|
|
||||||
sending.push(item);
|
|
||||||
continue;
|
|
||||||
|
|
||||||
// TODO if every SendQueueItem is waiting on data, use select_all to await
|
let body = &data.data[..data.len];
|
||||||
// something to do
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let header_id = RequestID::to_be_bytes(item.id);
|
let size_header = if data.may_have_more {
|
||||||
write.write_all(&header_id[..]).await?;
|
ChunkLength::to_be_bytes(data.len as u16 | CHUNK_HAS_CONTINUATION)
|
||||||
|
} else {
|
||||||
|
ChunkLength::to_be_bytes(data.len as u16)
|
||||||
|
};
|
||||||
|
|
||||||
let data = match data.as_ref() {
|
|
||||||
Some((data, len)) => &data[..*len],
|
|
||||||
None => &[],
|
|
||||||
};
|
|
||||||
|
|
||||||
if data.len() == MAX_CHUNK_LENGTH as usize {
|
|
||||||
let size_header =
|
|
||||||
ChunkLength::to_be_bytes(data.len() as u16 | CHUNK_HAS_CONTINUATION);
|
|
||||||
write.write_all(&size_header[..]).await?;
|
write.write_all(&size_header[..]).await?;
|
||||||
|
write.write_all(body).await?;
|
||||||
write.write_all(data).await?;
|
write.flush().await?;
|
||||||
|
|
||||||
sending.push(item);
|
|
||||||
} else {
|
|
||||||
let size_header = ChunkLength::to_be_bytes(data.len() as u16);
|
|
||||||
write.write_all(&size_header[..]).await?;
|
|
||||||
|
|
||||||
write.write_all(data).await?;
|
|
||||||
}
|
|
||||||
|
|
||||||
write.flush().await?;
|
|
||||||
} else {
|
|
||||||
let sth = msg_recv.recv().await;
|
|
||||||
if let Some((id, prio, data)) = sth {
|
|
||||||
match &data {
|
|
||||||
Data::Full(data) => {
|
|
||||||
trace!("send_loop: got {}, {} bytes", id, data.len());
|
|
||||||
}
|
|
||||||
Data::Streaming(_) => {
|
|
||||||
trace!("send_loop: got {}, unknown size", id);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sending.push(SendQueueItem {
|
|
||||||
id,
|
|
||||||
prio,
|
|
||||||
data: data.into(),
|
|
||||||
});
|
|
||||||
} else {
|
|
||||||
should_exit = true;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -19,6 +19,14 @@ pub type NodeKey = sodiumoxide::crypto::sign::ed25519::SecretKey;
|
||||||
/// A network key
|
/// A network key
|
||||||
pub type NetworkKey = sodiumoxide::crypto::auth::Key;
|
pub type NetworkKey = sodiumoxide::crypto::auth::Key;
|
||||||
|
|
||||||
|
/// A stream of associated data.
|
||||||
|
///
|
||||||
|
/// The Stream can continue after receiving an error.
|
||||||
|
/// When sent through Netapp, the Vec may be split in smaller chunk in such a way
|
||||||
|
/// consecutive Vec may get merged, but Vec and error code may not be reordered
|
||||||
|
///
|
||||||
|
/// The error code have no predefined meaning, it's up to you application to define their
|
||||||
|
/// semantic.
|
||||||
pub type AssociatedStream = Pin<Box<dyn Stream<Item = Vec<u8>> + Send>>;
|
pub type AssociatedStream = Pin<Box<dyn Stream<Item = Vec<u8>> + Send>>;
|
||||||
|
|
||||||
/// Utility function: encodes any serializable value in MessagePack binary format
|
/// Utility function: encodes any serializable value in MessagePack binary format
|
||||||
|
|
Loading…
Reference in a new issue