Merge pull request #171 from greatest-ape/ws-fix-backpressure-2

aquatic_ws: remove ineffective backpressure implementation, improve load tester, remove glob exports
This commit is contained in:
Joakim Frostegård 2024-01-08 20:37:52 +01:00 committed by GitHub
commit d8bd964a57
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
11 changed files with 214 additions and 246 deletions

View file

@ -3,7 +3,7 @@ use std::{net::IpAddr, sync::Arc};
use aquatic_common::access_list::AccessListArcSwap; use aquatic_common::access_list::AccessListArcSwap;
pub use aquatic_common::ValidUntil; pub use aquatic_common::ValidUntil;
use aquatic_ws_protocol::{InfoHash, PeerId}; use aquatic_ws_protocol::common::{InfoHash, PeerId};
#[derive(Copy, Clone, Debug)] #[derive(Copy, Clone, Debug)]
pub enum IpVersion { pub enum IpVersion {
@ -67,11 +67,10 @@ impl Into<OutMessageMeta> for InMessageMeta {
} }
} }
#[derive(Clone, Copy, Debug)] #[derive(Clone, Debug)]
pub enum SwarmControlMessage { pub enum SwarmControlMessage {
ConnectionClosed { ConnectionClosed {
info_hash: InfoHash,
peer_id: PeerId,
ip_version: IpVersion, ip_version: IpVersion,
announced_info_hashes: Vec<(InfoHash, PeerId)>,
}, },
} }

View file

@ -9,7 +9,13 @@ use anyhow::Context;
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache}; use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache};
use aquatic_common::rustls_config::RustlsConfig; use aquatic_common::rustls_config::RustlsConfig;
use aquatic_common::ServerStartInstant; use aquatic_common::ServerStartInstant;
use aquatic_ws_protocol::*; use aquatic_ws_protocol::common::{InfoHash, PeerId, ScrapeAction};
use aquatic_ws_protocol::incoming::{
AnnounceEvent, AnnounceRequest, InMessage, ScrapeRequest, ScrapeRequestInfoHashes,
};
use aquatic_ws_protocol::outgoing::{
ErrorResponse, ErrorResponseAction, OutMessage, ScrapeResponse, ScrapeStatistics,
};
use arc_swap::ArcSwap; use arc_swap::ArcSwap;
use async_tungstenite::WebSocketStream; use async_tungstenite::WebSocketStream;
use futures::stream::{SplitSink, SplitStream}; use futures::stream::{SplitSink, SplitStream};
@ -17,7 +23,7 @@ use futures::{AsyncWriteExt, StreamExt};
use futures_lite::future::race; use futures_lite::future::race;
use futures_rustls::TlsAcceptor; use futures_rustls::TlsAcceptor;
use glommio::channels::channel_mesh::Senders; use glommio::channels::channel_mesh::Senders;
use glommio::channels::local_channel::{new_bounded, LocalReceiver, LocalSender}; use glommio::channels::local_channel::{LocalReceiver, LocalSender};
use glommio::net::TcpStream; use glommio::net::TcpStream;
use glommio::timer::timeout; use glommio::timer::timeout;
use glommio::{enclose, prelude::*}; use glommio::{enclose, prelude::*};
@ -35,13 +41,6 @@ use crate::workers::socket::calculate_in_message_consumer_index;
#[cfg(feature = "metrics")] #[cfg(feature = "metrics")]
use crate::workers::socket::{ip_version_to_metrics_str, WORKER_INDEX}; use crate::workers::socket::{ip_version_to_metrics_str, WORKER_INDEX};
/// Length of ConnectionReader backpressure channel
///
/// ConnectionReader awaits a message in a channel before proceeding with
/// reading a request. For each response sent, a message is sent to the
/// channel, up to a maximum of this constant.
const READ_PASS_CHANNEL_LEN: usize = 4;
pub struct ConnectionRunner { pub struct ConnectionRunner {
pub config: Rc<Config>, pub config: Rc<Config>,
pub access_list: Arc<AccessListArcSwap>, pub access_list: Arc<AccessListArcSwap>,
@ -168,17 +167,6 @@ impl ConnectionRunner {
let pending_scrape_slab = Rc::new(RefCell::new(Slab::new())); let pending_scrape_slab = Rc::new(RefCell::new(Slab::new()));
let access_list_cache = create_access_list_cache(&self.access_list); let access_list_cache = create_access_list_cache(&self.access_list);
let (read_pass_sender, read_pass_receiver) = new_bounded(READ_PASS_CHANNEL_LEN);
for _ in 0..READ_PASS_CHANNEL_LEN {
if let Err(err) = read_pass_sender.try_send(()) {
panic!(
"couldn't add initial entries to read pass channel: {:#}",
err
)
};
}
let config = self.config.clone(); let config = self.config.clone();
let reader_future = enclose!((pending_scrape_slab, clean_up_data) async move { let reader_future = enclose!((pending_scrape_slab, clean_up_data) async move {
@ -187,7 +175,6 @@ impl ConnectionRunner {
access_list_cache, access_list_cache,
in_message_senders: self.in_message_senders, in_message_senders: self.in_message_senders,
out_message_sender: self.out_message_sender, out_message_sender: self.out_message_sender,
read_pass_receiver,
pending_scrape_slab, pending_scrape_slab,
out_message_consumer_id: self.out_message_consumer_id, out_message_consumer_id: self.out_message_consumer_id,
ws_in, ws_in,
@ -217,7 +204,6 @@ impl ConnectionRunner {
let mut writer = ConnectionWriter { let mut writer = ConnectionWriter {
config, config,
out_message_receiver: self.out_message_receiver, out_message_receiver: self.out_message_receiver,
read_pass_sender,
connection_valid_until: self.connection_valid_until, connection_valid_until: self.connection_valid_until,
ws_out, ws_out,
pending_scrape_slab, pending_scrape_slab,
@ -238,7 +224,6 @@ struct ConnectionReader<S> {
access_list_cache: AccessListCache, access_list_cache: AccessListCache,
in_message_senders: Rc<Senders<(InMessageMeta, InMessage)>>, in_message_senders: Rc<Senders<(InMessageMeta, InMessage)>>,
out_message_sender: Rc<LocalSender<(OutMessageMeta, OutMessage)>>, out_message_sender: Rc<LocalSender<(OutMessageMeta, OutMessage)>>,
read_pass_receiver: LocalReceiver<()>,
pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>, pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>,
out_message_consumer_id: ConsumerId, out_message_consumer_id: ConsumerId,
ws_in: SplitStream<WebSocketStream<S>>, ws_in: SplitStream<WebSocketStream<S>>,
@ -254,11 +239,6 @@ struct ConnectionReader<S> {
impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> { impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
async fn run_in_message_loop(&mut self) -> anyhow::Result<()> { async fn run_in_message_loop(&mut self) -> anyhow::Result<()> {
loop { loop {
self.read_pass_receiver
.recv()
.await
.ok_or_else(|| anyhow::anyhow!("read pass channel closed"))?;
let message = self let message = self
.ws_in .ws_in
.next() .next()
@ -496,7 +476,6 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
struct ConnectionWriter<S> { struct ConnectionWriter<S> {
config: Rc<Config>, config: Rc<Config>,
out_message_receiver: LocalReceiver<(OutMessageMeta, OutMessage)>, out_message_receiver: LocalReceiver<(OutMessageMeta, OutMessage)>,
read_pass_sender: LocalSender<()>,
connection_valid_until: Rc<RefCell<ValidUntil>>, connection_valid_until: Rc<RefCell<ValidUntil>>,
ws_out: SplitSink<WebSocketStream<S>, tungstenite::Message>, ws_out: SplitSink<WebSocketStream<S>, tungstenite::Message>,
pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>, pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>,
@ -549,10 +528,6 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionWriter<S> {
} }
}; };
if let Err(GlommioError::Closed(_)) = self.read_pass_sender.try_send(()) {
return Err(anyhow::anyhow!("read pass channel closed"));
}
yield_if_needed().await; yield_if_needed().await;
} }
} }
@ -630,19 +605,23 @@ impl ConnectionCleanupData {
config: &Config, config: &Config,
control_message_senders: Rc<Senders<SwarmControlMessage>>, control_message_senders: Rc<Senders<SwarmControlMessage>>,
) { ) {
// Use RefCell::take to avoid issues with Rc borrow across await let mut announced_info_hashes = HashMap::new();
let announced_info_hashes = self.announced_info_hashes.take();
// Tell swarm workers to remove peer
for (info_hash, peer_id) in announced_info_hashes.into_iter() {
let message = SwarmControlMessage::ConnectionClosed {
info_hash,
peer_id,
ip_version: self.ip_version,
};
for (info_hash, peer_id) in self.announced_info_hashes.take().into_iter() {
let consumer_index = calculate_in_message_consumer_index(&config, info_hash); let consumer_index = calculate_in_message_consumer_index(&config, info_hash);
announced_info_hashes
.entry(consumer_index)
.or_insert(Vec::new())
.push((info_hash, peer_id));
}
for (consumer_index, announced_info_hashes) in announced_info_hashes.into_iter() {
let message = SwarmControlMessage::ConnectionClosed {
ip_version: self.ip_version,
announced_info_hashes,
};
control_message_senders control_message_senders
.send_to(consumer_index, message) .send_to(consumer_index, message)
.await .await

View file

@ -8,7 +8,9 @@ use anyhow::Context;
use aquatic_common::privileges::PrivilegeDropper; use aquatic_common::privileges::PrivilegeDropper;
use aquatic_common::rustls_config::RustlsConfig; use aquatic_common::rustls_config::RustlsConfig;
use aquatic_common::{PanicSentinel, ServerStartInstant}; use aquatic_common::{PanicSentinel, ServerStartInstant};
use aquatic_ws_protocol::*; use aquatic_ws_protocol::common::InfoHash;
use aquatic_ws_protocol::incoming::InMessage;
use aquatic_ws_protocol::outgoing::OutMessage;
use arc_swap::ArcSwap; use arc_swap::ArcSwap;
use futures::StreamExt; use futures::StreamExt;
use glommio::channels::channel_mesh::{MeshBuilder, Partial, Role}; use glommio::channels::channel_mesh::{MeshBuilder, Partial, Role};

View file

@ -4,6 +4,8 @@ use std::cell::RefCell;
use std::rc::Rc; use std::rc::Rc;
use std::time::Duration; use std::time::Duration;
use aquatic_ws_protocol::incoming::InMessage;
use aquatic_ws_protocol::outgoing::OutMessage;
use futures::StreamExt; use futures::StreamExt;
use glommio::channels::channel_mesh::{MeshBuilder, Partial, Role, Senders}; use glommio::channels::channel_mesh::{MeshBuilder, Partial, Role, Senders};
use glommio::enclose; use glommio::enclose;
@ -12,7 +14,6 @@ use glommio::timer::TimerActionRepeat;
use rand::{rngs::SmallRng, SeedableRng}; use rand::{rngs::SmallRng, SeedableRng};
use aquatic_common::{PanicSentinel, ServerStartInstant}; use aquatic_common::{PanicSentinel, ServerStartInstant};
use aquatic_ws_protocol::*;
use crate::common::*; use crate::common::*;
use crate::config::Config; use crate::config::Config;
@ -102,13 +103,14 @@ where
while let Some(message) = stream.next().await { while let Some(message) = stream.next().await {
match message { match message {
SwarmControlMessage::ConnectionClosed { SwarmControlMessage::ConnectionClosed {
info_hash,
peer_id,
ip_version, ip_version,
announced_info_hashes,
} => { } => {
torrents let mut torrents = torrents.borrow_mut();
.borrow_mut()
.handle_connection_closed(info_hash, peer_id, ip_version); for (info_hash, peer_id) in announced_info_hashes {
torrents.handle_connection_closed(info_hash, peer_id, ip_version);
}
} }
} }
} }
@ -124,25 +126,8 @@ async fn handle_request_stream<S>(
S: futures_lite::Stream<Item = (InMessageMeta, InMessage)> + ::std::marker::Unpin, S: futures_lite::Stream<Item = (InMessageMeta, InMessage)> + ::std::marker::Unpin,
{ {
let rng = Rc::new(RefCell::new(SmallRng::from_entropy())); let rng = Rc::new(RefCell::new(SmallRng::from_entropy()));
let max_peer_age = config.cleaning.max_peer_age;
let peer_valid_until = Rc::new(RefCell::new(ValidUntil::new(
server_start_instant,
max_peer_age,
)));
// Periodically update peer_valid_until
TimerActionRepeat::repeat(enclose!((peer_valid_until) move || {
enclose!((peer_valid_until) move || async move {
*peer_valid_until.borrow_mut() = ValidUntil::new(server_start_instant, max_peer_age);
Some(Duration::from_secs(1))
})()
}));
let config = &config; let config = &config;
let torrents = &torrents; let torrents = &torrents;
let peer_valid_until = &peer_valid_until;
let rng = &rng; let rng = &rng;
let out_message_senders = &out_message_senders; let out_message_senders = &out_message_senders;
@ -159,7 +144,6 @@ async fn handle_request_stream<S>(
&mut rng.borrow_mut(), &mut rng.borrow_mut(),
&mut out_messages, &mut out_messages,
server_start_instant, server_start_instant,
peer_valid_until.borrow().to_owned(),
meta, meta,
request, request,
) )

View file

@ -1,6 +1,11 @@
use std::sync::Arc; use std::sync::Arc;
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache}; use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache};
use aquatic_ws_protocol::incoming::{AnnounceEvent, AnnounceRequest, ScrapeRequest};
use aquatic_ws_protocol::outgoing::{
AnnounceResponse, AnswerOutMessage, ErrorResponse, ErrorResponseAction, OfferOutMessage,
OutMessage, ScrapeResponse, ScrapeStatistics,
};
use hashbrown::HashMap; use hashbrown::HashMap;
use metrics::Gauge; use metrics::Gauge;
use rand::rngs::SmallRng; use rand::rngs::SmallRng;
@ -8,7 +13,7 @@ use rand::rngs::SmallRng;
use aquatic_common::{ use aquatic_common::{
extract_response_peers, IndexMap, SecondsSinceServerStart, ServerStartInstant, extract_response_peers, IndexMap, SecondsSinceServerStart, ServerStartInstant,
}; };
use aquatic_ws_protocol::*; use aquatic_ws_protocol::common::*;
use crate::common::*; use crate::common::*;
use crate::config::Config; use crate::config::Config;
@ -62,7 +67,6 @@ impl TorrentMaps {
rng: &mut SmallRng, rng: &mut SmallRng,
out_messages: &mut Vec<(OutMessageMeta, OutMessage)>, out_messages: &mut Vec<(OutMessageMeta, OutMessage)>,
server_start_instant: ServerStartInstant, server_start_instant: ServerStartInstant,
valid_until: ValidUntil,
request_sender_meta: InMessageMeta, request_sender_meta: InMessageMeta,
request: AnnounceRequest, request: AnnounceRequest,
) { ) {
@ -72,6 +76,8 @@ impl TorrentMaps {
self.ipv6.entry(request.info_hash).or_default() self.ipv6.entry(request.info_hash).or_default()
}; };
let valid_until = ValidUntil::new(server_start_instant, config.cleaning.max_peer_age);
// If there is already a peer with this peer_id, check that connection id // If there is already a peer with this peer_id, check that connection id
// is same as that of request sender. Otherwise, ignore request. Since // is same as that of request sender. Otherwise, ignore request. Since
// peers have access to each others peer_id's, they could send requests // peers have access to each others peer_id's, they could send requests

View file

@ -1,9 +1,8 @@
use std::sync::{atomic::AtomicUsize, Arc}; use std::sync::{atomic::AtomicUsize, Arc};
use aquatic_ws_protocol::common::InfoHash;
use rand_distr::Gamma; use rand_distr::Gamma;
pub use aquatic_ws_protocol::*;
#[derive(Default)] #[derive(Default)]
pub struct Statistics { pub struct Statistics {
pub requests: AtomicUsize, pub requests: AtomicUsize,
@ -18,7 +17,7 @@ pub struct Statistics {
#[derive(Clone)] #[derive(Clone)]
pub struct LoadTestState { pub struct LoadTestState {
pub info_hashes: Arc<Vec<InfoHash>>, pub info_hashes: Arc<[InfoHash]>,
pub statistics: Arc<Statistics>, pub statistics: Arc<Statistics>,
pub gamma: Arc<Gamma<f64>>, pub gamma: Arc<Gamma<f64>>,
} }

View file

@ -4,6 +4,7 @@ use std::time::{Duration, Instant};
use aquatic_common::cpu_pinning::glommio::{get_worker_placement, set_affinity_for_util_worker}; use aquatic_common::cpu_pinning::glommio::{get_worker_placement, set_affinity_for_util_worker};
use aquatic_common::cpu_pinning::WorkerIndex; use aquatic_common::cpu_pinning::WorkerIndex;
use aquatic_ws_protocol::common::InfoHash;
use glommio::LocalExecutorBuilder; use glommio::LocalExecutorBuilder;
use rand::prelude::*; use rand::prelude::*;
use rand_distr::Gamma; use rand_distr::Gamma;
@ -36,10 +37,10 @@ fn run(config: Config) -> ::anyhow::Result<()> {
println!("Starting client with config: {:#?}", config); println!("Starting client with config: {:#?}", config);
let mut info_hashes = Vec::with_capacity(config.torrents.number_of_torrents);
let mut rng = SmallRng::from_entropy(); let mut rng = SmallRng::from_entropy();
let mut info_hashes = Vec::with_capacity(config.torrents.number_of_torrents);
for _ in 0..config.torrents.number_of_torrents { for _ in 0..config.torrents.number_of_torrents {
info_hashes.push(InfoHash(rng.gen())); info_hashes.push(InfoHash(rng.gen()));
} }
@ -51,7 +52,7 @@ fn run(config: Config) -> ::anyhow::Result<()> {
.unwrap(); .unwrap();
let state = LoadTestState { let state = LoadTestState {
info_hashes: Arc::new(info_hashes), info_hashes: Arc::from(info_hashes.into_boxed_slice()),
statistics: Arc::new(Statistics::default()), statistics: Arc::new(Statistics::default()),
gamma: Arc::new(gamma), gamma: Arc::new(gamma),
}; };

View file

@ -6,15 +6,30 @@ use std::{
time::Duration, time::Duration,
}; };
use aquatic_ws_protocol::{InMessage, OfferId, OutMessage, PeerId, RtcAnswer, RtcAnswerType}; use aquatic_ws_protocol::incoming::{
AnnounceEvent, AnnounceRequest, AnnounceRequestOffer, InMessage, ScrapeRequestInfoHashes,
};
use aquatic_ws_protocol::outgoing::OutMessage;
use aquatic_ws_protocol::{
common::{
AnnounceAction, InfoHash, OfferId, PeerId, RtcAnswer, RtcAnswerType, RtcOffer,
RtcOfferType, ScrapeAction,
},
incoming::ScrapeRequest,
};
use async_tungstenite::{client_async, WebSocketStream}; use async_tungstenite::{client_async, WebSocketStream};
use futures::{SinkExt, StreamExt}; use futures::{SinkExt, StreamExt};
use futures_rustls::{client::TlsStream, TlsConnector}; use futures_rustls::{client::TlsStream, TlsConnector};
use glommio::net::TcpStream; use glommio::net::TcpStream;
use glommio::{prelude::*, timer::TimerActionRepeat}; use glommio::{prelude::*, timer::TimerActionRepeat};
use rand::{prelude::SmallRng, Rng, SeedableRng}; use rand::{prelude::SmallRng, Rng, SeedableRng};
use rand_distr::{Distribution, WeightedIndex};
use crate::{common::LoadTestState, config::Config, utils::create_random_request}; use crate::{
common::{LoadTestState, RequestType},
config::Config,
utils::select_info_hash_index,
};
pub async fn run_socket_thread( pub async fn run_socket_thread(
config: Config, config: Config,
@ -22,7 +37,10 @@ pub async fn run_socket_thread(
load_test_state: LoadTestState, load_test_state: LoadTestState,
) -> anyhow::Result<()> { ) -> anyhow::Result<()> {
let config = Rc::new(config); let config = Rc::new(config);
let rng = Rc::new(RefCell::new(SmallRng::from_entropy()));
let num_active_connections = Rc::new(RefCell::new(0usize)); let num_active_connections = Rc::new(RefCell::new(0usize));
let connection_creation_interval =
Duration::from_millis(config.connection_creation_interval_ms);
TimerActionRepeat::repeat(move || { TimerActionRepeat::repeat(move || {
periodically_open_connections( periodically_open_connections(
@ -30,10 +48,12 @@ pub async fn run_socket_thread(
tls_config.clone(), tls_config.clone(),
load_test_state.clone(), load_test_state.clone(),
num_active_connections.clone(), num_active_connections.clone(),
rng.clone(),
connection_creation_interval,
) )
}); })
.join()
futures::future::pending::<bool>().await; .await;
Ok(()) Ok(())
} }
@ -43,13 +63,19 @@ async fn periodically_open_connections(
tls_config: Arc<rustls::ClientConfig>, tls_config: Arc<rustls::ClientConfig>,
load_test_state: LoadTestState, load_test_state: LoadTestState,
num_active_connections: Rc<RefCell<usize>>, num_active_connections: Rc<RefCell<usize>>,
rng: Rc<RefCell<SmallRng>>,
connection_creation_interval: Duration,
) -> Option<Duration> { ) -> Option<Duration> {
let wait = Duration::from_millis(config.connection_creation_interval_ms);
if *num_active_connections.borrow() < config.num_connections_per_worker { if *num_active_connections.borrow() < config.num_connections_per_worker {
spawn_local(async move { spawn_local(async move {
if let Err(err) = if let Err(err) = Connection::run(
Connection::run(config, tls_config, load_test_state, num_active_connections).await config,
tls_config,
load_test_state,
num_active_connections,
rng,
)
.await
{ {
::log::info!("connection creation error: {:#}", err); ::log::info!("connection creation error: {:#}", err);
} }
@ -57,16 +83,15 @@ async fn periodically_open_connections(
.detach(); .detach();
} }
Some(wait) Some(connection_creation_interval)
} }
struct Connection { struct Connection {
config: Rc<Config>, config: Rc<Config>,
load_test_state: LoadTestState, load_test_state: LoadTestState,
rng: SmallRng, rng: Rc<RefCell<SmallRng>>,
can_send: bool,
peer_id: PeerId, peer_id: PeerId,
send_answer: Option<(PeerId, OfferId)>, can_send_answer: Option<(InfoHash, PeerId, OfferId)>,
stream: WebSocketStream<TlsStream<TcpStream>>, stream: WebSocketStream<TlsStream<TcpStream>>,
} }
@ -76,9 +101,9 @@ impl Connection {
tls_config: Arc<rustls::ClientConfig>, tls_config: Arc<rustls::ClientConfig>,
load_test_state: LoadTestState, load_test_state: LoadTestState,
num_active_connections: Rc<RefCell<usize>>, num_active_connections: Rc<RefCell<usize>>,
rng: Rc<RefCell<SmallRng>>,
) -> anyhow::Result<()> { ) -> anyhow::Result<()> {
let mut rng = SmallRng::from_entropy(); let peer_id = PeerId(rng.borrow_mut().gen());
let peer_id = PeerId(rng.gen());
let stream = TcpStream::connect(config.server_address) let stream = TcpStream::connect(config.server_address)
.await .await
.map_err(|err| anyhow::anyhow!("connect: {:?}", err))?; .map_err(|err| anyhow::anyhow!("connect: {:?}", err))?;
@ -99,9 +124,8 @@ impl Connection {
load_test_state, load_test_state,
rng, rng,
stream, stream,
can_send: true,
peer_id, peer_id,
send_answer: None, can_send_answer: None,
}; };
*num_active_connections.borrow_mut() += 1; *num_active_connections.borrow_mut() += 1;
@ -119,50 +143,103 @@ impl Connection {
async fn run_connection_loop(&mut self) -> anyhow::Result<()> { async fn run_connection_loop(&mut self) -> anyhow::Result<()> {
loop { loop {
if self.can_send { self.send_message().await?;
let request = create_random_request(
&self.config,
&self.load_test_state,
&mut self.rng,
self.peer_id,
self.send_answer.is_none(),
);
// If self.send_answer is set and request is announce request, make
// the request an offer answer
let request = if let InMessage::AnnounceRequest(mut r) = request {
if let Some((peer_id, offer_id)) = self.send_answer {
r.answer_to_peer_id = Some(peer_id);
r.answer_offer_id = Some(offer_id);
r.answer = Some(RtcAnswer {
t: RtcAnswerType::Answer,
sdp: "abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-".into()
});
r.event = None;
r.offers = None;
}
self.send_answer = None;
InMessage::AnnounceRequest(r)
} else {
request
};
self.stream.send(request.to_ws_message()).await?;
self.load_test_state
.statistics
.requests
.fetch_add(1, Ordering::Relaxed);
self.can_send = false;
}
self.read_message().await?; self.read_message().await?;
} }
} }
async fn send_message(&mut self) -> anyhow::Result<()> {
let mut rng = self.rng.borrow_mut();
let request = match random_request_type(&self.config, &mut *rng) {
RequestType::Announce => {
let (event, bytes_left) = {
if rng.gen_bool(self.config.torrents.peer_seeder_probability) {
(AnnounceEvent::Completed, 0)
} else {
(AnnounceEvent::Started, 50)
}
};
const SDP: &str = "abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg";
if let Some((info_hash, peer_id, offer_id)) = self.can_send_answer {
InMessage::AnnounceRequest(AnnounceRequest {
info_hash,
answer_to_peer_id: Some(peer_id),
answer_offer_id: Some(offer_id),
answer: Some(RtcAnswer {
t: RtcAnswerType::Answer,
sdp: SDP.into(),
}),
event: None,
offers: None,
action: AnnounceAction::Announce,
peer_id: self.peer_id,
bytes_left: Some(bytes_left),
numwant: Some(0),
})
} else {
let info_hash_index =
select_info_hash_index(&self.config, &self.load_test_state, &mut *rng);
let mut offers = Vec::with_capacity(self.config.torrents.offers_per_request);
for _ in 0..self.config.torrents.offers_per_request {
offers.push(AnnounceRequestOffer {
offer_id: OfferId(rng.gen()),
offer: RtcOffer {
t: RtcOfferType::Offer,
sdp: SDP.into(),
},
})
}
InMessage::AnnounceRequest(AnnounceRequest {
action: AnnounceAction::Announce,
info_hash: self.load_test_state.info_hashes[info_hash_index],
peer_id: self.peer_id,
bytes_left: Some(bytes_left),
event: Some(event),
numwant: Some(offers.len()),
offers: Some(offers),
answer: None,
answer_to_peer_id: None,
answer_offer_id: None,
})
}
}
RequestType::Scrape => {
let mut scrape_hashes = Vec::with_capacity(5);
for _ in 0..5 {
let info_hash_index =
select_info_hash_index(&self.config, &self.load_test_state, &mut *rng);
scrape_hashes.push(self.load_test_state.info_hashes[info_hash_index]);
}
InMessage::ScrapeRequest(ScrapeRequest {
action: ScrapeAction::Scrape,
info_hashes: Some(ScrapeRequestInfoHashes::Multiple(scrape_hashes)),
})
}
};
drop(rng);
self.can_send_answer = None;
self.stream.send(request.to_ws_message()).await?;
self.load_test_state
.statistics
.requests
.fetch_add(1, Ordering::Relaxed);
Ok(())
}
async fn read_message(&mut self) -> anyhow::Result<()> { async fn read_message(&mut self) -> anyhow::Result<()> {
let message = match self let message = match self
.stream .stream
@ -190,33 +267,25 @@ impl Connection {
.responses_offer .responses_offer
.fetch_add(1, Ordering::Relaxed); .fetch_add(1, Ordering::Relaxed);
self.send_answer = Some((offer.peer_id, offer.offer_id)); self.can_send_answer = Some((offer.info_hash, offer.peer_id, offer.offer_id));
self.can_send = true;
} }
Ok(OutMessage::AnswerOutMessage(_)) => { Ok(OutMessage::AnswerOutMessage(_)) => {
self.load_test_state self.load_test_state
.statistics .statistics
.responses_answer .responses_answer
.fetch_add(1, Ordering::Relaxed); .fetch_add(1, Ordering::Relaxed);
self.can_send = true;
} }
Ok(OutMessage::AnnounceResponse(_)) => { Ok(OutMessage::AnnounceResponse(_)) => {
self.load_test_state self.load_test_state
.statistics .statistics
.responses_announce .responses_announce
.fetch_add(1, Ordering::Relaxed); .fetch_add(1, Ordering::Relaxed);
self.can_send = true;
} }
Ok(OutMessage::ScrapeResponse(_)) => { Ok(OutMessage::ScrapeResponse(_)) => {
self.load_test_state self.load_test_state
.statistics .statistics
.responses_scrape .responses_scrape
.fetch_add(1, Ordering::Relaxed); .fetch_add(1, Ordering::Relaxed);
self.can_send = true;
} }
Ok(OutMessage::ErrorResponse(response)) => { Ok(OutMessage::ErrorResponse(response)) => {
self.load_test_state self.load_test_state
@ -225,8 +294,6 @@ impl Connection {
.fetch_add(1, Ordering::Relaxed); .fetch_add(1, Ordering::Relaxed);
::log::warn!("received error response: {:?}", response.failure_reason); ::log::warn!("received error response: {:?}", response.failure_reason);
self.can_send = true;
} }
Err(err) => { Err(err) => {
::log::error!("error deserializing message: {:#}", err); ::log::error!("error deserializing message: {:#}", err);
@ -236,3 +303,16 @@ impl Connection {
Ok(()) Ok(())
} }
} }
pub fn random_request_type(config: &Config, rng: &mut impl Rng) -> RequestType {
let weights = [
config.torrents.weight_announce as u32,
config.torrents.weight_scrape as u32,
];
let items = [RequestType::Announce, RequestType::Scrape];
let dist = WeightedIndex::new(&weights).expect("random request weighted index");
items[dist.sample(rng)]
}

View file

@ -1,104 +1,13 @@
use std::sync::Arc; use std::sync::Arc;
use rand::distributions::WeightedIndex;
use rand::prelude::*; use rand::prelude::*;
use rand_distr::Gamma; use rand_distr::Gamma;
use crate::common::*; use crate::common::*;
use crate::config::*; use crate::config::*;
pub fn create_random_request(
config: &Config,
state: &LoadTestState,
rng: &mut impl Rng,
peer_id: PeerId,
announce_gen_offers: bool,
) -> InMessage {
let weights = [
config.torrents.weight_announce as u32,
config.torrents.weight_scrape as u32,
];
let items = [RequestType::Announce, RequestType::Scrape];
let dist = WeightedIndex::new(&weights).expect("random request weighted index");
match items[dist.sample(rng)] {
RequestType::Announce => {
create_announce_request(config, state, rng, peer_id, announce_gen_offers)
}
RequestType::Scrape => create_scrape_request(config, state, rng),
}
}
#[inline] #[inline]
fn create_announce_request( pub fn select_info_hash_index(config: &Config, state: &LoadTestState, rng: &mut impl Rng) -> usize {
config: &Config,
state: &LoadTestState,
rng: &mut impl Rng,
peer_id: PeerId,
gen_offers: bool,
) -> InMessage {
let (event, bytes_left) = {
if rng.gen_bool(config.torrents.peer_seeder_probability) {
(AnnounceEvent::Completed, 0)
} else {
(AnnounceEvent::Started, 50)
}
};
let info_hash_index = select_info_hash_index(config, &state, rng);
let offers = if gen_offers {
let mut offers = Vec::with_capacity(config.torrents.offers_per_request);
for _ in 0..config.torrents.offers_per_request {
offers.push(AnnounceRequestOffer {
offer_id: OfferId(rng.gen()),
offer: RtcOffer {
t: RtcOfferType::Offer,
sdp: "abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-abcdefg-".into()
},
})
}
offers
} else {
Vec::new()
};
InMessage::AnnounceRequest(AnnounceRequest {
action: AnnounceAction::Announce,
info_hash: state.info_hashes[info_hash_index],
peer_id,
bytes_left: Some(bytes_left),
event: Some(event),
numwant: Some(offers.len()),
offers: Some(offers),
answer: None,
answer_to_peer_id: None,
answer_offer_id: None,
})
}
#[inline]
fn create_scrape_request(config: &Config, state: &LoadTestState, rng: &mut impl Rng) -> InMessage {
let mut scrape_hashes = Vec::with_capacity(5);
for _ in 0..5 {
let info_hash_index = select_info_hash_index(config, &state, rng);
scrape_hashes.push(state.info_hashes[info_hash_index]);
}
InMessage::ScrapeRequest(ScrapeRequest {
action: ScrapeAction::Scrape,
info_hashes: Some(ScrapeRequestInfoHashes::Multiple(scrape_hashes)),
})
}
#[inline]
fn select_info_hash_index(config: &Config, state: &LoadTestState, rng: &mut impl Rng) -> usize {
gamma_usize(rng, &state.gamma, config.torrents.number_of_torrents - 1) gamma_usize(rng, &state.gamma, config.torrents.number_of_torrents - 1)
} }

View file

@ -1,7 +1,10 @@
use criterion::{black_box, criterion_group, criterion_main, Criterion}; use criterion::{black_box, criterion_group, criterion_main, Criterion};
use std::time::Duration; use std::time::Duration;
use aquatic_ws_protocol::*; use aquatic_ws_protocol::{
common::*,
incoming::{AnnounceEvent, AnnounceRequest, AnnounceRequestOffer, InMessage},
};
pub fn bench(c: &mut Criterion) { pub fn bench(c: &mut Criterion) {
let info_hash = InfoHash([ let info_hash = InfoHash([

View file

@ -14,16 +14,22 @@ pub mod common;
pub mod incoming; pub mod incoming;
pub mod outgoing; pub mod outgoing;
pub use common::*;
pub use incoming::*;
pub use outgoing::*;
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use quickcheck::Arbitrary; use quickcheck::Arbitrary;
use quickcheck_macros::quickcheck; use quickcheck_macros::quickcheck;
use super::*; use crate::{
common::*,
incoming::{
AnnounceEvent, AnnounceRequest, AnnounceRequestOffer, InMessage, ScrapeRequest,
ScrapeRequestInfoHashes,
},
outgoing::{
AnnounceResponse, AnswerOutMessage, OfferOutMessage, OutMessage, ScrapeResponse,
ScrapeStatistics,
},
};
fn arbitrary_20_bytes(g: &mut quickcheck::Gen) -> [u8; 20] { fn arbitrary_20_bytes(g: &mut quickcheck::Gen) -> [u8; 20] {
let mut bytes = [0u8; 20]; let mut bytes = [0u8; 20];