Merge pull request #86 from greatest-ape/more-compact-peers

Reduce ValidUntil size; reduce size of various ws structs; improve aquatic_http_protocol
This commit is contained in:
Joakim Frostegård 2022-08-06 12:09:41 +02:00 committed by GitHub
commit c0ed0eb7db
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
24 changed files with 376 additions and 214 deletions

View file

@ -207,7 +207,11 @@ Implements:
* Doesn't allow full scrapes, i.e. of all registered info hashes * Doesn't allow full scrapes, i.e. of all registered info hashes
`aquatic_http` has not been tested as much as `aquatic_udp` but likely works `aquatic_http` has not been tested as much as `aquatic_udp` but likely works
fine. fine in production.
Running behind a reverse proxy is currently not supported due to the
[difficulties of determining the originating IP address](https://adam-p.ca/blog/2022/03/x-forwarded-for/)
without knowing the exact setup.
#### Performance #### Performance
@ -220,12 +224,14 @@ More details are available [here](./documents/aquatic-http-load-test-2022-04-11.
Aims for compatibility with [WebTorrent](https://github.com/webtorrent) Aims for compatibility with [WebTorrent](https://github.com/webtorrent)
clients. Notes: clients. Notes:
* Only runs over TLS
* Doesn't track the number of torrent downloads (0 is always sent). * Doesn't track the number of torrent downloads (0 is always sent).
* Doesn't allow full scrapes, i.e. of all registered info hashes * Doesn't allow full scrapes, i.e. of all registered info hashes
`aquatic_ws` has not been tested as much as `aquatic_udp` but likely works `aquatic_ws` has not been tested as much as `aquatic_udp` but likely works
fine. fine in production.
Running behind a reverse proxy is supported, as long as IPv4 requests are
proxied to IPv4 requests, and IPv6 requests to IPv6 requests.
#### Performance #### Performance

View file

@ -4,6 +4,9 @@
## Medium priority ## Medium priority
* Consider replacing unmaintained indexmap-amortized with plain indexmap
* Run cargo-fuzz on protocol crates
* quit whole program if any thread panics * quit whole program if any thread panics
* But it would be nice not to panic in workers, but to return errors instead. * But it would be nice not to panic in workers, but to return errors instead.
Once JoinHandle::is_finished is available in stable Rust (#90470), an Once JoinHandle::is_finished is available in stable Rust (#90470), an
@ -20,8 +23,6 @@
* stagger cleaning tasks? * stagger cleaning tasks?
* aquatic_ws * aquatic_ws
* Can peer IP address change after connection has been established
due to some kind of renegotition? It would cause issues.
* Add cleaning task for ConnectionHandle.announced_info_hashes? * Add cleaning task for ConnectionHandle.announced_info_hashes?
* RES memory still high after traffic stops, even if torrent maps and connection slabs go down to 0 len and capacity * RES memory still high after traffic stops, even if torrent maps and connection slabs go down to 0 len and capacity
* replacing indexmap_amortized / simd_json with equivalents doesn't help * replacing indexmap_amortized / simd_json with equivalents doesn't help

View file

@ -1,7 +1,7 @@
use std::net::{Ipv4Addr, SocketAddr, SocketAddrV4, SocketAddrV6}; use std::net::{Ipv4Addr, SocketAddr, SocketAddrV4, SocketAddrV6};
use std::sync::atomic::{AtomicBool, Ordering}; use std::sync::atomic::{AtomicBool, Ordering};
use std::sync::Arc; use std::sync::Arc;
use std::time::{Duration, Instant}; use std::time::Instant;
use ahash::RandomState; use ahash::RandomState;
use rand::Rng; use rand::Rng;
@ -16,23 +16,46 @@ pub mod rustls_config;
/// Amortized IndexMap using AHash hasher /// Amortized IndexMap using AHash hasher
pub type AmortizedIndexMap<K, V> = indexmap_amortized::IndexMap<K, V, RandomState>; pub type AmortizedIndexMap<K, V> = indexmap_amortized::IndexMap<K, V, RandomState>;
/// Peer or connection valid until this instant /// Peer, connection or similar valid until this instant
///
/// Used instead of "last seen" or similar to hopefully prevent arithmetic
/// overflow when cleaning.
#[derive(Debug, Clone, Copy)] #[derive(Debug, Clone, Copy)]
pub struct ValidUntil(pub Instant); pub struct ValidUntil(SecondsSinceServerStart);
impl ValidUntil { impl ValidUntil {
#[inline] #[inline]
pub fn new(offset_seconds: u64) -> Self { pub fn new(start_instant: ServerStartInstant, offset_seconds: u32) -> Self {
Self(Instant::now() + Duration::from_secs(offset_seconds)) Self(SecondsSinceServerStart(
start_instant.seconds_elapsed().0 + offset_seconds,
))
} }
pub fn new_with_now(now: Instant, offset_seconds: u64) -> Self { pub fn new_with_now(now: SecondsSinceServerStart, offset_seconds: u32) -> Self {
Self(now + Duration::from_secs(offset_seconds)) Self(SecondsSinceServerStart(now.0 + offset_seconds))
}
pub fn valid(&self, now: SecondsSinceServerStart) -> bool {
self.0 .0 > now.0
} }
} }
#[derive(Debug, Clone, Copy)]
pub struct ServerStartInstant(Instant);
impl ServerStartInstant {
pub fn new() -> Self {
Self(Instant::now())
}
pub fn seconds_elapsed(&self) -> SecondsSinceServerStart {
SecondsSinceServerStart(
self.0
.elapsed()
.as_secs()
.try_into()
.expect("server ran for more seconds than what fits in a u32"),
)
}
}
#[derive(Debug, Clone, Copy)]
pub struct SecondsSinceServerStart(u32);
pub struct PanicSentinelWatcher(Arc<AtomicBool>); pub struct PanicSentinelWatcher(Arc<AtomicBool>);
impl PanicSentinelWatcher { impl PanicSentinelWatcher {

View file

@ -112,9 +112,9 @@ pub struct CleaningConfig {
/// Clean connections this often (seconds) /// Clean connections this often (seconds)
pub connection_cleaning_interval: u64, pub connection_cleaning_interval: u64,
/// Remove peers that have not announced for this long (seconds) /// Remove peers that have not announced for this long (seconds)
pub max_peer_age: u64, pub max_peer_age: u32,
/// Remove connections that haven't seen valid requests for this long (seconds) /// Remove connections that haven't seen valid requests for this long (seconds)
pub max_connection_idle: u64, pub max_connection_idle: u32,
} }
impl Default for CleaningConfig { impl Default for CleaningConfig {

View file

@ -6,7 +6,7 @@ use aquatic_common::{
}, },
privileges::PrivilegeDropper, privileges::PrivilegeDropper,
rustls_config::create_rustls_config, rustls_config::create_rustls_config,
PanicSentinelWatcher, PanicSentinelWatcher, ServerStartInstant,
}; };
use common::State; use common::State;
use glommio::{channels::channel_mesh::MeshBuilder, prelude::*}; use glommio::{channels::channel_mesh::MeshBuilder, prelude::*};
@ -46,6 +46,8 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
&config.network.tls_private_key_path, &config.network.tls_private_key_path,
)?); )?);
let server_start_instant = ServerStartInstant::new();
let mut executors = Vec::new(); let mut executors = Vec::new();
for i in 0..(config.socket_workers) { for i in 0..(config.socket_workers) {
@ -73,6 +75,7 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
tls_config, tls_config,
request_mesh_builder, request_mesh_builder,
priv_dropper, priv_dropper,
server_start_instant,
) )
.await .await
}) })
@ -97,8 +100,14 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
let executor = builder let executor = builder
.spawn(move || async move { .spawn(move || async move {
workers::swarm::run_swarm_worker(sentinel, config, state, request_mesh_builder) workers::swarm::run_swarm_worker(
.await sentinel,
config,
state,
request_mesh_builder,
server_start_instant,
)
.await
}) })
.map_err(|err| anyhow::anyhow!("Spawning executor failed: {:#}", err))?; .map_err(|err| anyhow::anyhow!("Spawning executor failed: {:#}", err))?;

View file

@ -3,13 +3,13 @@ use std::collections::BTreeMap;
use std::os::unix::prelude::{FromRawFd, IntoRawFd}; use std::os::unix::prelude::{FromRawFd, IntoRawFd};
use std::rc::Rc; use std::rc::Rc;
use std::sync::Arc; use std::sync::Arc;
use std::time::{Duration, Instant}; use std::time::Duration;
use anyhow::Context; use anyhow::Context;
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache}; use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache};
use aquatic_common::privileges::PrivilegeDropper; use aquatic_common::privileges::PrivilegeDropper;
use aquatic_common::rustls_config::RustlsConfig; use aquatic_common::rustls_config::RustlsConfig;
use aquatic_common::{CanonicalSocketAddr, PanicSentinel}; use aquatic_common::{CanonicalSocketAddr, PanicSentinel, ServerStartInstant};
use aquatic_http_protocol::common::InfoHash; use aquatic_http_protocol::common::InfoHash;
use aquatic_http_protocol::request::{Request, RequestParseError, ScrapeRequest}; use aquatic_http_protocol::request::{Request, RequestParseError, ScrapeRequest};
use aquatic_http_protocol::response::{ use aquatic_http_protocol::response::{
@ -59,6 +59,7 @@ pub async fn run_socket_worker(
tls_config: Arc<RustlsConfig>, tls_config: Arc<RustlsConfig>,
request_mesh_builder: MeshBuilder<ChannelRequest, Partial>, request_mesh_builder: MeshBuilder<ChannelRequest, Partial>,
priv_dropper: PrivilegeDropper, priv_dropper: PrivilegeDropper,
server_start_instant: ServerStartInstant,
) { ) {
let config = Rc::new(config); let config = Rc::new(config);
let access_list = state.access_list; let access_list = state.access_list;
@ -74,6 +75,7 @@ pub async fn run_socket_worker(
clean_connections( clean_connections(
config.clone(), config.clone(),
connection_slab.clone(), connection_slab.clone(),
server_start_instant,
) )
})); }));
@ -84,7 +86,10 @@ pub async fn run_socket_worker(
Ok(stream) => { Ok(stream) => {
let key = connection_slab.borrow_mut().insert(ConnectionReference { let key = connection_slab.borrow_mut().insert(ConnectionReference {
task_handle: None, task_handle: None,
valid_until: ValidUntil::new(config.cleaning.max_connection_idle), valid_until: ValidUntil::new(
server_start_instant,
config.cleaning.max_connection_idle,
),
}); });
let task_handle = spawn_local(enclose!((config, access_list, request_senders, tls_config, connection_slab) async move { let task_handle = spawn_local(enclose!((config, access_list, request_senders, tls_config, connection_slab) async move {
@ -92,6 +97,7 @@ pub async fn run_socket_worker(
config, config,
access_list, access_list,
request_senders, request_senders,
server_start_instant,
ConnectionId(key), ConnectionId(key),
tls_config, tls_config,
connection_slab.clone(), connection_slab.clone(),
@ -118,11 +124,12 @@ pub async fn run_socket_worker(
async fn clean_connections( async fn clean_connections(
config: Rc<Config>, config: Rc<Config>,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
server_start_instant: ServerStartInstant,
) -> Option<Duration> { ) -> Option<Duration> {
let now = Instant::now(); let now = server_start_instant.seconds_elapsed();
connection_slab.borrow_mut().retain(|_, reference| { connection_slab.borrow_mut().retain(|_, reference| {
if reference.valid_until.0 > now { if reference.valid_until.valid(now) {
true true
} else { } else {
if let Some(ref handle) = reference.task_handle { if let Some(ref handle) = reference.task_handle {
@ -145,6 +152,7 @@ struct Connection {
access_list_cache: AccessListCache, access_list_cache: AccessListCache,
request_senders: Rc<Senders<ChannelRequest>>, request_senders: Rc<Senders<ChannelRequest>>,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
server_start_instant: ServerStartInstant,
stream: TlsStream<TcpStream>, stream: TlsStream<TcpStream>,
peer_addr: CanonicalSocketAddr, peer_addr: CanonicalSocketAddr,
connection_id: ConnectionId, connection_id: ConnectionId,
@ -158,6 +166,7 @@ impl Connection {
config: Rc<Config>, config: Rc<Config>,
access_list: Arc<AccessListArcSwap>, access_list: Arc<AccessListArcSwap>,
request_senders: Rc<Senders<ChannelRequest>>, request_senders: Rc<Senders<ChannelRequest>>,
server_start_instant: ServerStartInstant,
connection_id: ConnectionId, connection_id: ConnectionId,
tls_config: Arc<RustlsConfig>, tls_config: Arc<RustlsConfig>,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
@ -180,6 +189,7 @@ impl Connection {
access_list_cache: create_access_list_cache(&access_list), access_list_cache: create_access_list_cache(&access_list),
request_senders: request_senders.clone(), request_senders: request_senders.clone(),
connection_slab, connection_slab,
server_start_instant,
stream, stream,
peer_addr, peer_addr,
connection_id, connection_id,
@ -271,7 +281,10 @@ impl Connection {
async fn handle_request(&mut self, request: Request) -> anyhow::Result<Response> { async fn handle_request(&mut self, request: Request) -> anyhow::Result<Response> {
if let Ok(mut slab) = self.connection_slab.try_borrow_mut() { if let Ok(mut slab) = self.connection_slab.try_borrow_mut() {
if let Some(reference) = slab.get_mut(self.connection_id.0) { if let Some(reference) = slab.get_mut(self.connection_id.0) {
reference.valid_until = ValidUntil::new(self.config.cleaning.max_connection_idle); reference.valid_until = ValidUntil::new(
self.server_start_instant,
self.config.cleaning.max_connection_idle,
);
} }
} }

View file

@ -4,7 +4,6 @@ use std::net::{IpAddr, Ipv4Addr, Ipv6Addr};
use std::rc::Rc; use std::rc::Rc;
use std::sync::Arc; use std::sync::Arc;
use std::time::Duration; use std::time::Duration;
use std::time::Instant;
use either::Either; use either::Either;
use futures_lite::{Stream, StreamExt}; use futures_lite::{Stream, StreamExt};
@ -17,9 +16,9 @@ use rand::SeedableRng;
use smartstring::{LazyCompact, SmartString}; use smartstring::{LazyCompact, SmartString};
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache}; use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache};
use aquatic_common::ValidUntil;
use aquatic_common::{extract_response_peers, PanicSentinel}; use aquatic_common::{extract_response_peers, PanicSentinel};
use aquatic_common::{AmortizedIndexMap, CanonicalSocketAddr}; use aquatic_common::{AmortizedIndexMap, CanonicalSocketAddr};
use aquatic_common::{SecondsSinceServerStart, ServerStartInstant, ValidUntil};
use aquatic_http_protocol::common::*; use aquatic_http_protocol::common::*;
use aquatic_http_protocol::request::*; use aquatic_http_protocol::request::*;
use aquatic_http_protocol::response::ResponsePeer; use aquatic_http_protocol::response::ResponsePeer;
@ -107,20 +106,26 @@ pub struct TorrentMaps {
} }
impl TorrentMaps { impl TorrentMaps {
pub fn clean(&mut self, config: &Config, access_list: &Arc<AccessListArcSwap>) { pub fn clean(
&mut self,
config: &Config,
access_list: &Arc<AccessListArcSwap>,
server_start_instant: ServerStartInstant,
) {
let mut access_list_cache = create_access_list_cache(access_list); let mut access_list_cache = create_access_list_cache(access_list);
Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv4); let now = server_start_instant.seconds_elapsed();
Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv6);
Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv4, now);
Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv6, now);
} }
fn clean_torrent_map<I: Ip>( fn clean_torrent_map<I: Ip>(
config: &Config, config: &Config,
access_list_cache: &mut AccessListCache, access_list_cache: &mut AccessListCache,
torrent_map: &mut TorrentMap<I>, torrent_map: &mut TorrentMap<I>,
now: SecondsSinceServerStart,
) { ) {
let now = Instant::now();
torrent_map.retain(|info_hash, torrent_data| { torrent_map.retain(|info_hash, torrent_data| {
if !access_list_cache if !access_list_cache
.load() .load()
@ -133,7 +138,7 @@ impl TorrentMaps {
let num_leechers = &mut torrent_data.num_leechers; let num_leechers = &mut torrent_data.num_leechers;
torrent_data.peers.retain(|_, peer| { torrent_data.peers.retain(|_, peer| {
let keep = peer.valid_until.0 >= now; let keep = peer.valid_until.valid(now);
if !keep { if !keep {
match peer.status { match peer.status {
@ -162,6 +167,7 @@ pub async fn run_swarm_worker(
config: Config, config: Config,
state: State, state: State,
request_mesh_builder: MeshBuilder<ChannelRequest, Partial>, request_mesh_builder: MeshBuilder<ChannelRequest, Partial>,
server_start_instant: ServerStartInstant,
) { ) {
let (_, mut request_receivers) = request_mesh_builder.join(Role::Consumer).await.unwrap(); let (_, mut request_receivers) = request_mesh_builder.join(Role::Consumer).await.unwrap();
@ -171,19 +177,22 @@ pub async fn run_swarm_worker(
// Periodically clean torrents // Periodically clean torrents
TimerActionRepeat::repeat(enclose!((config, torrents, access_list) move || { TimerActionRepeat::repeat(enclose!((config, torrents, access_list) move || {
enclose!((config, torrents, access_list) move || async move { enclose!((config, torrents, access_list) move || async move {
torrents.borrow_mut().clean(&config, &access_list); torrents.borrow_mut().clean(&config, &access_list, server_start_instant);
Some(Duration::from_secs(config.cleaning.torrent_cleaning_interval)) Some(Duration::from_secs(config.cleaning.torrent_cleaning_interval))
})() })()
})); }));
let max_peer_age = config.cleaning.max_peer_age; let max_peer_age = config.cleaning.max_peer_age;
let peer_valid_until = Rc::new(RefCell::new(ValidUntil::new(max_peer_age))); let peer_valid_until = Rc::new(RefCell::new(ValidUntil::new(
server_start_instant,
max_peer_age,
)));
// Periodically update peer_valid_until // Periodically update peer_valid_until
TimerActionRepeat::repeat(enclose!((peer_valid_until) move || { TimerActionRepeat::repeat(enclose!((peer_valid_until) move || {
enclose!((peer_valid_until) move || async move { enclose!((peer_valid_until) move || async move {
*peer_valid_until.borrow_mut() = ValidUntil::new(max_peer_age); *peer_valid_until.borrow_mut() = ValidUntil::new(server_start_instant, max_peer_age);
Some(Duration::from_secs(1)) Some(Duration::from_secs(1))
})() })()

View file

@ -99,7 +99,7 @@ pub struct CleaningConfig {
/// Clean peers this often (seconds) /// Clean peers this often (seconds)
pub torrent_cleaning_interval: u64, pub torrent_cleaning_interval: u64,
/// Remove peers that have not announced for this long (seconds) /// Remove peers that have not announced for this long (seconds)
pub max_peer_age: u64, pub max_peer_age: u32,
} }
impl Default for CleaningConfig { impl Default for CleaningConfig {

View file

@ -6,6 +6,7 @@ use std::{collections::VecDeque, sync::Arc};
use aquatic_common::{ use aquatic_common::{
privileges::PrivilegeDropper, rustls_config::create_rustls_config, PanicSentinelWatcher, privileges::PrivilegeDropper, rustls_config::create_rustls_config, PanicSentinelWatcher,
ServerStartInstant,
}; };
use common::ChannelRequestSender; use common::ChannelRequestSender;
use dotenv::dotenv; use dotenv::dotenv;
@ -40,6 +41,8 @@ pub fn run(config: Config) -> anyhow::Result<()> {
let (sentinel_watcher, sentinel) = PanicSentinelWatcher::create_with_sentinel(); let (sentinel_watcher, sentinel) = PanicSentinelWatcher::create_with_sentinel();
let priv_dropper = PrivilegeDropper::new(config.privileges.clone(), config.socket_workers); let priv_dropper = PrivilegeDropper::new(config.privileges.clone(), config.socket_workers);
let server_start_instant = ServerStartInstant::new();
let mut handles = Vec::new(); let mut handles = Vec::new();
for _ in 0..config.socket_workers { for _ in 0..config.socket_workers {
@ -71,7 +74,14 @@ pub fn run(config: Config) -> anyhow::Result<()> {
let handle = ::std::thread::Builder::new() let handle = ::std::thread::Builder::new()
.name("request".into()) .name("request".into())
.spawn(move || workers::swarm::run_swarm_worker(sentinel, config, request_receiver))?; .spawn(move || {
workers::swarm::run_swarm_worker(
sentinel,
config,
request_receiver,
server_start_instant,
)
})?;
handles.push(handle); handles.push(handle);
} }

View file

@ -1,7 +1,6 @@
use std::net::{Ipv4Addr, Ipv6Addr}; use std::net::{Ipv4Addr, Ipv6Addr};
use std::time::Instant;
use aquatic_common::{AmortizedIndexMap, ValidUntil}; use aquatic_common::{AmortizedIndexMap, SecondsSinceServerStart, ServerStartInstant, ValidUntil};
use aquatic_http_protocol::common::{AnnounceEvent, InfoHash, PeerId}; use aquatic_http_protocol::common::{AnnounceEvent, InfoHash, PeerId};
use aquatic_http_protocol::response::ResponsePeer; use aquatic_http_protocol::response::ResponsePeer;
@ -84,20 +83,20 @@ pub struct TorrentMaps {
} }
impl TorrentMaps { impl TorrentMaps {
pub fn clean(&mut self) { pub fn clean(&mut self, server_start_instant: ServerStartInstant) {
Self::clean_torrent_map(&mut self.ipv4); let now = server_start_instant.seconds_elapsed();
Self::clean_torrent_map(&mut self.ipv6);
Self::clean_torrent_map(&mut self.ipv4, now);
Self::clean_torrent_map(&mut self.ipv6, now);
} }
fn clean_torrent_map<I: Ip>(torrent_map: &mut TorrentMap<I>) { fn clean_torrent_map<I: Ip>(torrent_map: &mut TorrentMap<I>, now: SecondsSinceServerStart) {
let now = Instant::now();
torrent_map.retain(|_, torrent_data| { torrent_map.retain(|_, torrent_data| {
let num_seeders = &mut torrent_data.num_seeders; let num_seeders = &mut torrent_data.num_seeders;
let num_leechers = &mut torrent_data.num_leechers; let num_leechers = &mut torrent_data.num_leechers;
torrent_data.peers.retain(|_, peer| { torrent_data.peers.retain(|_, peer| {
if peer.valid_until.0 >= now { if peer.valid_until.valid(now) {
true true
} else { } else {
match peer.status { match peer.status {

View file

@ -11,7 +11,9 @@ use tokio::sync::mpsc::Receiver;
use tokio::task::LocalSet; use tokio::task::LocalSet;
use tokio::time; use tokio::time;
use aquatic_common::{extract_response_peers, CanonicalSocketAddr, PanicSentinel, ValidUntil}; use aquatic_common::{
extract_response_peers, CanonicalSocketAddr, PanicSentinel, ServerStartInstant, ValidUntil,
};
use aquatic_http_protocol::response::{ use aquatic_http_protocol::response::{
AnnounceResponse, Response, ResponsePeer, ResponsePeerListV4, ResponsePeerListV6, AnnounceResponse, Response, ResponsePeer, ResponsePeerListV4, ResponsePeerListV6,
}; };
@ -25,12 +27,13 @@ pub fn run_swarm_worker(
_sentinel: PanicSentinel, _sentinel: PanicSentinel,
config: Config, config: Config,
request_receiver: Receiver<ChannelAnnounceRequest>, request_receiver: Receiver<ChannelAnnounceRequest>,
server_start_instant: ServerStartInstant,
) -> anyhow::Result<()> { ) -> anyhow::Result<()> {
let runtime = tokio::runtime::Builder::new_current_thread() let runtime = tokio::runtime::Builder::new_current_thread()
.enable_all() .enable_all()
.build()?; .build()?;
runtime.block_on(run_inner(config, request_receiver))?; runtime.block_on(run_inner(config, request_receiver, server_start_instant))?;
Ok(()) Ok(())
} }
@ -38,6 +41,7 @@ pub fn run_swarm_worker(
async fn run_inner( async fn run_inner(
config: Config, config: Config,
mut request_receiver: Receiver<ChannelAnnounceRequest>, mut request_receiver: Receiver<ChannelAnnounceRequest>,
server_start_instant: ServerStartInstant,
) -> anyhow::Result<()> { ) -> anyhow::Result<()> {
let torrents = Rc::new(RefCell::new(TorrentMaps::default())); let torrents = Rc::new(RefCell::new(TorrentMaps::default()));
let mut rng = SmallRng::from_entropy(); let mut rng = SmallRng::from_entropy();
@ -45,6 +49,7 @@ async fn run_inner(
LocalSet::new().spawn_local(periodically_clean_torrents( LocalSet::new().spawn_local(periodically_clean_torrents(
config.clone(), config.clone(),
torrents.clone(), torrents.clone(),
server_start_instant,
)); ));
loop { loop {
@ -53,7 +58,7 @@ async fn run_inner(
.await .await
.ok_or_else(|| anyhow::anyhow!("request channel closed"))?; .ok_or_else(|| anyhow::anyhow!("request channel closed"))?;
let valid_until = ValidUntil::new(config.cleaning.max_peer_age); let valid_until = ValidUntil::new(server_start_instant, config.cleaning.max_peer_age);
let response = handle_announce_request( let response = handle_announce_request(
&config, &config,
@ -68,7 +73,11 @@ async fn run_inner(
} }
} }
async fn periodically_clean_torrents(config: Config, torrents: Rc<RefCell<TorrentMaps>>) { async fn periodically_clean_torrents(
config: Config,
torrents: Rc<RefCell<TorrentMaps>>,
server_start_instant: ServerStartInstant,
) {
let mut interval = time::interval(time::Duration::from_secs( let mut interval = time::interval(time::Duration::from_secs(
config.cleaning.torrent_cleaning_interval, config.cleaning.torrent_cleaning_interval,
)); ));
@ -76,7 +85,7 @@ async fn periodically_clean_torrents(config: Config, torrents: Rc<RefCell<Torren
loop { loop {
interval.tick().await; interval.tick().await;
torrents.borrow_mut().clean(); torrents.borrow_mut().clean(server_start_instant);
} }
} }

View file

@ -250,6 +250,17 @@ pub enum RequestParseError {
Invalid(anyhow::Error), Invalid(anyhow::Error),
} }
impl ::std::fmt::Display for RequestParseError {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
match self {
Self::NeedMoreData => write!(f, "Incomplete request, more data needed"),
Self::Invalid(err) => write!(f, "Invalid request: {:#}", err),
}
}
}
impl ::std::error::Error for RequestParseError {}
#[derive(Debug, Clone, PartialEq, Eq)] #[derive(Debug, Clone, PartialEq, Eq)]
pub enum Request { pub enum Request {
Announce(AnnounceRequest), Announce(AnnounceRequest),
@ -262,28 +273,20 @@ impl Request {
let mut headers = [httparse::EMPTY_HEADER; 16]; let mut headers = [httparse::EMPTY_HEADER; 16];
let mut http_request = httparse::Request::new(&mut headers); let mut http_request = httparse::Request::new(&mut headers);
let path = match http_request.parse(bytes) { match http_request.parse(bytes) {
Ok(httparse::Status::Complete(_)) => { Ok(httparse::Status::Complete(_)) => {
if let Some(path) = http_request.path { if let Some(path) = http_request.path {
path Self::from_http_get_path(path).map_err(RequestParseError::Invalid)
} else { } else {
return Err(RequestParseError::Invalid(anyhow::anyhow!("no http path"))); Err(RequestParseError::Invalid(anyhow::anyhow!("no http path")))
} }
} }
Ok(httparse::Status::Partial) => { Ok(httparse::Status::Partial) => Err(RequestParseError::NeedMoreData),
if let Some(path) = http_request.path { Err(err) => Err(RequestParseError::Invalid(anyhow::Error::from(err))),
path }
} else {
return Err(RequestParseError::NeedMoreData);
}
}
Err(err) => return Err(RequestParseError::Invalid(anyhow::Error::from(err))),
};
Self::from_http_get_path(path).map_err(RequestParseError::Invalid)
} }
/// Parse Request from http path (GET `/announce?info_hash=...`) /// Parse Request from http GET path (`/announce?info_hash=...`)
/// ///
/// Existing serde-url decode crates were insufficient, so the decision was /// Existing serde-url decode crates were insufficient, so the decision was
/// made to create a custom parser. serde_urlencoded doesn't support multiple /// made to create a custom parser. serde_urlencoded doesn't support multiple
@ -308,10 +311,12 @@ impl Request {
Ok(Request::Announce(AnnounceRequest::from_query_string( Ok(Request::Announce(AnnounceRequest::from_query_string(
query_string, query_string,
)?)) )?))
} else { } else if location == "/scrape" {
Ok(Request::Scrape(ScrapeRequest::from_query_string( Ok(Request::Scrape(ScrapeRequest::from_query_string(
query_string, query_string,
)?)) )?))
} else {
Err(anyhow::anyhow!("Path must be /announce or /scrape"))
} }
} }

View file

@ -181,10 +181,10 @@ pub struct CleaningConfig {
/// Allow clients to use a connection token for this long (seconds) /// Allow clients to use a connection token for this long (seconds)
pub max_connection_age: u32, pub max_connection_age: u32,
/// Remove peers who have not announced for this long (seconds) /// Remove peers who have not announced for this long (seconds)
pub max_peer_age: u64, pub max_peer_age: u32,
/// Remove pending scrape responses that have not been returned from swarm /// Remove pending scrape responses that have not been returned from swarm
/// workers for this long (seconds) /// workers for this long (seconds)
pub max_pending_scrape_age: u64, pub max_pending_scrape_age: u32,
} }
impl Default for CleaningConfig { impl Default for CleaningConfig {

View file

@ -14,7 +14,7 @@ use aquatic_common::access_list::update_access_list;
#[cfg(feature = "cpu-pinning")] #[cfg(feature = "cpu-pinning")]
use aquatic_common::cpu_pinning::{pin_current_if_configured_to, WorkerIndex}; use aquatic_common::cpu_pinning::{pin_current_if_configured_to, WorkerIndex};
use aquatic_common::privileges::PrivilegeDropper; use aquatic_common::privileges::PrivilegeDropper;
use aquatic_common::PanicSentinelWatcher; use aquatic_common::{PanicSentinelWatcher, ServerStartInstant};
use common::{ use common::{
ConnectedRequestSender, ConnectedResponseSender, SocketWorkerIndex, State, SwarmWorkerIndex, ConnectedRequestSender, ConnectedResponseSender, SocketWorkerIndex, State, SwarmWorkerIndex,
@ -41,6 +41,8 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
let mut response_senders = Vec::new(); let mut response_senders = Vec::new();
let mut response_receivers = BTreeMap::new(); let mut response_receivers = BTreeMap::new();
let server_start_instant = ServerStartInstant::new();
for i in 0..config.swarm_workers { for i in 0..config.swarm_workers {
let (request_sender, request_receiver) = if config.worker_channel_size == 0 { let (request_sender, request_receiver) = if config.worker_channel_size == 0 {
unbounded() unbounded()
@ -85,6 +87,7 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
sentinel, sentinel,
config, config,
state, state,
server_start_instant,
request_receiver, request_receiver,
response_sender, response_sender,
SwarmWorkerIndex(i), SwarmWorkerIndex(i),
@ -120,6 +123,7 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
config, config,
i, i,
connection_validator, connection_validator,
server_start_instant,
request_sender, request_sender,
response_receiver, response_receiver,
priv_dropper, priv_dropper,

View file

@ -6,6 +6,7 @@ pub mod validator;
use std::time::{Duration, Instant}; use std::time::{Duration, Instant};
use anyhow::Context; use anyhow::Context;
use aquatic_common::ServerStartInstant;
use crossbeam_channel::Receiver; use crossbeam_channel::Receiver;
use mio::net::UdpSocket; use mio::net::UdpSocket;
use mio::{Events, Interest, Poll, Token}; use mio::{Events, Interest, Poll, Token};
@ -31,6 +32,7 @@ pub fn run_socket_worker(
config: Config, config: Config,
token_num: usize, token_num: usize,
mut connection_validator: ConnectionValidator, mut connection_validator: ConnectionValidator,
server_start_instant: ServerStartInstant,
request_sender: ConnectedRequestSender, request_sender: ConnectedRequestSender,
response_receiver: Receiver<(ConnectedResponse, CanonicalSocketAddr)>, response_receiver: Receiver<(ConnectedResponse, CanonicalSocketAddr)>,
priv_dropper: PrivilegeDropper, priv_dropper: PrivilegeDropper,
@ -59,7 +61,8 @@ pub fn run_socket_worker(
let pending_scrape_cleaning_duration = let pending_scrape_cleaning_duration =
Duration::from_secs(config.cleaning.pending_scrape_cleaning_interval); Duration::from_secs(config.cleaning.pending_scrape_cleaning_interval);
let mut pending_scrape_valid_until = ValidUntil::new(config.cleaning.max_pending_scrape_age); let mut pending_scrape_valid_until =
ValidUntil::new(server_start_instant, config.cleaning.max_pending_scrape_age);
let mut last_pending_scrape_cleaning = Instant::now(); let mut last_pending_scrape_cleaning = Instant::now();
let mut iter_counter = 0usize; let mut iter_counter = 0usize;
@ -100,13 +103,17 @@ pub fn run_socket_worker(
// Run periodic ValidUntil updates and state cleaning // Run periodic ValidUntil updates and state cleaning
if iter_counter % 256 == 0 { if iter_counter % 256 == 0 {
let seconds_since_start = server_start_instant.seconds_elapsed();
pending_scrape_valid_until = ValidUntil::new_with_now(
seconds_since_start,
config.cleaning.max_pending_scrape_age,
);
let now = Instant::now(); let now = Instant::now();
pending_scrape_valid_until =
ValidUntil::new_with_now(now, config.cleaning.max_pending_scrape_age);
if now > last_pending_scrape_cleaning + pending_scrape_cleaning_duration { if now > last_pending_scrape_cleaning + pending_scrape_cleaning_duration {
pending_scrape_responses.clean(); pending_scrape_responses.clean(seconds_since_start);
last_pending_scrape_cleaning = now; last_pending_scrape_cleaning = now;
} }

View file

@ -1,10 +1,9 @@
use std::collections::BTreeMap; use std::collections::BTreeMap;
use std::time::Instant;
use hashbrown::HashMap; use hashbrown::HashMap;
use slab::Slab; use slab::Slab;
use aquatic_common::ValidUntil; use aquatic_common::{SecondsSinceServerStart, ValidUntil};
use aquatic_udp_protocol::*; use aquatic_udp_protocol::*;
use crate::common::*; use crate::common::*;
@ -97,11 +96,9 @@ impl PendingScrapeResponseSlab {
} }
} }
pub fn clean(&mut self) { pub fn clean(&mut self, now: SecondsSinceServerStart) {
let now = Instant::now();
self.0.retain(|k, v| { self.0.retain(|k, v| {
if v.valid_until.0 > now { if v.valid_until.valid(now) {
true true
} else { } else {
::log::warn!( ::log::warn!(
@ -120,6 +117,7 @@ impl PendingScrapeResponseSlab {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use aquatic_common::ServerStartInstant;
use quickcheck::TestResult; use quickcheck::TestResult;
use quickcheck_macros::quickcheck; use quickcheck_macros::quickcheck;
@ -138,7 +136,7 @@ mod tests {
config.swarm_workers = swarm_workers as usize; config.swarm_workers = swarm_workers as usize;
let valid_until = ValidUntil::new(1); let valid_until = ValidUntil::new(ServerStartInstant::new(), 1);
let mut map = PendingScrapeResponseSlab::default(); let mut map = PendingScrapeResponseSlab::default();

View file

@ -5,6 +5,7 @@ use std::sync::atomic::Ordering;
use std::time::Duration; use std::time::Duration;
use std::time::Instant; use std::time::Instant;
use aquatic_common::ServerStartInstant;
use crossbeam_channel::Receiver; use crossbeam_channel::Receiver;
use rand::{rngs::SmallRng, SeedableRng}; use rand::{rngs::SmallRng, SeedableRng};
@ -21,6 +22,7 @@ pub fn run_swarm_worker(
_sentinel: PanicSentinel, _sentinel: PanicSentinel,
config: Config, config: Config,
state: State, state: State,
server_start_instant: ServerStartInstant,
request_receiver: Receiver<(SocketWorkerIndex, ConnectedRequest, CanonicalSocketAddr)>, request_receiver: Receiver<(SocketWorkerIndex, ConnectedRequest, CanonicalSocketAddr)>,
response_sender: ConnectedResponseSender, response_sender: ConnectedResponseSender,
worker_index: SwarmWorkerIndex, worker_index: SwarmWorkerIndex,
@ -29,7 +31,7 @@ pub fn run_swarm_worker(
let mut rng = SmallRng::from_entropy(); let mut rng = SmallRng::from_entropy();
let timeout = Duration::from_millis(config.request_channel_recv_timeout_ms); let timeout = Duration::from_millis(config.request_channel_recv_timeout_ms);
let mut peer_valid_until = ValidUntil::new(config.cleaning.max_peer_age); let mut peer_valid_until = ValidUntil::new(server_start_instant, config.cleaning.max_peer_age);
let cleaning_interval = Duration::from_secs(config.cleaning.torrent_cleaning_interval); let cleaning_interval = Duration::from_secs(config.cleaning.torrent_cleaning_interval);
let statistics_update_interval = Duration::from_secs(config.statistics.interval); let statistics_update_interval = Duration::from_secs(config.statistics.interval);
@ -81,10 +83,14 @@ pub fn run_swarm_worker(
if iter_counter % 128 == 0 { if iter_counter % 128 == 0 {
let now = Instant::now(); let now = Instant::now();
peer_valid_until = ValidUntil::new_with_now(now, config.cleaning.max_peer_age); peer_valid_until = ValidUntil::new(server_start_instant, config.cleaning.max_peer_age);
if now > last_cleaning + cleaning_interval { if now > last_cleaning + cleaning_interval {
let (ipv4, ipv6) = torrents.clean_and_get_num_peers(&config, &state.access_list); let (ipv4, ipv6) = torrents.clean_and_get_num_peers(
&config,
&state.access_list,
server_start_instant,
);
if config.statistics.active() { if config.statistics.active() {
state.statistics_ipv4.peers[worker_index.0].store(ipv4, Ordering::Release); state.statistics_ipv4.peers[worker_index.0].store(ipv4, Ordering::Release);

View file

@ -1,8 +1,9 @@
use std::net::Ipv4Addr; use std::net::Ipv4Addr;
use std::net::Ipv6Addr; use std::net::Ipv6Addr;
use std::sync::Arc; use std::sync::Arc;
use std::time::Instant;
use aquatic_common::SecondsSinceServerStart;
use aquatic_common::ServerStartInstant;
use aquatic_common::{ use aquatic_common::{
access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache, AccessListMode}, access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache, AccessListMode},
extract_response_peers, AmortizedIndexMap, ValidUntil, extract_response_peers, AmortizedIndexMap, ValidUntil,
@ -99,9 +100,9 @@ impl<I: Ip> TorrentData<I> {
} }
/// Remove inactive peers and reclaim space /// Remove inactive peers and reclaim space
fn clean(&mut self, now: Instant) { fn clean(&mut self, now: SecondsSinceServerStart) {
self.peers.retain(|_, peer| { self.peers.retain(|_, peer| {
if peer.valid_until.0 > now { if peer.valid_until.valid(now) {
true true
} else { } else {
match peer.status { match peer.status {
@ -143,7 +144,7 @@ impl<I: Ip> TorrentMap<I> {
&mut self, &mut self,
access_list_cache: &mut AccessListCache, access_list_cache: &mut AccessListCache,
access_list_mode: AccessListMode, access_list_mode: AccessListMode,
now: Instant, now: SecondsSinceServerStart,
) -> usize { ) -> usize {
let mut num_peers = 0; let mut num_peers = 0;
@ -192,10 +193,11 @@ impl TorrentMaps {
&mut self, &mut self,
config: &Config, config: &Config,
access_list: &Arc<AccessListArcSwap>, access_list: &Arc<AccessListArcSwap>,
server_start_instant: ServerStartInstant,
) -> (usize, usize) { ) -> (usize, usize) {
let mut cache = create_access_list_cache(access_list); let mut cache = create_access_list_cache(access_list);
let mode = config.access_list.mode; let mode = config.access_list.mode;
let now = Instant::now(); let now = server_start_instant.seconds_elapsed();
let ipv4 = self.ipv4.clean_and_get_num_peers(&mut cache, mode, now); let ipv4 = self.ipv4.clean_and_get_num_peers(&mut cache, mode, now);
let ipv6 = self.ipv6.clean_and_get_num_peers(&mut cache, mode, now); let ipv6 = self.ipv6.clean_and_get_num_peers(&mut cache, mode, now);
@ -226,7 +228,7 @@ mod tests {
ip_address: Ipv4Addr::from(i.to_be_bytes()), ip_address: Ipv4Addr::from(i.to_be_bytes()),
port: Port(1), port: Port(1),
status: PeerStatus::Leeching, status: PeerStatus::Leeching,
valid_until: ValidUntil::new(0), valid_until: ValidUntil::new(ServerStartInstant::new(), 0),
} }
} }

View file

@ -7,7 +7,7 @@
//! Scrape: 1 873 545 requests/second, 533.75 ns/request //! Scrape: 1 873 545 requests/second, 533.75 ns/request
//! ``` //! ```
use aquatic_common::PanicSentinelWatcher; use aquatic_common::{PanicSentinelWatcher, ServerStartInstant};
use aquatic_udp::workers::swarm::run_swarm_worker; use aquatic_udp::workers::swarm::run_swarm_worker;
use crossbeam_channel::unbounded; use crossbeam_channel::unbounded;
use num_format::{Locale, ToFormattedString}; use num_format::{Locale, ToFormattedString};
@ -51,6 +51,8 @@ pub fn run(bench_config: BenchConfig) -> ::anyhow::Result<()> {
let response_sender = ConnectedResponseSender::new(vec![response_sender]); let response_sender = ConnectedResponseSender::new(vec![response_sender]);
let server_start_instant = ServerStartInstant::new();
{ {
let config = aquatic_config.clone(); let config = aquatic_config.clone();
let state = State::new(config.swarm_workers); let state = State::new(config.swarm_workers);
@ -60,6 +62,7 @@ pub fn run(bench_config: BenchConfig) -> ::anyhow::Result<()> {
sentinel, sentinel,
config, config,
state, state,
server_start_instant,
request_receiver, request_receiver,
response_sender, response_sender,
SwarmWorkerIndex(0), SwarmWorkerIndex(0),

View file

@ -29,16 +29,16 @@ pub struct State {
} }
#[derive(Copy, Clone, Debug)] #[derive(Copy, Clone, Debug)]
pub struct PendingScrapeId(pub usize); pub struct PendingScrapeId(pub u8);
#[derive(Copy, Clone, Debug)] #[derive(Copy, Clone, Debug)]
pub struct ConsumerId(pub usize); pub struct ConsumerId(pub u8);
#[derive(Clone, Copy, Debug, PartialEq)] #[derive(Clone, Copy, Debug, PartialEq)]
pub struct ConnectionId(pub usize); pub struct ConnectionId(pub usize);
#[derive(Clone, Copy, Debug)] #[derive(Clone, Copy, Debug)]
pub struct ConnectionMeta { pub struct InMessageMeta {
/// Index of socket worker responsible for this connection. Required for /// Index of socket worker responsible for this connection. Required for
/// sending back response through correct channel to correct worker. /// sending back response through correct channel to correct worker.
pub out_message_consumer_id: ConsumerId, pub out_message_consumer_id: ConsumerId,
@ -47,6 +47,25 @@ pub struct ConnectionMeta {
pub pending_scrape_id: Option<PendingScrapeId>, pub pending_scrape_id: Option<PendingScrapeId>,
} }
#[derive(Clone, Copy, Debug)]
pub struct OutMessageMeta {
/// Index of socket worker responsible for this connection. Required for
/// sending back response through correct channel to correct worker.
pub out_message_consumer_id: ConsumerId,
pub connection_id: ConnectionId,
pub pending_scrape_id: Option<PendingScrapeId>,
}
impl Into<OutMessageMeta> for InMessageMeta {
fn into(self) -> OutMessageMeta {
OutMessageMeta {
out_message_consumer_id: self.out_message_consumer_id,
connection_id: self.connection_id,
pending_scrape_id: self.pending_scrape_id,
}
}
}
#[derive(Clone, Copy, Debug)] #[derive(Clone, Copy, Debug)]
pub enum SwarmControlMessage { pub enum SwarmControlMessage {
ConnectionClosed { ConnectionClosed {

View file

@ -124,11 +124,11 @@ pub struct CleaningConfig {
/// Clean peers this often (seconds) /// Clean peers this often (seconds)
pub torrent_cleaning_interval: u64, pub torrent_cleaning_interval: u64,
/// Remove peers that have not announced for this long (seconds) /// Remove peers that have not announced for this long (seconds)
pub max_peer_age: u64, pub max_peer_age: u32,
// Clean connections this often (seconds) // Clean connections this often (seconds)
pub connection_cleaning_interval: u64, pub connection_cleaning_interval: u64,
/// Close connections if no responses have been sent to them for this long (seconds) /// Close connections if no responses have been sent to them for this long (seconds)
pub max_connection_idle: u64, pub max_connection_idle: u32,
} }
impl Default for CleaningConfig { impl Default for CleaningConfig {

View file

@ -8,7 +8,7 @@ use anyhow::Context;
use aquatic_common::cpu_pinning::glommio::{get_worker_placement, set_affinity_for_util_worker}; use aquatic_common::cpu_pinning::glommio::{get_worker_placement, set_affinity_for_util_worker};
use aquatic_common::cpu_pinning::WorkerIndex; use aquatic_common::cpu_pinning::WorkerIndex;
use aquatic_common::rustls_config::create_rustls_config; use aquatic_common::rustls_config::create_rustls_config;
use aquatic_common::PanicSentinelWatcher; use aquatic_common::{PanicSentinelWatcher, ServerStartInstant};
use glommio::{channels::channel_mesh::MeshBuilder, prelude::*}; use glommio::{channels::channel_mesh::MeshBuilder, prelude::*};
use signal_hook::{ use signal_hook::{
consts::{SIGTERM, SIGUSR1}, consts::{SIGTERM, SIGUSR1},
@ -49,14 +49,19 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
let priv_dropper = PrivilegeDropper::new(config.privileges.clone(), config.socket_workers); let priv_dropper = PrivilegeDropper::new(config.privileges.clone(), config.socket_workers);
let opt_tls_config = if config.network.enable_tls { let opt_tls_config = if config.network.enable_tls {
Some(Arc::new(create_rustls_config( Some(Arc::new(
&config.network.tls_certificate_path, create_rustls_config(
&config.network.tls_private_key_path, &config.network.tls_certificate_path,
).with_context(|| "create rustls config")?)) &config.network.tls_private_key_path,
)
.with_context(|| "create rustls config")?,
))
} else { } else {
None None
}; };
let server_start_instant = ServerStartInstant::new();
let mut executors = Vec::new(); let mut executors = Vec::new();
for i in 0..(config.socket_workers) { for i in 0..(config.socket_workers) {
@ -88,6 +93,7 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
request_mesh_builder, request_mesh_builder,
response_mesh_builder, response_mesh_builder,
priv_dropper, priv_dropper,
server_start_instant,
) )
.await .await
}) })
@ -123,6 +129,7 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
control_mesh_builder, control_mesh_builder,
request_mesh_builder, request_mesh_builder,
response_mesh_builder, response_mesh_builder,
server_start_instant,
) )
.await .await
}) })

View file

@ -4,13 +4,13 @@ use std::collections::BTreeMap;
use std::os::unix::prelude::{FromRawFd, IntoRawFd}; use std::os::unix::prelude::{FromRawFd, IntoRawFd};
use std::rc::Rc; use std::rc::Rc;
use std::sync::Arc; use std::sync::Arc;
use std::time::{Duration, Instant}; use std::time::Duration;
use anyhow::Context; use anyhow::Context;
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache}; use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache};
use aquatic_common::privileges::PrivilegeDropper; use aquatic_common::privileges::PrivilegeDropper;
use aquatic_common::rustls_config::RustlsConfig; use aquatic_common::rustls_config::RustlsConfig;
use aquatic_common::PanicSentinel; use aquatic_common::{PanicSentinel, ServerStartInstant};
use aquatic_ws_protocol::*; use aquatic_ws_protocol::*;
use async_tungstenite::WebSocketStream; use async_tungstenite::WebSocketStream;
use futures::stream::{SplitSink, SplitStream}; use futures::stream::{SplitSink, SplitStream};
@ -42,7 +42,7 @@ struct ConnectionReference {
task_handle: Option<JoinHandle<()>>, task_handle: Option<JoinHandle<()>>,
/// Sender part of channel used to pass on outgoing messages from request /// Sender part of channel used to pass on outgoing messages from request
/// worker /// worker
out_message_sender: Rc<LocalSender<(ConnectionMeta, OutMessage)>>, out_message_sender: Rc<LocalSender<(OutMessageMeta, OutMessage)>>,
/// Updated after sending message to peer /// Updated after sending message to peer
valid_until: ValidUntil, valid_until: ValidUntil,
peer_id: Option<PeerId>, peer_id: Option<PeerId>,
@ -56,9 +56,10 @@ pub async fn run_socket_worker(
state: State, state: State,
opt_tls_config: Option<Arc<RustlsConfig>>, opt_tls_config: Option<Arc<RustlsConfig>>,
control_message_mesh_builder: MeshBuilder<SwarmControlMessage, Partial>, control_message_mesh_builder: MeshBuilder<SwarmControlMessage, Partial>,
in_message_mesh_builder: MeshBuilder<(ConnectionMeta, InMessage), Partial>, in_message_mesh_builder: MeshBuilder<(InMessageMeta, InMessage), Partial>,
out_message_mesh_builder: MeshBuilder<(ConnectionMeta, OutMessage), Partial>, out_message_mesh_builder: MeshBuilder<(OutMessageMeta, OutMessage), Partial>,
priv_dropper: PrivilegeDropper, priv_dropper: PrivilegeDropper,
server_start_instant: ServerStartInstant,
) { ) {
let config = Rc::new(config); let config = Rc::new(config);
let access_list = state.access_list; let access_list = state.access_list;
@ -86,7 +87,13 @@ pub async fn run_socket_worker(
let (_, mut out_message_receivers) = let (_, mut out_message_receivers) =
out_message_mesh_builder.join(Role::Consumer).await.unwrap(); out_message_mesh_builder.join(Role::Consumer).await.unwrap();
let out_message_consumer_id = ConsumerId(out_message_receivers.consumer_id().unwrap()); let out_message_consumer_id = ConsumerId(
out_message_receivers
.consumer_id()
.unwrap()
.try_into()
.unwrap(),
);
::log::info!("joined channels"); ::log::info!("joined channels");
@ -98,6 +105,7 @@ pub async fn run_socket_worker(
clean_connections( clean_connections(
config.clone(), config.clone(),
connection_slab.clone(), connection_slab.clone(),
server_start_instant,
) )
}), }),
tq_prioritized, tq_prioritized,
@ -133,13 +141,16 @@ pub async fn run_socket_worker(
let key = RefCell::borrow_mut(&connection_slab).insert(ConnectionReference { let key = RefCell::borrow_mut(&connection_slab).insert(ConnectionReference {
task_handle: None, task_handle: None,
out_message_sender: out_message_sender.clone(), out_message_sender: out_message_sender.clone(),
valid_until: ValidUntil::new(config.cleaning.max_connection_idle), valid_until: ValidUntil::new(
server_start_instant,
config.cleaning.max_connection_idle,
),
peer_id: None, peer_id: None,
announced_info_hashes: Default::default(), announced_info_hashes: Default::default(),
ip_version, ip_version,
}); });
::log::info!("accepting stream, assigning id {}", key); ::log::trace!("accepting stream, assigning id {}", key);
let task_handle = spawn_local_into(enclose!((config, access_list, control_message_senders, in_message_senders, connection_slab, opt_tls_config) async move { let task_handle = spawn_local_into(enclose!((config, access_list, control_message_senders, in_message_senders, connection_slab, opt_tls_config) async move {
if let Err(err) = run_connection( if let Err(err) = run_connection(
@ -151,6 +162,7 @@ pub async fn run_socket_worker(
connection_slab.clone(), connection_slab.clone(),
out_message_sender, out_message_sender,
out_message_receiver, out_message_receiver,
server_start_instant,
out_message_consumer_id, out_message_consumer_id,
ConnectionId(key), ConnectionId(key),
opt_tls_config, opt_tls_config,
@ -208,11 +220,12 @@ pub async fn run_socket_worker(
async fn clean_connections( async fn clean_connections(
config: Rc<Config>, config: Rc<Config>,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
server_start_instant: ServerStartInstant,
) -> Option<Duration> { ) -> Option<Duration> {
let now = Instant::now(); let now = server_start_instant.seconds_elapsed();
connection_slab.borrow_mut().retain(|_, reference| { connection_slab.borrow_mut().retain(|_, reference| {
if reference.valid_until.0 > now { if reference.valid_until.valid(now) {
true true
} else { } else {
if let Some(ref handle) = reference.task_handle { if let Some(ref handle) = reference.task_handle {
@ -231,14 +244,14 @@ async fn clean_connections(
} }
async fn receive_out_messages( async fn receive_out_messages(
mut out_message_receiver: ConnectedReceiver<(ConnectionMeta, OutMessage)>, mut out_message_receiver: ConnectedReceiver<(OutMessageMeta, OutMessage)>,
connection_references: Rc<RefCell<Slab<ConnectionReference>>>, connection_references: Rc<RefCell<Slab<ConnectionReference>>>,
) { ) {
let connection_references = &connection_references; let connection_references = &connection_references;
while let Some((meta, out_message)) = out_message_receiver.next().await { while let Some((meta, out_message)) = out_message_receiver.next().await {
if let Some(reference) = connection_references.borrow().get(meta.connection_id.0) { if let Some(reference) = connection_references.borrow().get(meta.connection_id.0) {
::log::info!( ::log::trace!(
"local channel {} len: {}", "local channel {} len: {}",
meta.connection_id.0, meta.connection_id.0,
reference.out_message_sender.len() reference.out_message_sender.len()
@ -249,7 +262,7 @@ async fn receive_out_messages(
Err(GlommioError::Closed(_)) => {} Err(GlommioError::Closed(_)) => {}
Err(GlommioError::WouldBlock(_)) => {} Err(GlommioError::WouldBlock(_)) => {}
Err(err) => { Err(err) => {
::log::info!( ::log::debug!(
"Couldn't send out_message from shared channel to local receiver: {:?}", "Couldn't send out_message from shared channel to local receiver: {:?}",
err err
); );
@ -262,12 +275,13 @@ async fn receive_out_messages(
async fn run_connection( async fn run_connection(
config: Rc<Config>, config: Rc<Config>,
access_list: Arc<AccessListArcSwap>, access_list: Arc<AccessListArcSwap>,
in_message_senders: Rc<Senders<(ConnectionMeta, InMessage)>>, in_message_senders: Rc<Senders<(InMessageMeta, InMessage)>>,
tq_prioritized: TaskQueueHandle, tq_prioritized: TaskQueueHandle,
tq_regular: TaskQueueHandle, tq_regular: TaskQueueHandle,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
out_message_sender: Rc<LocalSender<(ConnectionMeta, OutMessage)>>, out_message_sender: Rc<LocalSender<(OutMessageMeta, OutMessage)>>,
out_message_receiver: LocalReceiver<(ConnectionMeta, OutMessage)>, out_message_receiver: LocalReceiver<(OutMessageMeta, OutMessage)>,
server_start_instant: ServerStartInstant,
out_message_consumer_id: ConsumerId, out_message_consumer_id: ConsumerId,
connection_id: ConnectionId, connection_id: ConnectionId,
opt_tls_config: Option<Arc<RustlsConfig>>, opt_tls_config: Option<Arc<RustlsConfig>>,
@ -288,6 +302,7 @@ async fn run_connection(
connection_slab.clone(), connection_slab.clone(),
out_message_sender, out_message_sender,
out_message_receiver, out_message_receiver,
server_start_instant,
out_message_consumer_id, out_message_consumer_id,
connection_id, connection_id,
stream, stream,
@ -333,6 +348,7 @@ async fn run_connection(
connection_slab.clone(), connection_slab.clone(),
out_message_sender, out_message_sender,
out_message_receiver, out_message_receiver,
server_start_instant,
out_message_consumer_id, out_message_consumer_id,
connection_id, connection_id,
stream, stream,
@ -347,12 +363,13 @@ async fn run_stream_agnostic_connection<
>( >(
config: Rc<Config>, config: Rc<Config>,
access_list: Arc<AccessListArcSwap>, access_list: Arc<AccessListArcSwap>,
in_message_senders: Rc<Senders<(ConnectionMeta, InMessage)>>, in_message_senders: Rc<Senders<(InMessageMeta, InMessage)>>,
tq_prioritized: TaskQueueHandle, tq_prioritized: TaskQueueHandle,
tq_regular: TaskQueueHandle, tq_regular: TaskQueueHandle,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
out_message_sender: Rc<LocalSender<(ConnectionMeta, OutMessage)>>, out_message_sender: Rc<LocalSender<(OutMessageMeta, OutMessage)>>,
out_message_receiver: LocalReceiver<(ConnectionMeta, OutMessage)>, out_message_receiver: LocalReceiver<(OutMessageMeta, OutMessage)>,
server_start_instant: ServerStartInstant,
out_message_consumer_id: ConsumerId, out_message_consumer_id: ConsumerId,
connection_id: ConnectionId, connection_id: ConnectionId,
stream: S, stream: S,
@ -404,6 +421,7 @@ async fn run_stream_agnostic_connection<
ws_out, ws_out,
pending_scrape_slab, pending_scrape_slab,
connection_id, connection_id,
server_start_instant,
}; };
let result = writer.run_out_message_loop().await; let result = writer.run_out_message_loop().await;
@ -422,8 +440,8 @@ struct ConnectionReader<S> {
config: Rc<Config>, config: Rc<Config>,
access_list_cache: AccessListCache, access_list_cache: AccessListCache,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
in_message_senders: Rc<Senders<(ConnectionMeta, InMessage)>>, in_message_senders: Rc<Senders<(InMessageMeta, InMessage)>>,
out_message_sender: Rc<LocalSender<(ConnectionMeta, OutMessage)>>, out_message_sender: Rc<LocalSender<(OutMessageMeta, OutMessage)>>,
pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>, pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>,
out_message_consumer_id: ConsumerId, out_message_consumer_id: ConsumerId,
ws_in: SplitStream<WebSocketStream<S>>, ws_in: SplitStream<WebSocketStream<S>>,
@ -434,8 +452,6 @@ struct ConnectionReader<S> {
impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> { impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
async fn run_in_message_loop(&mut self) -> anyhow::Result<()> { async fn run_in_message_loop(&mut self) -> anyhow::Result<()> {
loop { loop {
::log::debug!("read_in_message");
while self.out_message_sender.is_full() { while self.out_message_sender.is_full() {
sleep(Duration::from_millis(100)).await; sleep(Duration::from_millis(100)).await;
@ -446,8 +462,6 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
match InMessage::from_ws_message(message) { match InMessage::from_ws_message(message) {
Ok(in_message) => { Ok(in_message) => {
::log::debug!("parsed in_message");
self.handle_in_message(in_message).await?; self.handle_in_message(in_message).await?;
} }
Err(err) => { Err(err) => {
@ -520,7 +534,6 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
) )
.await .await
.unwrap(); .unwrap();
::log::info!("sent message to swarm worker");
} else { } else {
self.send_error_response( self.send_error_response(
"Info hash not allowed".into(), "Info hash not allowed".into(),
@ -563,11 +576,14 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
stats: Default::default(), stats: Default::default(),
}; };
let pending_scrape_id = PendingScrapeId( let pending_scrape_id: u8 = self
RefCell::borrow_mut(&mut self.pending_scrape_slab) .pending_scrape_slab
.insert(pending_scrape_response), .borrow_mut()
); .insert(pending_scrape_response)
let meta = self.make_connection_meta(Some(pending_scrape_id)); .try_into()
.with_context(|| "Reached 256 pending scrape responses")?;
let meta = self.make_connection_meta(Some(PendingScrapeId(pending_scrape_id)));
for (consumer_index, info_hashes) in info_hashes_by_worker { for (consumer_index, info_hashes) in info_hashes_by_worker {
let in_message = InMessage::ScrapeRequest(ScrapeRequest { let in_message = InMessage::ScrapeRequest(ScrapeRequest {
@ -580,7 +596,6 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
.send_to(consumer_index, (meta, in_message)) .send_to(consumer_index, (meta, in_message))
.await .await
.unwrap(); .unwrap();
::log::info!("sent message to swarm worker");
} }
} }
} }
@ -601,13 +616,13 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
}); });
self.out_message_sender self.out_message_sender
.send((self.make_connection_meta(None), out_message)) .send((self.make_connection_meta(None).into(), out_message))
.await .await
.map_err(|err| anyhow::anyhow!("ConnectionReader::send_error_response failed: {}", err)) .map_err(|err| anyhow::anyhow!("ConnectionReader::send_error_response failed: {}", err))
} }
fn make_connection_meta(&self, pending_scrape_id: Option<PendingScrapeId>) -> ConnectionMeta { fn make_connection_meta(&self, pending_scrape_id: Option<PendingScrapeId>) -> InMessageMeta {
ConnectionMeta { InMessageMeta {
connection_id: self.connection_id, connection_id: self.connection_id,
out_message_consumer_id: self.out_message_consumer_id, out_message_consumer_id: self.out_message_consumer_id,
ip_version: self.ip_version, ip_version: self.ip_version,
@ -618,10 +633,11 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionReader<S> {
struct ConnectionWriter<S> { struct ConnectionWriter<S> {
config: Rc<Config>, config: Rc<Config>,
out_message_receiver: LocalReceiver<(ConnectionMeta, OutMessage)>, out_message_receiver: LocalReceiver<(OutMessageMeta, OutMessage)>,
connection_slab: Rc<RefCell<Slab<ConnectionReference>>>, connection_slab: Rc<RefCell<Slab<ConnectionReference>>>,
ws_out: SplitSink<WebSocketStream<S>, tungstenite::Message>, ws_out: SplitSink<WebSocketStream<S>, tungstenite::Message>,
pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>, pending_scrape_slab: Rc<RefCell<Slab<PendingScrapeResponse>>>,
server_start_instant: ServerStartInstant,
connection_id: ConnectionId, connection_id: ConnectionId,
} }
@ -640,7 +656,7 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionWriter<S> {
let finished = if let Some(pending) = Slab::get_mut( let finished = if let Some(pending) = Slab::get_mut(
&mut RefCell::borrow_mut(&self.pending_scrape_slab), &mut RefCell::borrow_mut(&self.pending_scrape_slab),
pending_scrape_id.0, pending_scrape_id.0 as usize,
) { ) {
pending.stats.extend(out_message.files); pending.stats.extend(out_message.files);
pending.pending_worker_out_messages -= 1; pending.pending_worker_out_messages -= 1;
@ -654,7 +670,7 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionWriter<S> {
let out_message = { let out_message = {
let mut slab = RefCell::borrow_mut(&self.pending_scrape_slab); let mut slab = RefCell::borrow_mut(&self.pending_scrape_slab);
let pending = slab.remove(pending_scrape_id.0); let pending = slab.remove(pending_scrape_id.0 as usize);
slab.shrink_to_fit(); slab.shrink_to_fit();
@ -694,13 +710,16 @@ impl<S: futures::AsyncRead + futures::AsyncWrite + Unpin> ConnectionWriter<S> {
self.connection_id.0 self.connection_id.0
) )
})? })?
.valid_until = ValidUntil::new(self.config.cleaning.max_connection_idle); .valid_until = ValidUntil::new(
self.server_start_instant,
self.config.cleaning.max_connection_idle,
);
Ok(()) Ok(())
} }
Ok(Err(err)) => Err(err.into()), Ok(Err(err)) => Err(err.into()),
Err(err) => { Err(err) => {
::log::info!("send_out_message: sending to peer took to long: {}", err); ::log::debug!("send_out_message: sending to peer took to long: {}", err);
Ok(()) Ok(())
} }

View file

@ -1,7 +1,7 @@
use std::cell::RefCell; use std::cell::RefCell;
use std::rc::Rc; use std::rc::Rc;
use std::sync::Arc; use std::sync::Arc;
use std::time::{Duration, Instant}; use std::time::Duration;
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache}; use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap, AccessListCache};
use futures::StreamExt; use futures::StreamExt;
@ -12,7 +12,10 @@ use glommio::timer::TimerActionRepeat;
use hashbrown::HashMap; use hashbrown::HashMap;
use rand::{rngs::SmallRng, SeedableRng}; use rand::{rngs::SmallRng, SeedableRng};
use aquatic_common::{extract_response_peers, AmortizedIndexMap, PanicSentinel}; use aquatic_common::{
extract_response_peers, AmortizedIndexMap, PanicSentinel, SecondsSinceServerStart,
ServerStartInstant,
};
use aquatic_ws_protocol::*; use aquatic_ws_protocol::*;
use crate::common::*; use crate::common::*;
@ -44,8 +47,9 @@ impl PeerStatus {
#[derive(Clone, Copy)] #[derive(Clone, Copy)]
struct Peer { struct Peer {
pub connection_meta: ConnectionMeta, pub consumer_id: ConsumerId,
pub status: PeerStatus, pub connection_id: ConnectionId,
pub seeder: bool,
pub valid_until: ValidUntil, pub valid_until: ValidUntil,
} }
@ -71,14 +75,10 @@ impl Default for TorrentData {
impl TorrentData { impl TorrentData {
pub fn remove_peer(&mut self, peer_id: PeerId) { pub fn remove_peer(&mut self, peer_id: PeerId) {
if let Some(peer) = self.peers.remove(&peer_id) { if let Some(peer) = self.peers.remove(&peer_id) {
match peer.status { if peer.seeder {
PeerStatus::Leeching => { self.num_seeders -= 1;
self.num_leechers -= 1; } else {
} self.num_leechers -= 1;
PeerStatus::Seeding => {
self.num_seeders -= 1;
}
PeerStatus::Stopped => (),
} }
} }
} }
@ -93,20 +93,25 @@ struct TorrentMaps {
} }
impl TorrentMaps { impl TorrentMaps {
fn clean(&mut self, config: &Config, access_list: &Arc<AccessListArcSwap>) { fn clean(
&mut self,
config: &Config,
access_list: &Arc<AccessListArcSwap>,
server_start_instant: ServerStartInstant,
) {
let mut access_list_cache = create_access_list_cache(access_list); let mut access_list_cache = create_access_list_cache(access_list);
let now = server_start_instant.seconds_elapsed();
Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv4); Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv4, now);
Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv6); Self::clean_torrent_map(config, &mut access_list_cache, &mut self.ipv6, now);
} }
fn clean_torrent_map( fn clean_torrent_map(
config: &Config, config: &Config,
access_list_cache: &mut AccessListCache, access_list_cache: &mut AccessListCache,
torrent_map: &mut TorrentMap, torrent_map: &mut TorrentMap,
now: SecondsSinceServerStart,
) { ) {
let now = Instant::now();
torrent_map.retain(|info_hash, torrent_data| { torrent_map.retain(|info_hash, torrent_data| {
if !access_list_cache if !access_list_cache
.load() .load()
@ -119,18 +124,14 @@ impl TorrentMaps {
let num_leechers = &mut torrent_data.num_leechers; let num_leechers = &mut torrent_data.num_leechers;
torrent_data.peers.retain(|_, peer| { torrent_data.peers.retain(|_, peer| {
let keep = peer.valid_until.0 >= now; let keep = peer.valid_until.valid(now);
if !keep { if !keep {
match peer.status { if peer.seeder {
PeerStatus::Seeding => { *num_seeders -= 1;
*num_seeders -= 1; } else {
} *num_leechers -= 1;
PeerStatus::Leeching => { }
*num_leechers -= 1;
}
_ => (),
};
} }
keep keep
@ -148,8 +149,9 @@ pub async fn run_swarm_worker(
config: Config, config: Config,
state: State, state: State,
control_message_mesh_builder: MeshBuilder<SwarmControlMessage, Partial>, control_message_mesh_builder: MeshBuilder<SwarmControlMessage, Partial>,
in_message_mesh_builder: MeshBuilder<(ConnectionMeta, InMessage), Partial>, in_message_mesh_builder: MeshBuilder<(InMessageMeta, InMessage), Partial>,
out_message_mesh_builder: MeshBuilder<(ConnectionMeta, OutMessage), Partial>, out_message_mesh_builder: MeshBuilder<(OutMessageMeta, OutMessage), Partial>,
server_start_instant: ServerStartInstant,
) { ) {
let (_, mut control_message_receivers) = control_message_mesh_builder let (_, mut control_message_receivers) = control_message_mesh_builder
.join(Role::Consumer) .join(Role::Consumer)
@ -167,7 +169,7 @@ pub async fn run_swarm_worker(
// Periodically clean torrents // Periodically clean torrents
TimerActionRepeat::repeat(enclose!((config, torrents, access_list) move || { TimerActionRepeat::repeat(enclose!((config, torrents, access_list) move || {
enclose!((config, torrents, access_list) move || async move { enclose!((config, torrents, access_list) move || async move {
torrents.borrow_mut().clean(&config, &access_list); torrents.borrow_mut().clean(&config, &access_list, server_start_instant);
Some(Duration::from_secs(config.cleaning.torrent_cleaning_interval)) Some(Duration::from_secs(config.cleaning.torrent_cleaning_interval))
})() })()
@ -186,6 +188,7 @@ pub async fn run_swarm_worker(
let handle = spawn_local(handle_request_stream( let handle = spawn_local(handle_request_stream(
config.clone(), config.clone(),
torrents.clone(), torrents.clone(),
server_start_instant,
out_message_senders.clone(), out_message_senders.clone(),
receiver, receiver,
)) ))
@ -229,19 +232,23 @@ where
async fn handle_request_stream<S>( async fn handle_request_stream<S>(
config: Config, config: Config,
torrents: Rc<RefCell<TorrentMaps>>, torrents: Rc<RefCell<TorrentMaps>>,
out_message_senders: Rc<Senders<(ConnectionMeta, OutMessage)>>, server_start_instant: ServerStartInstant,
out_message_senders: Rc<Senders<(OutMessageMeta, OutMessage)>>,
stream: S, stream: S,
) where ) where
S: futures_lite::Stream<Item = (ConnectionMeta, InMessage)> + ::std::marker::Unpin, S: futures_lite::Stream<Item = (InMessageMeta, InMessage)> + ::std::marker::Unpin,
{ {
let rng = Rc::new(RefCell::new(SmallRng::from_entropy())); let rng = Rc::new(RefCell::new(SmallRng::from_entropy()));
let max_peer_age = config.cleaning.max_peer_age; let max_peer_age = config.cleaning.max_peer_age;
let peer_valid_until = Rc::new(RefCell::new(ValidUntil::new(max_peer_age))); let peer_valid_until = Rc::new(RefCell::new(ValidUntil::new(
server_start_instant,
max_peer_age,
)));
TimerActionRepeat::repeat(enclose!((peer_valid_until) move || { TimerActionRepeat::repeat(enclose!((peer_valid_until) move || {
enclose!((peer_valid_until) move || async move { enclose!((peer_valid_until) move || async move {
*peer_valid_until.borrow_mut() = ValidUntil::new(max_peer_age); *peer_valid_until.borrow_mut() = ValidUntil::new(server_start_instant, max_peer_age);
Some(Duration::from_secs(1)) Some(Duration::from_secs(1))
})() })()
@ -279,14 +286,12 @@ async fn handle_request_stream<S>(
}; };
for (meta, out_message) in out_messages.drain(..) { for (meta, out_message) in out_messages.drain(..) {
::log::info!("swarm worker trying to send OutMessage to socket worker");
out_message_senders out_message_senders
.send_to(meta.out_message_consumer_id.0, (meta, out_message)) .send_to(meta.out_message_consumer_id.0 as usize, (meta, out_message))
.await .await
.expect("failed sending out_message to socket worker"); .expect("failed sending out_message to socket worker");
::log::info!("swarm worker sent OutMessage to socket worker"); ::log::debug!("swarm worker sent OutMessage to socket worker");
} }
}, },
) )
@ -297,9 +302,9 @@ fn handle_announce_request(
config: &Config, config: &Config,
rng: &mut SmallRng, rng: &mut SmallRng,
torrent_maps: &mut TorrentMaps, torrent_maps: &mut TorrentMaps,
out_messages: &mut Vec<(ConnectionMeta, OutMessage)>, out_messages: &mut Vec<(OutMessageMeta, OutMessage)>,
valid_until: ValidUntil, valid_until: ValidUntil,
request_sender_meta: ConnectionMeta, request_sender_meta: InMessageMeta,
request: AnnounceRequest, request: AnnounceRequest,
) { ) {
let torrent_data: &mut TorrentData = if let IpVersion::V4 = request_sender_meta.ip_version { let torrent_data: &mut TorrentData = if let IpVersion::V4 = request_sender_meta.ip_version {
@ -313,7 +318,7 @@ fn handle_announce_request(
// peers have access to each others peer_id's, they could send requests // peers have access to each others peer_id's, they could send requests
// using them, causing all sorts of issues. // using them, causing all sorts of issues.
if let Some(previous_peer) = torrent_data.peers.get(&request.peer_id) { if let Some(previous_peer) = torrent_data.peers.get(&request.peer_id) {
if request_sender_meta.connection_id != previous_peer.connection_meta.connection_id { if request_sender_meta.connection_id != previous_peer.connection_id {
return; return;
} }
} }
@ -327,31 +332,39 @@ fn handle_announce_request(
request.bytes_left, request.bytes_left,
); );
let peer = Peer {
connection_meta: request_sender_meta,
status: peer_status,
valid_until,
};
let opt_removed_peer = match peer_status { let opt_removed_peer = match peer_status {
PeerStatus::Leeching => { PeerStatus::Leeching => {
torrent_data.num_leechers += 1; torrent_data.num_leechers += 1;
let peer = Peer {
connection_id: request_sender_meta.connection_id,
consumer_id: request_sender_meta.out_message_consumer_id,
seeder: false,
valid_until,
};
torrent_data.peers.insert(request.peer_id, peer) torrent_data.peers.insert(request.peer_id, peer)
} }
PeerStatus::Seeding => { PeerStatus::Seeding => {
torrent_data.num_seeders += 1; torrent_data.num_seeders += 1;
let peer = Peer {
connection_id: request_sender_meta.connection_id,
consumer_id: request_sender_meta.out_message_consumer_id,
seeder: true,
valid_until,
};
torrent_data.peers.insert(request.peer_id, peer) torrent_data.peers.insert(request.peer_id, peer)
} }
PeerStatus::Stopped => torrent_data.peers.remove(&request.peer_id), PeerStatus::Stopped => torrent_data.peers.remove(&request.peer_id),
}; };
match opt_removed_peer.map(|peer| peer.status) { match opt_removed_peer.map(|peer| peer.seeder) {
Some(PeerStatus::Leeching) => { Some(false) => {
torrent_data.num_leechers -= 1; torrent_data.num_leechers -= 1;
} }
Some(PeerStatus::Seeding) => { Some(true) => {
torrent_data.num_seeders -= 1; torrent_data.num_seeders -= 1;
} }
_ => {} _ => {}
@ -385,14 +398,14 @@ fn handle_announce_request(
offer_id: offer.offer_id, offer_id: offer.offer_id,
}; };
out_messages.push(( let meta = OutMessageMeta {
offer_receiver.connection_meta, out_message_consumer_id: offer_receiver.consumer_id,
OutMessage::Offer(middleman_offer), connection_id: offer_receiver.connection_id,
)); pending_scrape_id: None,
::log::trace!( };
"sending middleman offer to {:?}",
offer_receiver.connection_meta out_messages.push((meta, OutMessage::Offer(middleman_offer)));
); ::log::trace!("sending middleman offer to {:?}", meta);
} }
} }
@ -409,14 +422,14 @@ fn handle_announce_request(
offer_id, offer_id,
}; };
out_messages.push(( let meta = OutMessageMeta {
answer_receiver.connection_meta, out_message_consumer_id: answer_receiver.consumer_id,
OutMessage::Answer(middleman_answer), connection_id: answer_receiver.connection_id,
)); pending_scrape_id: None,
::log::trace!( };
"sending middleman answer to {:?}",
answer_receiver.connection_meta out_messages.push((meta, OutMessage::Answer(middleman_answer)));
); ::log::trace!("sending middleman answer to {:?}", meta);
} }
} }
@ -428,14 +441,14 @@ fn handle_announce_request(
announce_interval: config.protocol.peer_announce_interval, announce_interval: config.protocol.peer_announce_interval,
}); });
out_messages.push((request_sender_meta, out_message)); out_messages.push((request_sender_meta.into(), out_message));
} }
fn handle_scrape_request( fn handle_scrape_request(
config: &Config, config: &Config,
torrent_maps: &mut TorrentMaps, torrent_maps: &mut TorrentMaps,
out_messages: &mut Vec<(ConnectionMeta, OutMessage)>, out_messages: &mut Vec<(OutMessageMeta, OutMessage)>,
meta: ConnectionMeta, meta: InMessageMeta,
request: ScrapeRequest, request: ScrapeRequest,
) { ) {
let info_hashes = if let Some(info_hashes) = request.info_hashes { let info_hashes = if let Some(info_hashes) = request.info_hashes {
@ -469,5 +482,5 @@ fn handle_scrape_request(
} }
} }
out_messages.push((meta, OutMessage::ScrapeResponse(out_message))); out_messages.push((meta.into(), OutMessage::ScrapeResponse(out_message)));
} }