mirror of
https://github.com/YGGverse/aquatic.git
synced 2026-03-31 17:55:36 +00:00
udp: implement torrent map cleaning for new, sharded torrent state
This commit is contained in:
parent
54149ed3eb
commit
99632d4be5
8 changed files with 38 additions and 107 deletions
1
Cargo.lock
generated
1
Cargo.lock
generated
|
|
@ -184,7 +184,6 @@ dependencies = [
|
||||||
"cfg-if",
|
"cfg-if",
|
||||||
"crossbeam-channel",
|
"crossbeam-channel",
|
||||||
"hex",
|
"hex",
|
||||||
"histogram",
|
|
||||||
"io-uring",
|
"io-uring",
|
||||||
"libc",
|
"libc",
|
||||||
"log",
|
"log",
|
||||||
|
|
|
||||||
1
TODO.md
1
TODO.md
|
|
@ -22,7 +22,6 @@
|
||||||
with 6 socket and 4 request workers. performance is great overall and faster than
|
with 6 socket and 4 request workers. performance is great overall and faster than
|
||||||
without sharding. io_uring impl is a lot behind mio impl with new load tester
|
without sharding. io_uring impl is a lot behind mio impl with new load tester
|
||||||
* what poll event capacity is actually needed?
|
* what poll event capacity is actually needed?
|
||||||
* clean torrent map in workers, remove it from shared state
|
|
||||||
* mio
|
* mio
|
||||||
* stagger connection cleaning intervals?
|
* stagger connection cleaning intervals?
|
||||||
* uring
|
* uring
|
||||||
|
|
|
||||||
|
|
@ -28,7 +28,6 @@ aquatic_udp_protocol = "0.1.0"
|
||||||
cfg-if = "1"
|
cfg-if = "1"
|
||||||
crossbeam-channel = "0.5"
|
crossbeam-channel = "0.5"
|
||||||
hex = "0.4"
|
hex = "0.4"
|
||||||
histogram = "0.6"
|
|
||||||
log = "0.4"
|
log = "0.4"
|
||||||
mimalloc = { version = "0.1", default-features = false }
|
mimalloc = { version = "0.1", default-features = false }
|
||||||
parking_lot = "0.11"
|
parking_lot = "0.11"
|
||||||
|
|
|
||||||
|
|
@ -6,7 +6,6 @@ use std::sync::Arc;
|
||||||
use std::time::Instant;
|
use std::time::Instant;
|
||||||
|
|
||||||
use crossbeam_channel::Sender;
|
use crossbeam_channel::Sender;
|
||||||
use parking_lot::Mutex;
|
|
||||||
use socket2::{Domain, Protocol, Socket, Type};
|
use socket2::{Domain, Protocol, Socket, Type};
|
||||||
|
|
||||||
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap};
|
use aquatic_common::access_list::{create_access_list_cache, AccessListArcSwap};
|
||||||
|
|
@ -304,7 +303,6 @@ pub struct Statistics {
|
||||||
#[derive(Clone)]
|
#[derive(Clone)]
|
||||||
pub struct State {
|
pub struct State {
|
||||||
pub access_list: Arc<AccessListArcSwap>,
|
pub access_list: Arc<AccessListArcSwap>,
|
||||||
pub torrents: Arc<Mutex<TorrentMaps>>,
|
|
||||||
pub statistics: Arc<Statistics>,
|
pub statistics: Arc<Statistics>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -312,7 +310,6 @@ impl Default for State {
|
||||||
fn default() -> Self {
|
fn default() -> Self {
|
||||||
Self {
|
Self {
|
||||||
access_list: Arc::new(AccessListArcSwap::default()),
|
access_list: Arc::new(AccessListArcSwap::default()),
|
||||||
torrents: Arc::new(Mutex::new(TorrentMaps::default())),
|
|
||||||
statistics: Arc::new(Statistics::default()),
|
statistics: Arc::new(Statistics::default()),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -2,6 +2,7 @@ use std::collections::BTreeMap;
|
||||||
use std::net::IpAddr;
|
use std::net::IpAddr;
|
||||||
use std::net::SocketAddr;
|
use std::net::SocketAddr;
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
|
use std::time::Instant;
|
||||||
|
|
||||||
use aquatic_common::ValidUntil;
|
use aquatic_common::ValidUntil;
|
||||||
use crossbeam_channel::Receiver;
|
use crossbeam_channel::Receiver;
|
||||||
|
|
@ -16,6 +17,7 @@ use crate::config::Config;
|
||||||
|
|
||||||
pub fn run_request_worker(
|
pub fn run_request_worker(
|
||||||
config: Config,
|
config: Config,
|
||||||
|
state: State,
|
||||||
request_receiver: Receiver<(SocketWorkerIndex, ConnectedRequest, SocketAddr)>,
|
request_receiver: Receiver<(SocketWorkerIndex, ConnectedRequest, SocketAddr)>,
|
||||||
response_sender: ConnectedResponseSender,
|
response_sender: ConnectedResponseSender,
|
||||||
) {
|
) {
|
||||||
|
|
@ -23,11 +25,15 @@ pub fn run_request_worker(
|
||||||
let mut small_rng = SmallRng::from_entropy();
|
let mut small_rng = SmallRng::from_entropy();
|
||||||
|
|
||||||
let timeout = Duration::from_millis(config.handlers.channel_recv_timeout_ms);
|
let timeout = Duration::from_millis(config.handlers.channel_recv_timeout_ms);
|
||||||
|
let mut peer_valid_until = ValidUntil::new(config.cleaning.max_peer_age);
|
||||||
|
|
||||||
|
let cleaning_interval = Duration::from_secs(config.cleaning.torrent_cleaning_interval);
|
||||||
|
|
||||||
|
let mut iter_counter = 0usize;
|
||||||
|
let mut last_cleaning = Instant::now();
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
if let Ok((sender_index, request, src)) = request_receiver.recv_timeout(timeout) {
|
if let Ok((sender_index, request, src)) = request_receiver.recv_timeout(timeout) {
|
||||||
let peer_valid_until = ValidUntil::new(config.cleaning.max_peer_age);
|
|
||||||
|
|
||||||
let response = match request {
|
let response = match request {
|
||||||
ConnectedRequest::Announce(request) => handle_announce_request(
|
ConnectedRequest::Announce(request) => handle_announce_request(
|
||||||
&config,
|
&config,
|
||||||
|
|
@ -45,7 +51,19 @@ pub fn run_request_worker(
|
||||||
response_sender.try_send_to(sender_index, response, src);
|
response_sender.try_send_to(sender_index, response, src);
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: clean torrent map, update peer_valid_until
|
if iter_counter % 128 == 0 {
|
||||||
|
peer_valid_until = ValidUntil::new(config.cleaning.max_peer_age);
|
||||||
|
|
||||||
|
let now = Instant::now();
|
||||||
|
|
||||||
|
if now > last_cleaning + cleaning_interval {
|
||||||
|
torrents.clean(&config, &state.access_list);
|
||||||
|
|
||||||
|
last_cleaning = now;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
iter_counter = iter_counter.wrapping_add(1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -35,33 +35,6 @@ pub fn run(config: Config) -> ::anyhow::Result<()> {
|
||||||
|
|
||||||
let mut signals = Signals::new(::std::iter::once(SIGUSR1))?;
|
let mut signals = Signals::new(::std::iter::once(SIGUSR1))?;
|
||||||
|
|
||||||
{
|
|
||||||
let config = config.clone();
|
|
||||||
let state = state.clone();
|
|
||||||
|
|
||||||
::std::thread::spawn(move || run_inner(config, state));
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(feature = "cpu-pinning")]
|
|
||||||
pin_current_if_configured_to(
|
|
||||||
&config.cpu_pinning,
|
|
||||||
config.socket_workers,
|
|
||||||
WorkerIndex::Other,
|
|
||||||
);
|
|
||||||
|
|
||||||
for signal in &mut signals {
|
|
||||||
match signal {
|
|
||||||
SIGUSR1 => {
|
|
||||||
let _ = update_access_list(&config.access_list, &state.access_list);
|
|
||||||
}
|
|
||||||
_ => unreachable!(),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn run_inner(config: Config, state: State) -> ::anyhow::Result<()> {
|
|
||||||
let num_bound_sockets = Arc::new(AtomicUsize::new(0));
|
let num_bound_sockets = Arc::new(AtomicUsize::new(0));
|
||||||
|
|
||||||
let mut request_senders = Vec::new();
|
let mut request_senders = Vec::new();
|
||||||
|
|
@ -86,6 +59,7 @@ pub fn run_inner(config: Config, state: State) -> ::anyhow::Result<()> {
|
||||||
|
|
||||||
for i in 0..config.request_workers {
|
for i in 0..config.request_workers {
|
||||||
let config = config.clone();
|
let config = config.clone();
|
||||||
|
let state = state.clone();
|
||||||
let request_receiver = request_receivers.remove(&i).unwrap().clone();
|
let request_receiver = request_receivers.remove(&i).unwrap().clone();
|
||||||
let response_sender = ConnectedResponseSender::new(response_senders.clone());
|
let response_sender = ConnectedResponseSender::new(response_senders.clone());
|
||||||
|
|
||||||
|
|
@ -99,7 +73,7 @@ pub fn run_inner(config: Config, state: State) -> ::anyhow::Result<()> {
|
||||||
WorkerIndex::RequestWorker(i),
|
WorkerIndex::RequestWorker(i),
|
||||||
);
|
);
|
||||||
|
|
||||||
handlers::run_request_worker(config, request_receiver, response_sender)
|
handlers::run_request_worker(config, state, request_receiver, response_sender)
|
||||||
})
|
})
|
||||||
.with_context(|| "spawn request worker")?;
|
.with_context(|| "spawn request worker")?;
|
||||||
}
|
}
|
||||||
|
|
@ -146,12 +120,6 @@ pub fn run_inner(config: Config, state: State) -> ::anyhow::Result<()> {
|
||||||
.with_context(|| "spawn socket worker")?;
|
.with_context(|| "spawn socket worker")?;
|
||||||
}
|
}
|
||||||
|
|
||||||
::std::mem::drop(request_senders);
|
|
||||||
::std::mem::drop(request_receivers);
|
|
||||||
|
|
||||||
::std::mem::drop(response_senders);
|
|
||||||
::std::mem::drop(response_receivers);
|
|
||||||
|
|
||||||
if config.statistics.interval != 0 {
|
if config.statistics.interval != 0 {
|
||||||
let state = state.clone();
|
let state = state.clone();
|
||||||
let config = config.clone();
|
let config = config.clone();
|
||||||
|
|
@ -189,11 +157,14 @@ pub fn run_inner(config: Config, state: State) -> ::anyhow::Result<()> {
|
||||||
WorkerIndex::Other,
|
WorkerIndex::Other,
|
||||||
);
|
);
|
||||||
|
|
||||||
loop {
|
for signal in &mut signals {
|
||||||
::std::thread::sleep(Duration::from_secs(
|
match signal {
|
||||||
config.cleaning.torrent_cleaning_interval,
|
SIGUSR1 => {
|
||||||
));
|
let _ = update_access_list(&config.access_list, &state.access_list);
|
||||||
|
}
|
||||||
state.torrents.lock().clean(&config, &state.access_list);
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,5 @@
|
||||||
use std::sync::atomic::Ordering;
|
use std::sync::atomic::Ordering;
|
||||||
|
|
||||||
use histogram::Histogram;
|
|
||||||
|
|
||||||
use super::common::*;
|
use super::common::*;
|
||||||
use crate::config::Config;
|
use crate::config::Config;
|
||||||
|
|
||||||
|
|
@ -38,59 +36,5 @@ pub fn gather_and_print_statistics(state: &State, config: &Config) {
|
||||||
bytes_sent_per_second * 8.0 / 1_000_000.0,
|
bytes_sent_per_second * 8.0 / 1_000_000.0,
|
||||||
);
|
);
|
||||||
|
|
||||||
let mut total_num_torrents_ipv4 = 0usize;
|
|
||||||
let mut total_num_torrents_ipv6 = 0usize;
|
|
||||||
let mut total_num_peers_ipv4 = 0usize;
|
|
||||||
let mut total_num_peers_ipv6 = 0usize;
|
|
||||||
|
|
||||||
let mut peers_per_torrent = Histogram::new();
|
|
||||||
|
|
||||||
{
|
|
||||||
let torrents = &mut state.torrents.lock();
|
|
||||||
|
|
||||||
for torrent in torrents.ipv4.values() {
|
|
||||||
let num_peers = torrent.num_seeders + torrent.num_leechers;
|
|
||||||
|
|
||||||
if let Err(err) = peers_per_torrent.increment(num_peers as u64) {
|
|
||||||
::log::error!("error incrementing peers_per_torrent histogram: {}", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
total_num_peers_ipv4 += num_peers;
|
|
||||||
}
|
|
||||||
for torrent in torrents.ipv6.values() {
|
|
||||||
let num_peers = torrent.num_seeders + torrent.num_leechers;
|
|
||||||
|
|
||||||
if let Err(err) = peers_per_torrent.increment(num_peers as u64) {
|
|
||||||
::log::error!("error incrementing peers_per_torrent histogram: {}", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
total_num_peers_ipv6 += num_peers;
|
|
||||||
}
|
|
||||||
|
|
||||||
total_num_torrents_ipv4 += torrents.ipv4.len();
|
|
||||||
total_num_torrents_ipv6 += torrents.ipv6.len();
|
|
||||||
}
|
|
||||||
|
|
||||||
println!(
|
|
||||||
"ipv4 torrents: {}, peers: {}; ipv6 torrents: {}, peers: {}",
|
|
||||||
total_num_torrents_ipv4,
|
|
||||||
total_num_peers_ipv4,
|
|
||||||
total_num_torrents_ipv6,
|
|
||||||
total_num_peers_ipv6,
|
|
||||||
);
|
|
||||||
|
|
||||||
if peers_per_torrent.entries() != 0 {
|
|
||||||
println!(
|
|
||||||
"peers per torrent: min: {}, p50: {}, p75: {}, p90: {}, p99: {}, p999: {}, max: {}",
|
|
||||||
peers_per_torrent.minimum().unwrap(),
|
|
||||||
peers_per_torrent.percentile(50.0).unwrap(),
|
|
||||||
peers_per_torrent.percentile(75.0).unwrap(),
|
|
||||||
peers_per_torrent.percentile(90.0).unwrap(),
|
|
||||||
peers_per_torrent.percentile(99.0).unwrap(),
|
|
||||||
peers_per_torrent.percentile(99.9).unwrap(),
|
|
||||||
peers_per_torrent.maximum().unwrap(),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
println!();
|
println!();
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -39,7 +39,9 @@ fn main() {
|
||||||
pub fn run(bench_config: BenchConfig) -> ::anyhow::Result<()> {
|
pub fn run(bench_config: BenchConfig) -> ::anyhow::Result<()> {
|
||||||
// Setup common state, spawn request handlers
|
// Setup common state, spawn request handlers
|
||||||
|
|
||||||
let aquatic_config = Config::default();
|
let mut aquatic_config = Config::default();
|
||||||
|
|
||||||
|
aquatic_config.cleaning.torrent_cleaning_interval = 60 * 60 * 24;
|
||||||
|
|
||||||
let (request_sender, request_receiver) = unbounded();
|
let (request_sender, request_receiver) = unbounded();
|
||||||
let (response_sender, response_receiver) = unbounded();
|
let (response_sender, response_receiver) = unbounded();
|
||||||
|
|
@ -49,7 +51,9 @@ pub fn run(bench_config: BenchConfig) -> ::anyhow::Result<()> {
|
||||||
{
|
{
|
||||||
let config = aquatic_config.clone();
|
let config = aquatic_config.clone();
|
||||||
|
|
||||||
::std::thread::spawn(move || run_request_worker(config, request_receiver, response_sender));
|
::std::thread::spawn(move || {
|
||||||
|
run_request_worker(config, State::default(), request_receiver, response_sender)
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
// Run benchmarks
|
// Run benchmarks
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue