mirror of
https://github.com/YGGverse/aquatic.git
synced 2026-03-31 17:55:36 +00:00
aquatic_ws: handler: implement scrape handler; other small changes
This commit is contained in:
parent
1bb9caa8d1
commit
7ebbb311e1
2 changed files with 109 additions and 37 deletions
|
|
@ -1,9 +1,11 @@
|
||||||
use std::net::SocketAddr;
|
use std::net::SocketAddr;
|
||||||
use std::time::Instant;
|
use std::time::Instant;
|
||||||
|
use std::sync::Arc;
|
||||||
|
|
||||||
use flume::{Sender, Receiver};
|
use flume::{Sender, Receiver};
|
||||||
use hashbrown::HashMap;
|
use hashbrown::HashMap;
|
||||||
use indexmap::IndexMap;
|
use indexmap::IndexMap;
|
||||||
|
use parking_lot::Mutex;
|
||||||
|
|
||||||
use crate::protocol::*;
|
use crate::protocol::*;
|
||||||
|
|
||||||
|
|
@ -33,14 +35,14 @@ pub type TorrentMap = HashMap<InfoHash, TorrentData>;
|
||||||
|
|
||||||
|
|
||||||
pub struct State {
|
pub struct State {
|
||||||
pub torrents: TorrentMap,
|
pub torrents: Arc<Mutex<TorrentMap>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
impl Default for State {
|
impl Default for State {
|
||||||
fn default() -> Self {
|
fn default() -> Self {
|
||||||
Self {
|
Self {
|
||||||
torrents: HashMap::new(),
|
torrents: Arc::new(Mutex::new(HashMap::new())),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -1,6 +1,8 @@
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
|
use std::vec::Drain;
|
||||||
|
|
||||||
use hashbrown::HashMap;
|
use hashbrown::HashMap;
|
||||||
|
use parking_lot::MutexGuard;
|
||||||
|
|
||||||
use crate::common::*;
|
use crate::common::*;
|
||||||
use crate::protocol::*;
|
use crate::protocol::*;
|
||||||
|
|
@ -11,49 +13,117 @@ pub fn run_request_worker(
|
||||||
in_message_receiver: InMessageReceiver,
|
in_message_receiver: InMessageReceiver,
|
||||||
out_message_sender: OutMessageSender,
|
out_message_sender: OutMessageSender,
|
||||||
){
|
){
|
||||||
let mut in_messages = Vec::new();
|
|
||||||
let mut out_messages = Vec::new();
|
let mut out_messages = Vec::new();
|
||||||
|
|
||||||
|
let mut announce_requests = Vec::new();
|
||||||
|
let mut scrape_requests = Vec::new();
|
||||||
|
|
||||||
let timeout = Duration::from_micros(200);
|
let timeout = Duration::from_micros(200);
|
||||||
|
|
||||||
for i in 0..1000 {
|
loop {
|
||||||
if i == 0 {
|
let mut opt_torrent_map_guard: Option<MutexGuard<TorrentMap>> = None;
|
||||||
if let Ok((meta, in_message)) = in_message_receiver.recv(){
|
|
||||||
in_messages.push((meta, in_message));
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
let res_in_message = in_message_receiver.recv_timeout(timeout);
|
|
||||||
|
|
||||||
if let Ok((meta, in_message)) = res_in_message {
|
for i in 0..1000 {
|
||||||
in_messages.push((meta, in_message));
|
if i == 0 {
|
||||||
|
if let Ok((meta, in_message)) = in_message_receiver.recv(){
|
||||||
|
match in_message {
|
||||||
|
InMessage::AnnounceRequest(r) => {
|
||||||
|
announce_requests.push((meta, r));
|
||||||
|
},
|
||||||
|
InMessage::ScrapeRequest(r) => {
|
||||||
|
scrape_requests.push((meta, r));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
break
|
let res_in_message = in_message_receiver.recv_timeout(timeout);
|
||||||
}
|
|
||||||
|
if let Ok((meta, in_message)) = res_in_message {
|
||||||
|
match in_message {
|
||||||
|
InMessage::AnnounceRequest(r) => {
|
||||||
|
announce_requests.push((meta, r));
|
||||||
|
},
|
||||||
|
InMessage::ScrapeRequest(r) => {
|
||||||
|
scrape_requests.push((meta, r));
|
||||||
|
},
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if let Some(torrent_guard) = state.torrents.try_lock(){
|
||||||
|
opt_torrent_map_guard = Some(torrent_guard);
|
||||||
|
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
let mut torrent_map_guard = opt_torrent_map_guard
|
||||||
|
.unwrap_or_else(|| state.torrents.lock());
|
||||||
|
|
||||||
|
handle_announce_requests(
|
||||||
|
&mut torrent_map_guard,
|
||||||
|
&mut out_messages,
|
||||||
|
announce_requests.drain(..)
|
||||||
|
);
|
||||||
|
|
||||||
|
handle_scrape_requests(
|
||||||
|
&mut torrent_map_guard,
|
||||||
|
&mut out_messages,
|
||||||
|
scrape_requests.drain(..)
|
||||||
|
);
|
||||||
|
|
||||||
|
::std::mem::drop(torrent_map_guard);
|
||||||
|
|
||||||
|
for (meta, out_message) in out_messages.drain(..){
|
||||||
|
out_message_sender.send(meta, out_message);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
pub fn handle_announce_requests(
|
||||||
|
torrents: &mut TorrentMap,
|
||||||
|
messages_out: &mut Vec<(ConnectionMeta, OutMessage)>,
|
||||||
|
requests: Drain<(ConnectionMeta, AnnounceRequest)>,
|
||||||
|
){
|
||||||
|
// if offers are set, fetch same number of peers, send offers to all of them
|
||||||
|
// if answer is set, fetch that peer, send answer to it
|
||||||
|
// finally, return announce response, I think
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
pub fn handle_scrape_requests(
|
||||||
|
torrents: &mut TorrentMap,
|
||||||
|
messages_out: &mut Vec<(ConnectionMeta, OutMessage)>,
|
||||||
|
requests: Drain<(ConnectionMeta, ScrapeRequest)>,
|
||||||
|
){
|
||||||
|
messages_out.extend(requests.map(|(meta, request)| {
|
||||||
|
let num_info_hashes = request.info_hashes
|
||||||
|
.as_ref()
|
||||||
|
.map(|v| v.len())
|
||||||
|
.unwrap_or(0);
|
||||||
|
|
||||||
|
let mut response = ScrapeResponse {
|
||||||
|
files: HashMap::with_capacity(num_info_hashes),
|
||||||
};
|
};
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: lock torrent mutex
|
// If request.info_hashes is None, don't return scrape for all
|
||||||
|
// torrents, even though that is done in reference server, it is
|
||||||
|
// too expensive.
|
||||||
|
if let Some(info_hashes) = request.info_hashes {
|
||||||
|
for info_hash in info_hashes {
|
||||||
|
if let Some(torrent_data) = torrents.get(&info_hash){
|
||||||
|
let stats = ScrapeStatistics {
|
||||||
|
complete: torrent_data.seeders,
|
||||||
|
downloaded: 0, // No implementation planned
|
||||||
|
incomplete: torrent_data.leechers,
|
||||||
|
};
|
||||||
|
|
||||||
// This should be separate function. Possibly requests should be separated
|
response.files.insert(info_hash, stats);
|
||||||
// above and one announce and one scrape handler used
|
}
|
||||||
for (meta, in_message) in in_messages.drain(..){
|
}
|
||||||
// announce requests:
|
}
|
||||||
// if offers are set, fetch same number of peers, send offers to all of them
|
|
||||||
// if answer is set, fetch that peer, send answer to it
|
|
||||||
// finally, return announce response, seemingly
|
|
||||||
|
|
||||||
// scrape: just fetch stats for all info_hashes sent
|
(meta, OutMessage::ScrapeResponse(response))
|
||||||
|
}));
|
||||||
let out_message = OutMessage::ScrapeResponse(ScrapeResponse {
|
|
||||||
files: HashMap::new(),
|
|
||||||
});
|
|
||||||
|
|
||||||
out_messages.push((meta, out_message));
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: unlock torrent mutex
|
|
||||||
|
|
||||||
for (meta, out_message) in out_messages.drain(..){
|
|
||||||
out_message_sender.send(meta, out_message);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
Loading…
Add table
Add a link
Reference in a new issue