|
@@ -1,7 +1,8 @@
|
|
|
-use troll_patrol::{request_handler::handle, *};
|
|
|
+use troll_patrol::{request_handler::*, *};
|
|
|
|
|
|
use clap::Parser;
|
|
|
use futures::future;
|
|
|
+use futures::join;
|
|
|
use hyper::{
|
|
|
server::conn::AddrStream,
|
|
|
service::{make_service_fn, service_fn},
|
|
@@ -10,14 +11,20 @@ use hyper::{
|
|
|
use serde::Deserialize;
|
|
|
use sled::Db;
|
|
|
use std::{
|
|
|
- collections::BTreeMap, convert::Infallible, fs::File, io::BufReader, net::SocketAddr,
|
|
|
- path::PathBuf, time::Duration,
|
|
|
+ collections::{BTreeMap, HashMap, HashSet},
|
|
|
+ convert::Infallible,
|
|
|
+ fs::File,
|
|
|
+ io::BufReader,
|
|
|
+ net::SocketAddr,
|
|
|
+ path::PathBuf,
|
|
|
+ time::Duration,
|
|
|
};
|
|
|
use tokio::{
|
|
|
signal, spawn,
|
|
|
sync::{broadcast, mpsc, oneshot},
|
|
|
time::sleep,
|
|
|
};
|
|
|
+#[cfg(not(features = "simulation"))]
|
|
|
use tokio_cron::{Job, Scheduler};
|
|
|
|
|
|
async fn shutdown_signal() {
|
|
@@ -60,6 +67,7 @@ pub struct Config {
|
|
|
|
|
|
//require_bridge_token: bool,
|
|
|
port: u16,
|
|
|
+ updater_port: u16,
|
|
|
updater_schedule: String,
|
|
|
}
|
|
|
|
|
@@ -86,7 +94,7 @@ async fn update_daily_info(
|
|
|
scaling_factor: f64,
|
|
|
min_historical_days: u32,
|
|
|
max_historical_days: u32,
|
|
|
-) {
|
|
|
+) -> HashMap<[u8; 20], HashSet<String>> {
|
|
|
update_extra_infos(&db, &extra_infos_base_url)
|
|
|
.await
|
|
|
.unwrap();
|
|
@@ -99,15 +107,22 @@ async fn update_daily_info(
|
|
|
min_historical_days,
|
|
|
max_historical_days,
|
|
|
);
|
|
|
- report_blockages(&distributors, new_blockages).await;
|
|
|
+ report_blockages(&distributors, new_blockages.clone()).await;
|
|
|
|
|
|
// Generate tomorrow's key if we don't already have it
|
|
|
new_negative_report_key(&db, get_date() + 1);
|
|
|
+
|
|
|
+ // Return new detected blockages
|
|
|
+ new_blockages
|
|
|
}
|
|
|
|
|
|
+/*
|
|
|
async fn run_updater(updater_tx: mpsc::Sender<Command>) {
|
|
|
- updater_tx.send(Command::Update {}).await.unwrap();
|
|
|
+ updater_tx.send(Command::Update {
|
|
|
+
|
|
|
+ }).await.unwrap();
|
|
|
}
|
|
|
+*/
|
|
|
|
|
|
async fn create_context_manager(
|
|
|
db_config: DbConfig,
|
|
@@ -155,8 +170,8 @@ async fn context_manager(
|
|
|
drop(shutdown_sig);
|
|
|
println!("Shutdown Sent.");
|
|
|
}
|
|
|
- Update {} => {
|
|
|
- update_daily_info(
|
|
|
+ Update { _req, sender } => {
|
|
|
+ let blockages = update_daily_info(
|
|
|
&db,
|
|
|
&distributors,
|
|
|
&extra_infos_base_url,
|
|
@@ -167,6 +182,23 @@ async fn context_manager(
|
|
|
max_historical_days,
|
|
|
)
|
|
|
.await;
|
|
|
+ let response = if cfg!(feature = "simulation") {
|
|
|
+ // Convert map keys from [u8; 20] to 40-character hex strings
|
|
|
+ let mut blockages_str = HashMap::<String, HashSet<String>>::new();
|
|
|
+ for (fingerprint, countries) in blockages {
|
|
|
+ let fpr_string = array_bytes::bytes2hex("", fingerprint);
|
|
|
+ blockages_str.insert(fpr_string, countries);
|
|
|
+ }
|
|
|
+ Ok(prepare_header(
|
|
|
+ serde_json::to_string(&blockages_str).unwrap(),
|
|
|
+ ))
|
|
|
+ } else {
|
|
|
+ Ok(prepare_header("OK".to_string()))
|
|
|
+ };
|
|
|
+ if let Err(e) = sender.send(response) {
|
|
|
+ eprintln!("Update Response Error: {:?}", e);
|
|
|
+ };
|
|
|
+ sleep(Duration::from_millis(1)).await;
|
|
|
}
|
|
|
}
|
|
|
}
|
|
@@ -182,7 +214,10 @@ enum Command {
|
|
|
Shutdown {
|
|
|
shutdown_sig: broadcast::Sender<()>,
|
|
|
},
|
|
|
- Update {},
|
|
|
+ Update {
|
|
|
+ _req: Request<Body>,
|
|
|
+ sender: oneshot::Sender<Result<Response<Body>, Infallible>>,
|
|
|
+ },
|
|
|
}
|
|
|
|
|
|
#[tokio::main]
|
|
@@ -218,14 +253,17 @@ async fn main() {
|
|
|
}
|
|
|
});
|
|
|
|
|
|
- let updater = spawn(async move {
|
|
|
- // Run updater once per day
|
|
|
- let mut sched = Scheduler::utc();
|
|
|
- sched.add(Job::new(config.updater_schedule, move || {
|
|
|
- run_updater(updater_tx.clone())
|
|
|
- }));
|
|
|
- });
|
|
|
-
|
|
|
+ // TODO: Reintroduce this
|
|
|
+ /*
|
|
|
+ #[cfg(not(feature = "simulation"))]
|
|
|
+ let updater = spawn(async move {
|
|
|
+ // Run updater once per day
|
|
|
+ let mut sched = Scheduler::utc();
|
|
|
+ sched.add(Job::new(config.updater_schedule, move || {
|
|
|
+ run_updater(updater_tx.clone())
|
|
|
+ }));
|
|
|
+ });
|
|
|
+ */
|
|
|
let context_manager = spawn(async move {
|
|
|
create_context_manager(
|
|
|
config.db,
|
|
@@ -259,12 +297,37 @@ async fn main() {
|
|
|
async move { Ok::<_, Infallible>(service) }
|
|
|
});
|
|
|
|
|
|
+ let updater_make_service = make_service_fn(move |_conn: &AddrStream| {
|
|
|
+ let request_tx = updater_tx.clone();
|
|
|
+ let service = service_fn(move |_req| {
|
|
|
+ let request_tx = request_tx.clone();
|
|
|
+ let (response_tx, response_rx) = oneshot::channel();
|
|
|
+ let cmd = Command::Update {
|
|
|
+ _req,
|
|
|
+ sender: response_tx,
|
|
|
+ };
|
|
|
+ async move {
|
|
|
+ request_tx.send(cmd).await.unwrap();
|
|
|
+ response_rx.await.unwrap()
|
|
|
+ }
|
|
|
+ });
|
|
|
+ async move { Ok::<_, Infallible>(service) }
|
|
|
+ });
|
|
|
+
|
|
|
let addr = SocketAddr::from(([0, 0, 0, 0], config.port));
|
|
|
let server = Server::bind(&addr).serve(make_service);
|
|
|
let graceful = server.with_graceful_shutdown(shutdown_signal());
|
|
|
+ let updater_addr = SocketAddr::from(([127, 0, 0, 1], config.updater_port));
|
|
|
+ let updater_server = Server::bind(&updater_addr).serve(updater_make_service);
|
|
|
+ let updater_graceful = updater_server.with_graceful_shutdown(shutdown_signal());
|
|
|
println!("Listening on {}", addr);
|
|
|
- if let Err(e) = graceful.await {
|
|
|
- eprintln!("server error: {}", e);
|
|
|
+ println!("Updater listening on {}", updater_addr);
|
|
|
+ let (a, b) = join!(graceful, updater_graceful);
|
|
|
+ if a.is_err() {
|
|
|
+ eprintln!("server error: {}", a.unwrap_err());
|
|
|
+ }
|
|
|
+ if b.is_err() {
|
|
|
+ eprintln!("server error: {}", b.unwrap_err());
|
|
|
}
|
|
|
- future::join_all([context_manager, updater, shutdown_handler]).await;
|
|
|
+ future::join_all([context_manager, shutdown_handler]).await;
|
|
|
}
|