Browse Source

Add functions to verify reports, update db

Vecna 2 months ago
parent
commit
3ef9c04141
3 changed files with 239 additions and 43 deletions
  1. 11 0
      config.json
  2. 8 38
      src/bin/server.rs
  3. 220 5
      src/lib.rs

+ 11 - 0
config.json

@@ -0,0 +1,11 @@
+{
+    "db": {
+    "db_path": "server_db"
+
+    },
+    "distributors": {
+    "Lox": "127.0.0.1:8002"
+    },
+    "port": 8003,
+    "require_bridge_token": false
+}

+ 8 - 38
src/bin/server.rs

@@ -1,10 +1,4 @@
-use troll_patrol::{
-    extra_info::{self, ExtraInfo},
-    //negative_report::SerializableNegativeReport,
-    //positive_report::SerializablePositiveReport,
-    request_handler::handle,
-    *,
-};
+use troll_patrol::{request_handler::handle, *};
 
 use clap::Parser;
 use futures::future;
@@ -16,7 +10,7 @@ use hyper::{
 use serde::Deserialize;
 use sled::Db;
 use std::{
-    collections::HashSet, convert::Infallible, fs::File, io::BufReader, net::SocketAddr,
+    collections::BTreeMap, convert::Infallible, fs::File, io::BufReader, net::SocketAddr,
     path::PathBuf, time::Duration,
 };
 use tokio::{
@@ -43,6 +37,8 @@ struct Args {
 #[derive(Debug, Deserialize)]
 pub struct Config {
     pub db: DbConfig,
+    // map of distributor name to IP:port to contact it
+    pub distributors: BTreeMap<BridgeDistributor, String>,
     //require_bridge_token: bool,
     port: u16,
 }
@@ -61,36 +57,10 @@ impl Default for DbConfig {
     }
 }
 
-async fn update_extra_infos(db: &Db) {
-    // Track which files have been processed. This is slightly redundant
-    // because we're only downloading files we don't already have, but it
-    // might be a good idea to check in case we downloaded a file but didn't
-    // process it for some reason.
-    let mut processed_extra_infos_files = match db.get(b"extra_infos_files").unwrap() {
-        Some(v) => bincode::deserialize(&v).unwrap(),
-        None => HashSet::<String>::new(),
-    };
-
-    let new_files = extra_info::download_extra_infos().await.unwrap();
-
-    let mut new_extra_infos = HashSet::<ExtraInfo>::new();
-
-    // Make set of new extra-infos
-    for extra_info_file in &new_files {
-        extra_info::add_extra_infos(&extra_info_file, &mut new_extra_infos);
-        processed_extra_infos_files.insert(extra_info_file.to_string());
-    }
-
-    // Add new extra-infos data to database
-    for extra_info in new_extra_infos {
-        add_extra_info_to_db(&db, extra_info);
-    }
-
-    db.insert(
-        b"extra_infos_files",
-        bincode::serialize(&processed_extra_infos_files).unwrap(),
-    )
-    .unwrap();
+async fn update_daily_info(db: &Db, distributors: &BTreeMap<BridgeDistributor, String>) {
+    update_extra_infos(&db).await;
+    update_negative_reports(&db, &distributors).await;
+    update_positive_reports(&db, &distributors).await;
 }
 
 async fn create_context_manager(

+ 220 - 5
src/lib.rs

@@ -1,3 +1,4 @@
+use hyper::{Body, Client, Method, Request};
 use lazy_static::lazy_static;
 use serde::{Deserialize, Serialize};
 use sled::Db;
@@ -31,7 +32,7 @@ pub fn get_date() -> u32 {
         .unwrap()
 }
 
-#[derive(Clone, Copy, Eq, PartialEq, Ord, PartialOrd, Serialize, Deserialize)]
+#[derive(Clone, Copy, Debug, Eq, PartialEq, Ord, PartialOrd, Serialize, Deserialize)]
 pub enum BridgeDistributor {
     Lox,
 }
@@ -48,8 +49,8 @@ pub struct BridgeInfo {
     /// first Julian date we started collecting data on this bridge
     pub first_seen: u32,
 
-    /// flag indicating whether the bridge is believed to be blocked
-    pub is_blocked: bool,
+    /// list of countries where the bridge is believed to be blocked
+    pub blocked_in: Vec<String>,
 
     /// map of dates to data for that day
     pub info_by_day: HashMap<u32, DailyBridgeInfo>,
@@ -61,7 +62,7 @@ impl BridgeInfo {
             fingerprint: fingerprint,
             nickname: nickname.to_string(),
             first_seen: get_date(),
-            is_blocked: false,
+            blocked_in: Vec::<String>::new(),
             info_by_day: HashMap::<u32, DailyBridgeInfo>::new(),
         }
     }
@@ -75,7 +76,10 @@ impl fmt::Display for BridgeInfo {
         );
         str.push_str(format!("nickname: {}\n", self.nickname).as_str());
         str.push_str(format!("first_seen: {}\n", self.first_seen).as_str());
-        str.push_str(format!("is_blocked: {}\n", self.is_blocked).as_str());
+        str.push_str("blocked_in:");
+        for country in &self.blocked_in {
+            str.push_str(format!("\n  {}", country).as_str());
+        }
         str.push_str("info_by_day:");
         for day in self.info_by_day.keys() {
             str.push_str(format!("\n  day: {}", day).as_str());
@@ -180,6 +184,8 @@ impl fmt::Display for DailyBridgeInfo {
     }
 }
 
+// Process extra-infos
+
 /// Adds the extra-info data for a single bridge to the database. If the
 /// database already contains an extra-info for this bridge for thid date,
 /// but this extra-info contains different data for some reason, use the
@@ -210,6 +216,41 @@ pub fn add_extra_info_to_db(db: &Db, extra_info: ExtraInfo) {
         .unwrap();
 }
 
+/// Download new extra-infos files and add their data to the database
+pub async fn update_extra_infos(db: &Db) {
+    // Track which files have been processed. This is slightly redundant
+    // because we're only downloading files we don't already have, but it
+    // might be a good idea to check in case we downloaded a file but didn't
+    // process it for some reason.
+    let mut processed_extra_infos_files = match db.get(b"extra_infos_files").unwrap() {
+        Some(v) => bincode::deserialize(&v).unwrap(),
+        None => HashSet::<String>::new(),
+    };
+
+    let new_files = extra_info::download_extra_infos().await.unwrap();
+
+    let mut new_extra_infos = HashSet::<ExtraInfo>::new();
+
+    // Make set of new extra-infos
+    for extra_info_file in &new_files {
+        extra_info::add_extra_infos(&extra_info_file, &mut new_extra_infos);
+        processed_extra_infos_files.insert(extra_info_file.to_string());
+    }
+
+    // Add new extra-infos data to database
+    for extra_info in new_extra_infos {
+        add_extra_info_to_db(&db, extra_info);
+    }
+
+    db.insert(
+        b"extra_infos_files",
+        bincode::serialize(&processed_extra_infos_files).unwrap(),
+    )
+    .unwrap();
+}
+
+// Process negative reports
+
 /// Negative reports can be deduplicated, so we store to-be-processed
 /// negative reports as a map of [report] to [count of report]. Add this
 /// NR to that map (or create a new map if necessary).
@@ -244,6 +285,94 @@ pub fn save_negative_report_to_process(db: &Db, nr: NegativeReport) {
         .unwrap();
 }
 
+/// Sends a collection of negative reports to the Lox Authority and returns the
+/// number of valid reports returned by the server. The negative reports in the
+/// collection should all have the same bridge fingerprint, date, country, and
+/// distributor.
+pub async fn verify_negative_reports(
+    distributors: &BTreeMap<BridgeDistributor, String>,
+    reports: &BTreeMap<String, u32>,
+) -> u32 {
+    // Don't make a network call if we don't have any reports anyway
+    if reports.is_empty() {
+        return 0;
+    }
+    // Get one report, assume the rest have the same distributor
+    let first_report: SerializableNegativeReport =
+        serde_json::from_str(reports.first_key_value().unwrap().0).unwrap();
+    let distributor = first_report.distributor;
+    let client = Client::new();
+    let uri: String = (distributors.get(&distributor).unwrap().to_owned() + "/verifynegative")
+        .parse()
+        .unwrap();
+    let req = Request::builder()
+        .method(Method::POST)
+        .uri(uri)
+        .body(Body::from(serde_json::to_string(&reports).unwrap()))
+        .unwrap();
+    let resp = client.request(req).await.unwrap();
+    let buf = hyper::body::to_bytes(resp).await.unwrap();
+    serde_json::from_slice(&buf).unwrap()
+}
+
+/// Process today's negative reports and store the count of verified reports in
+/// the database.
+pub async fn update_negative_reports(db: &Db, distributors: &BTreeMap<BridgeDistributor, String>) {
+    let mut all_negative_reports = match db.get("nrs-to-process").unwrap() {
+        Some(v) => bincode::deserialize(&v).unwrap(),
+        None => BTreeMap::<String, BTreeMap<String, u32>>::new(),
+    };
+    for bridge_date in all_negative_reports.keys() {
+        // We could parse the fingerprint and date:
+        //let fingerprint: [u8; 20] = array_bytes::hex2array(&bridge_date[0..40]).unwrap();
+        //let date: u32 = &bridge_date[41..].parse().unwrap();
+        // but instead, let's just get it from the first report
+        let reports = all_negative_reports.get(bridge_date).unwrap();
+        if !reports.is_empty() {
+            let first_report: SerializableNegativeReport =
+                serde_json::from_str(reports.first_key_value().unwrap().0).unwrap();
+            let fingerprint = first_report.fingerprint;
+            let date = first_report.date;
+            let country = first_report.country;
+            let count_valid = verify_negative_reports(&distributors, reports).await;
+            let mut count_per_country = BTreeMap::<String, u32>::new();
+            count_per_country.insert(country, count_valid).unwrap();
+            let mut bridge_info = match db.get(&fingerprint).unwrap() {
+                Some(v) => bincode::deserialize(&v).unwrap(),
+                // It should already exist, unless the bridge hasn't published
+                // any bridge stats.
+                None => BridgeInfo::new(fingerprint, &"".to_string()),
+            };
+            // Add the new report count to it
+            if bridge_info.info_by_day.contains_key(&date) {
+                let daily_bridge_info = bridge_info.info_by_day.get_mut(&date).unwrap();
+                daily_bridge_info.add_info(BridgeInfoType::NegativeReports, &count_per_country);
+                // Commit changes to database
+                db.insert(fingerprint, bincode::serialize(&bridge_info).unwrap())
+                    .unwrap();
+            } else {
+                // No existing entry; make a new one.
+                let mut daily_bridge_info = DailyBridgeInfo::new();
+                daily_bridge_info.add_info(BridgeInfoType::NegativeReports, &count_per_country);
+                bridge_info.info_by_day.insert(date, daily_bridge_info);
+                // Commit changes to database
+                db.insert(fingerprint, bincode::serialize(&bridge_info).unwrap())
+                    .unwrap();
+            }
+        }
+    }
+    // TODO: Would it be cheaper to just recreate it?
+    all_negative_reports.clear();
+    // Remove the now-processed reports from the database
+    db.insert(
+        "nrs-to-process",
+        bincode::serialize(&all_negative_reports).unwrap(),
+    )
+    .unwrap();
+}
+
+// Process positive reports
+
 /// We store to-be-processed positive reports as a vector. Add this PR
 /// to that vector (or create a new vector if necessary).
 pub fn save_positive_report_to_process(db: &Db, pr: PositiveReport) {
@@ -271,3 +400,89 @@ pub fn save_positive_report_to_process(db: &Db, pr: PositiveReport) {
     db.insert("prs-to-process", bincode::serialize(&reports).unwrap())
         .unwrap();
 }
+
+/// Sends a collection of positive reports to the Lox Authority and returns the
+/// number of valid reports returned by the server. The positive reports in the
+/// collection should all have the same bridge fingerprint, date, and country.
+pub async fn verify_positive_reports(
+    distributors: &BTreeMap<BridgeDistributor, String>,
+    reports: &Vec<SerializablePositiveReport>,
+) -> u32 {
+    // Don't make a network call if we don't have any reports anyway
+    if reports.is_empty() {
+        return 0;
+    }
+    let client = Client::new();
+    let uri: String = (distributors
+        .get(&BridgeDistributor::Lox)
+        .unwrap()
+        .to_owned()
+        + "/verifypositive")
+        .parse()
+        .unwrap();
+    let req = Request::builder()
+        .method(Method::POST)
+        .uri(uri)
+        .body(Body::from(serde_json::to_string(&reports).unwrap()))
+        .unwrap();
+    let resp = client.request(req).await.unwrap();
+    let buf = hyper::body::to_bytes(resp).await.unwrap();
+    serde_json::from_slice(&buf).unwrap()
+}
+
+/// Process today's positive reports and store the count of verified reports in
+/// the database.
+pub async fn update_positive_reports(db: &Db, distributors: &BTreeMap<BridgeDistributor, String>) {
+    let mut all_positive_reports = match db.get("prs-to-process").unwrap() {
+        Some(v) => bincode::deserialize(&v).unwrap(),
+        None => BTreeMap::<String, Vec<SerializablePositiveReport>>::new(),
+    };
+    for bridge_date in all_positive_reports.keys() {
+        // We could parse the fingerprint and date:
+        //let fingerprint: [u8; 20] = array_bytes::hex2array(&bridge_date[0..40]).unwrap();
+        //let date: u32 = &bridge_date[41..].parse().unwrap();
+        // but instead, let's just get it from the first report
+        let reports = all_positive_reports.get(bridge_date).unwrap();
+        if !reports.is_empty() {
+            let first_report = &reports[0];
+            let fingerprint = first_report.fingerprint;
+            let date = first_report.date;
+            let country = first_report.country.clone();
+            let count_valid = verify_positive_reports(&distributors, reports).await;
+            let mut count_per_country = BTreeMap::<String, u32>::new();
+            count_per_country.insert(country, count_valid).unwrap();
+            let mut bridge_info = match db.get(&fingerprint).unwrap() {
+                Some(v) => bincode::deserialize(&v).unwrap(),
+                // It should already exist, unless the bridge hasn't published
+                // any bridge stats.
+                None => BridgeInfo::new(fingerprint, &"".to_string()),
+            };
+            // Add the new report count to it
+            if bridge_info.info_by_day.contains_key(&date) {
+                let daily_bridge_info = bridge_info.info_by_day.get_mut(&date).unwrap();
+                daily_bridge_info.add_info(BridgeInfoType::PositiveReports, &count_per_country);
+                // Commit changes to database
+                db.insert(fingerprint, bincode::serialize(&bridge_info).unwrap())
+                    .unwrap();
+            } else {
+                // No existing entry; make a new one.
+                let mut daily_bridge_info = DailyBridgeInfo::new();
+                daily_bridge_info.add_info(BridgeInfoType::PositiveReports, &count_per_country);
+                bridge_info.info_by_day.insert(date, daily_bridge_info);
+                // Commit changes to database
+                db.insert(fingerprint, bincode::serialize(&bridge_info).unwrap())
+                    .unwrap();
+            }
+        }
+    }
+    // TODO: Would it be cheaper to just recreate it?
+    all_positive_reports.clear();
+    // Remove the now-processed reports from the database
+    db.insert(
+        "prs-to-process",
+        bincode::serialize(&all_positive_reports).unwrap(),
+    )
+    .unwrap();
+}
+
+// TODO: function to mark a bridge as blocked