Add better error messages

This commit is contained in:
Kiana Sheibani 2023-10-01 14:55:15 -04:00
parent e16b0be447
commit a4f130b36e
Signed by: toki
GPG key ID: 6CB106C25E86A9F7
8 changed files with 309 additions and 117 deletions

View file

@ -72,7 +72,7 @@ pub fn new_dataset(connection: &Connection, dataset: &str) -> sqlite::Result<()>
connection.execute(query)
}
pub fn get_last_sync(connection: &Connection, dataset: &str) -> sqlite::Result<Option<u64>> {
pub fn get_last_sync(connection: &Connection, dataset: &str) -> sqlite::Result<Option<Timestamp>> {
let query = "SELECT last_sync FROM datasets WHERE name = ?";
Ok(connection
@ -81,7 +81,8 @@ pub fn get_last_sync(connection: &Connection, dataset: &str) -> sqlite::Result<O
.bind((1, dataset))?
.map(|x| x.map(|r| r.read::<i64, _>("last_sync").to_owned() as u64))
.next()
.and_then(Result::ok))
.and_then(Result::ok)
.map(Timestamp))
}
pub fn update_last_sync(connection: &Connection, dataset: &str, sync: u64) -> sqlite::Result<()> {

View file

@ -3,6 +3,7 @@
use clap::{Parser, Subcommand};
use std::io::{self, Write};
use std::path::PathBuf;
use std::process::exit;
use std::time::SystemTime;
mod queries;
@ -12,6 +13,11 @@ use datasets::*;
mod sync;
use sync::*;
pub fn error(msg: &str, code: i32) -> ! {
println!("\nERROR: {}", msg);
exit(code)
}
/// ## CLI Structs
#[derive(Parser)]
@ -104,65 +110,91 @@ fn main() {
}
fn dataset_list() {
let config_dir = dirs::config_dir().unwrap();
let config_dir = dirs::config_dir().expect("Could not determine config directory");
let connection = open_datasets(&config_dir).unwrap();
let datasets = list_datasets(&connection).unwrap();
let connection =
open_datasets(&config_dir).unwrap_or_else(|_| error("Could not open datasets file", 1));
let datasets = list_datasets(&connection).expect("Error communicating with SQLite");
println!("{:?}", datasets);
}
fn dataset_new(name: Option<String>) {
let config_dir = dirs::config_dir().unwrap();
let name = name.unwrap_or_else(|| {
fn read_string() -> String {
let mut line = String::new();
print!("Name of new dataset: ");
io::stdout().flush().expect("Could not access stdout");
io::stdin()
.read_line(&mut line)
.expect("Could not read from stdin");
line.trim().to_owned()
}
fn dataset_new(name: Option<String>) {
let config_dir = dirs::config_dir().expect("Could not determine config directory");
let name = name.unwrap_or_else(|| {
print!("Name of new dataset: ");
read_string()
});
let connection = open_datasets(&config_dir).unwrap();
new_dataset(&connection, &name).unwrap();
let connection =
open_datasets(&config_dir).unwrap_or_else(|_| error("Could not open datasets file", 1));
new_dataset(&connection, &name).expect("Error communicating with SQLite");
}
fn dataset_delete(name: Option<String>) {
let config_dir = dirs::config_dir().unwrap();
let config_dir = dirs::config_dir().expect("Could not determine config directory");
let name = name.unwrap_or_else(|| {
let mut line = String::new();
print!("Dataset to delete: ");
io::stdout().flush().expect("Could not access stdout");
io::stdin()
.read_line(&mut line)
.expect("Could not read from stdin");
line.trim().to_owned()
read_string()
});
let connection = open_datasets(&config_dir).unwrap();
delete_dataset(&connection, &name).unwrap();
let connection =
open_datasets(&config_dir).unwrap_or_else(|_| error("Could not open datasets file", 1));
delete_dataset(&connection, &name).expect("Error communicating with SQLite");
}
fn sync(datasets: Vec<String>, all: bool, auth_token: Option<String>) {
let config_dir = dirs::config_dir().unwrap();
let auth = auth_token.or_else(|| get_auth_token(&config_dir)).unwrap();
let auth = auth_token
.or_else(|| get_auth_token(&config_dir))
.unwrap_or_else(|| error("Access token not provided", 1));
let connection = open_datasets(&config_dir).unwrap();
let connection =
open_datasets(&config_dir).unwrap_or_else(|_| error("Could not open datasets file", 1));
#[allow(unused_must_use)]
let datasets = if all {
list_datasets(&connection).unwrap()
} else if datasets.len() == 0 {
new_dataset(&connection, "default").unwrap();
new_dataset(&connection, "default");
vec![String::from("default")]
} else {
datasets
};
for dataset in datasets {
let last_sync = get_last_sync(&connection, &dataset).unwrap().unwrap();
let last_sync = get_last_sync(&connection, &dataset)
.expect("Error communicating with SQLite")
.unwrap_or_else(|| error(&format!("Dataset {} does not exist!", dataset), 1));
sync_dataset(
&connection,
&dataset,
last_sync,
VideogameId(1386),
Some("GA"),
&auth,
)
.expect("Error communicating with SQLite");
let current_time = SystemTime::now()
.duration_since(SystemTime::UNIX_EPOCH)
.unwrap_or_else(|_| error("System time is before the Unix epoch!", 2))
.as_secs();
update_last_sync(&connection, &dataset, current_time)
.expect("Error communicating with SQLite");
}
}

View file

@ -2,6 +2,8 @@ use cynic::{GraphQlResponse, QueryBuilder};
use serde::{Deserialize, Serialize};
use std::fmt::Debug;
use std::path::Path;
use std::thread::sleep;
use std::time::Duration;
pub mod search_games;
pub use search_games::*;
@ -61,7 +63,7 @@ pub struct EntrantId(pub u64);
#[cynic(graphql_type = "ID")]
pub struct PlayerId(pub u64);
#[derive(cynic::Scalar, Debug, Clone)]
#[derive(cynic::Scalar, Debug, Copy, Clone)]
pub struct Timestamp(pub u64);
// Query machinery
@ -75,19 +77,28 @@ pub trait QueryUnwrap<Vars>: 'static + QueryBuilder<Vars> {
// Generic function for running start.gg queries
pub fn run_query<Builder, Vars>(vars: Vars, auth_token: &str) -> Option<Builder::Unwrapped>
where
Builder: Debug,
Vars: Clone,
Builder: QueryUnwrap<Vars>,
Vars: Serialize,
for<'de> Builder: Deserialize<'de>,
{
use cynic::http::ReqwestBlockingExt;
let query = Builder::build(vars);
let response = reqwest::blocking::Client::new()
let mut response = reqwest::blocking::Client::new()
.post("https://api.start.gg/gql/alpha")
.header("Authorization", String::from("Bearer ") + auth_token)
.run_graphql(query);
.run_graphql(Builder::build(vars.clone()));
Builder::unwrap_response(response.unwrap())
for _ in 1..10 {
sleep(Duration::from_secs(2));
response = reqwest::blocking::Client::new()
.post("https://api.start.gg/gql/alpha")
.header("Authorization", String::from("Bearer ") + auth_token)
.run_graphql(Builder::build(vars.clone()));
if response.is_ok() {
break;
}
}
Builder::unwrap_response(response.ok()?)
}

View file

@ -6,10 +6,10 @@ pub type Teams<T> = Vec<Vec<T>>;
// Variables
#[derive(cynic::QueryVariables, Debug)]
#[derive(cynic::QueryVariables, Debug, Clone)]
pub struct EventSetsVars {
pub event: EventId,
pub sets_page: i32,
pub page: i32,
}
// Query
@ -24,16 +24,22 @@ pub struct EventSets {
#[derive(cynic::QueryFragment, Debug)]
#[cynic(variables = "EventSetsVars")]
struct Event {
#[arguments(page: $sets_page, perPage: 11)]
#[arguments(page: $page, perPage: 50)]
sets: Option<SetConnection>,
}
#[derive(cynic::QueryFragment, Debug)]
struct SetConnection {
page_info: Option<PageInfo>,
#[cynic(flatten)]
nodes: Vec<Set>,
}
#[derive(cynic::QueryFragment, Debug)]
struct PageInfo {
total_pages: Option<i32>,
}
#[derive(cynic::QueryFragment, Debug)]
struct Set {
#[arguments(includeByes: true)]
@ -68,21 +74,24 @@ struct Player {
// Unwrap
pub struct EventSetsResponse {
pub pages: u64,
pub sets: Vec<SetData>,
}
pub struct SetData {
teams: Teams<PlayerData>,
winner: usize,
pub teams: Teams<PlayerData>,
pub winner: usize,
}
impl QueryUnwrap<EventSetsVars> for EventSets {
type Unwrapped = Vec<SetData>;
type Unwrapped = EventSetsResponse;
// This might be the most spaghetti code I've ever written
fn unwrap_response(response: GraphQlResponse<EventSets>) -> Option<Vec<SetData>> {
Some(
response
.data?
.event?
.sets?
fn unwrap_response(response: GraphQlResponse<EventSets>) -> Option<EventSetsResponse> {
let response_sets = response.data?.event?.sets?;
let sets = response_sets
.nodes
.into_iter()
.filter_map(|set| {
@ -114,7 +123,11 @@ impl QueryUnwrap<EventSetsVars> for EventSets {
.try_collect()?;
Some(SetData { teams, winner })
})
.collect::<Vec<_>>(),
)
.collect::<Vec<_>>();
Some(EventSetsResponse {
pages: response_sets.page_info?.total_pages? as u64,
sets,
})
}
}

View file

@ -4,7 +4,7 @@ use schema::schema;
// Variables
#[derive(cynic::QueryVariables, Debug)]
#[derive(cynic::QueryVariables, Debug, Clone)]
pub struct PlayerInfoVars {
pub id: PlayerId,
}

View file

@ -4,7 +4,7 @@ use schema::schema;
// Variables
#[derive(cynic::QueryVariables)]
#[derive(cynic::QueryVariables, Clone)]
pub struct VideogameSearchVars<'a> {
pub name: &'a str,
}

View file

@ -5,14 +5,16 @@ use schema::schema;
// Variables
#[derive(cynic::QueryVariables, Debug)]
pub struct TournamentEventsVars {
#[derive(cynic::QueryVariables, Debug, Clone)]
pub struct TournamentEventsVars<'a> {
// HACK: This should really be an optional variable, but there seems to be a
// server-side bug that completely breaks everything when this isn't passed.
// We can use a dummy value of 1 when we don't want to filter by time.
pub last_query: Timestamp,
pub last_sync: Timestamp,
pub game_id: VideogameId,
pub page: i32,
pub state: Option<&'a str>,
}
// Query
@ -22,12 +24,13 @@ pub struct TournamentEventsVars {
pub struct TournamentEvents {
#[arguments(query: {
page: $page,
perPage: 300,
perPage: 250,
sortBy: "endAt asc",
filter: {
past: true,
afterDate: $last_query,
afterDate: $last_sync,
videogameIds: [$game_id],
addrState: $state
}})]
tournaments: Option<TournamentConnection>,
}
@ -35,10 +38,16 @@ pub struct TournamentEvents {
#[derive(cynic::QueryFragment, Debug)]
#[cynic(variables = "TournamentEventsVars")]
struct TournamentConnection {
page_info: Option<PageInfo>,
#[cynic(flatten)]
nodes: Vec<Tournament>,
}
#[derive(cynic::QueryFragment, Debug)]
struct PageInfo {
total_pages: Option<i32>,
}
#[derive(cynic::QueryFragment, Debug)]
#[cynic(variables = "TournamentEventsVars")]
struct Tournament {
@ -56,20 +65,27 @@ struct Event {
// Unwrap
#[derive(Debug, Clone)]
pub struct TournamentEventResponse {
pub pages: i32,
pub tournaments: Vec<TournamentData>,
}
#[derive(Debug, Clone)]
pub struct TournamentData {
pub name: String,
pub events: Vec<EventId>,
}
impl QueryUnwrap<TournamentEventsVars> for TournamentEvents {
type Unwrapped = Vec<TournamentData>;
impl<'a> QueryUnwrap<TournamentEventsVars<'a>> for TournamentEvents {
type Unwrapped = TournamentEventResponse;
fn unwrap_response(response: GraphQlResponse<TournamentEvents>) -> Option<Vec<TournamentData>> {
Some(
response
.data?
.tournaments?
fn unwrap_response(
response: GraphQlResponse<TournamentEvents>,
) -> Option<TournamentEventResponse> {
let response_tournaments = response.data?.tournaments?;
let tournaments = response_tournaments
.nodes
.into_iter()
.filter_map(|tour| {
@ -82,7 +98,11 @@ impl QueryUnwrap<TournamentEventsVars> for TournamentEvents {
.collect(),
})
})
.collect(),
)
.collect::<Vec<_>>();
Some(TournamentEventResponse {
pages: response_tournaments.page_info?.total_pages?,
tournaments,
})
}
}

View file

@ -1,4 +1,8 @@
use std::thread::sleep;
use std::time::Duration;
use crate::datasets::*;
use crate::error;
use crate::queries::*;
use sqlite::*;
@ -33,13 +37,105 @@ fn adjust_ratings(ratings: Teams<&mut f64>, winner: usize) {
// Extract set data
fn get_event_sets(event: EventId, auth: &str) -> Option<Vec<SetData>> {
let sets = run_query::<EventSets, _>(EventSetsVars {
sleep(Duration::from_millis(700));
let sets_response = run_query::<EventSets, _>(EventSetsVars { event, page: 1 }, auth)?;
let pages = sets_response.pages;
if pages == 0 {
Some(vec![])
} else if pages == 1 {
Some(sets_response.sets)
} else {
println!(" (Page 1)");
let mut sets = sets_response.sets;
for page in 2..=pages {
println!(" (Page {})", page);
let next_response = run_query::<EventSets, _>(
EventSetsVars {
event,
sets_page: 1,
});
page: page as i32,
},
auth,
)?;
sleep(Duration::from_millis(700));
sets.extend(next_response.sets);
}
/*
Some(sets)
}
}
fn get_tournament_events(
last_sync: Timestamp,
game_id: VideogameId,
state: Option<&str>,
auth: &str,
) -> Option<Vec<EventId>> {
println!("Accessing tournaments...");
let tour_response = run_query::<TournamentEvents, _>(
TournamentEventsVars {
last_sync,
game_id,
state,
page: 1,
},
auth,
)?;
let pages = tour_response.pages;
if pages == 0 {
Some(vec![])
} else if pages == 1 {
Some(
tour_response
.tournaments
.into_iter()
.flat_map(|tour| tour.events)
.collect::<Vec<_>>(),
)
} else {
println!(" (Page 1)");
let mut tournaments = tour_response
.tournaments
.into_iter()
.flat_map(|tour| tour.events)
.collect::<Vec<_>>();
for page in 2..=pages {
println!(" (Page {})", page);
let next_response = run_query::<TournamentEvents, _>(
TournamentEventsVars {
last_sync,
game_id,
state,
page,
},
auth,
)?;
tournaments.extend(
next_response
.tournaments
.into_iter()
.flat_map(|tour| tour.events),
);
}
Some(tournaments)
}
}
// Dataset syncing
fn update_from_set(connection: &Connection, dataset: &str, results: SetData) -> sqlite::Result<()> {
let players_data = results.teams;
add_players(connection, dataset, &players_data)?;
@ -54,14 +150,33 @@ fn update_from_set(connection: &Connection, dataset: &str, results: SetData) ->
update_ratings(connection, dataset, elos)
}
pub fn update_from_tournament(
pub fn sync_dataset(
connection: &Connection,
dataset: &str,
results: TournamentData,
last_sync: Timestamp,
game_id: VideogameId,
state: Option<&str>,
auth: &str,
) -> sqlite::Result<()> {
results
.sets
.into_iter()
.try_for_each(|set| update_from_set(connection, dataset, set))
let events = get_tournament_events(last_sync, game_id, state, auth)
.unwrap_or_else(|| error("Could not access start.gg", 1));
connection.execute("BEGIN;")?;
let num_events = events.len();
for (i, event) in events.into_iter().enumerate() {
println!(
"Accessing sets from event ID {}... ({}/{})",
event.0, i, num_events
);
let sets =
get_event_sets(event, auth).unwrap_or_else(|| error("Could not access start.gg", 1));
println!(" Updating ratings from event...");
sets.into_iter()
.try_for_each(|set| update_from_set(connection, dataset, set))?;
}
connection.execute("COMMIT;")
}
*/