From b7dd7e6aad2d504597a22b75c4da967b19bd6acf Mon Sep 17 00:00:00 2001 From: Patrick Wang Date: Tue, 30 Apr 2024 10:11:24 -0400 Subject: [PATCH] fmt and clippy --- optd-perftest/src/datafusion_dbms.rs | 25 ++++++++++++++----------- optd-perftest/src/tpch.rs | 3 ++- 2 files changed, 16 insertions(+), 12 deletions(-) diff --git a/optd-perftest/src/datafusion_dbms.rs b/optd-perftest/src/datafusion_dbms.rs index c72d3ef8..88853b22 100644 --- a/optd-perftest/src/datafusion_dbms.rs +++ b/optd-perftest/src/datafusion_dbms.rs @@ -55,7 +55,7 @@ impl CardtestRunnerDBMSHelper for DatafusionDBMS { ) -> anyhow::Result> { let base_table_stats = self.get_benchmark_stats(benchmark).await?; // clear_state() is how we "load" the stats into datafusion - self.clear_state(Some(base_table_stats), &benchmark).await?; + self.clear_state(Some(base_table_stats), benchmark).await?; if self.adaptive { // We need to load the stats if we're doing adaptivity because that involves executing the queries in datafusion. @@ -95,7 +95,11 @@ impl DatafusionDBMS { /// /// A more ideal way to generate statistics would be to use the `ANALYZE` /// command in SQL, but DataFusion does not support that yet. - async fn clear_state(&mut self, stats: Option, benchmark: &Benchmark) -> anyhow::Result<()> { + async fn clear_state( + &mut self, + stats: Option, + benchmark: &Benchmark, + ) -> anyhow::Result<()> { let with_logical = match benchmark { Benchmark::Tpch(_) => WITH_LOGICAL_FOR_TPCH, Benchmark::Job(_) | Benchmark::Joblight(_) => WITH_LOGICAL_FOR_JOB, @@ -263,7 +267,7 @@ impl DatafusionDBMS { /// This is used to execute the query in order to load the true cardinalities back into optd /// in order to use the adaptive cost model. async fn execute_query(&self, sql: &str) -> anyhow::Result<()> { - Self::execute(&self.get_ctx(), sql).await?; + Self::execute(self.get_ctx(), sql).await?; Ok(()) } @@ -319,7 +323,7 @@ impl DatafusionDBMS { } Benchmark::Job(_) | Benchmark::Joblight(_) => { let job_kit = JobKit::build(&self.workspace_dpath)?; - Self::create_job_tables(&self.get_ctx(), &job_kit).await?; + Self::create_job_tables(self.get_ctx(), &job_kit).await?; } }; Ok(()) @@ -333,7 +337,7 @@ impl DatafusionDBMS { .filter(|s| !s.is_empty()) .collect::>(); for ddl in ddls { - Self::execute(&self.get_ctx(), ddl).await?; + Self::execute(self.get_ctx(), ddl).await?; } Ok(()) } @@ -367,7 +371,7 @@ impl DatafusionDBMS { for tbl_fpath in tbl_fpath_iter { let tbl_name = tbl_fpath.file_stem().unwrap().to_str().unwrap(); Self::execute( - &self.get_ctx(), + self.get_ctx(), &format!( "create external table {}_tbl stored as csv delimiter '|' location '{}';", tbl_name, @@ -392,7 +396,7 @@ impl DatafusionDBMS { .collect::>() .join(", "); Self::execute( - &self.get_ctx(), + self.get_ctx(), &format!( "insert into {} select {} from {}_tbl;", tbl_name, projection_list, tbl_name, @@ -413,7 +417,7 @@ impl DatafusionDBMS { // Download the tables. let job_kit = JobKit::build(&self.workspace_dpath)?; - job_kit.download_tables(&job_kit_config)?; + job_kit.download_tables(job_kit_config)?; // Create the tables. Self::create_job_tables(&ctx, &job_kit).await?; @@ -422,8 +426,7 @@ impl DatafusionDBMS { let tbl_fpath_iter = job_kit.get_tbl_fpath_iter().unwrap(); for tbl_fpath in tbl_fpath_iter { let tbl_name = tbl_fpath.file_stem().unwrap().to_str().unwrap(); - let schema = - ctx + let schema = ctx .catalog("datafusion") .unwrap() .schema("public") @@ -434,7 +437,7 @@ impl DatafusionDBMS { .schema(); self.get_ctx() .register_csv( - &tbl_name, + tbl_name, tbl_fpath.to_str().unwrap(), CsvReadOptions::new() .schema(&schema) diff --git a/optd-perftest/src/tpch.rs b/optd-perftest/src/tpch.rs index a0c7db4f..9260a194 100644 --- a/optd-perftest/src/tpch.rs +++ b/optd-perftest/src/tpch.rs @@ -13,7 +13,8 @@ use std::path::{Path, PathBuf}; const TPCH_KIT_REPO_URL: &str = "https://github.com/wangpatrick57/tpch-kit.git"; pub const TPCH_KIT_POSTGRES: &str = "POSTGRESQL"; const NUM_TPCH_QUERIES: usize = 22; -pub const WORKING_QUERY_IDS: &[&str] = &["2", "3", "5", "7", "8", "9", "10", "12", "13", "14", "17"]; +pub const WORKING_QUERY_IDS: &[&str] = + &["2", "3", "5", "7", "8", "9", "10", "12", "13", "14", "17"]; #[derive(Clone, Debug, Serialize, Deserialize)] pub struct TpchKitConfig {