diff --git a/benchmarks/src/bin/tpch.rs b/benchmarks/src/bin/tpch.rs index f0977fe3267f..838709cf03b5 100644 --- a/benchmarks/src/bin/tpch.rs +++ b/benchmarks/src/bin/tpch.rs @@ -247,7 +247,7 @@ async fn benchmark_query( } let elapsed = start.elapsed().as_secs_f64() * 1000.0; - millis.push(elapsed as f64); + millis.push(elapsed); let row_count = result.iter().map(|b| b.num_rows()).sum(); println!( "Query {} iteration {} took {:.1} ms and returned {} rows", diff --git a/datafusion/core/benches/data_utils/mod.rs b/datafusion/core/benches/data_utils/mod.rs index 66603fd42322..575e1831c838 100644 --- a/datafusion/core/benches/data_utils/mod.rs +++ b/datafusion/core/benches/data_utils/mod.rs @@ -131,7 +131,7 @@ fn create_record_batch( schema, vec![ Arc::new(StringArray::from(keys)), - Arc::new(Float32Array::from_slice(&vec![i as f32; batch_size])), + Arc::new(Float32Array::from_slice(vec![i as f32; batch_size])), Arc::new(Float64Array::from(values)), Arc::new(UInt64Array::from(integer_values_wide)), Arc::new(UInt64Array::from(integer_values_narrow)), diff --git a/datafusion/core/benches/filter_query_sql.rs b/datafusion/core/benches/filter_query_sql.rs index 8dc981700598..8b0142078d90 100644 --- a/datafusion/core/benches/filter_query_sql.rs +++ b/datafusion/core/benches/filter_query_sql.rs @@ -49,8 +49,8 @@ fn create_context(array_len: usize, batch_size: usize) -> Result RecordBatch::try_new( schema.clone(), vec![ - Arc::new(Float32Array::from_slice(&vec![i as f32; batch_size])), - Arc::new(Float64Array::from_slice(&vec![i as f64; batch_size])), + Arc::new(Float32Array::from_slice(vec![i as f32; batch_size])), + Arc::new(Float64Array::from_slice(vec![i as f64; batch_size])), ], ) .unwrap() diff --git a/datafusion/core/benches/math_query_sql.rs b/datafusion/core/benches/math_query_sql.rs index 11b107f02f4b..87c7a363efc0 100644 --- a/datafusion/core/benches/math_query_sql.rs +++ b/datafusion/core/benches/math_query_sql.rs @@ -61,8 +61,8 @@ fn create_context( RecordBatch::try_new( schema.clone(), vec![ - Arc::new(Float32Array::from_slice(&vec![i as f32; batch_size])), - Arc::new(Float64Array::from_slice(&vec![i as f64; batch_size])), + Arc::new(Float32Array::from_slice(vec![i as f32; batch_size])), + Arc::new(Float64Array::from_slice(vec![i as f64; batch_size])), ], ) .unwrap() diff --git a/datafusion/core/src/catalog/information_schema.rs b/datafusion/core/src/catalog/information_schema.rs index eca399e27093..957cac53a93b 100644 --- a/datafusion/core/src/catalog/information_schema.rs +++ b/datafusion/core/src/catalog/information_schema.rs @@ -264,7 +264,7 @@ impl SchemaProvider for InformationSchemaProvider { } fn table_exist(&self, name: &str) -> bool { - return matches!(name.to_ascii_lowercase().as_str(), TABLES | VIEWS | COLUMNS); + matches!(name.to_ascii_lowercase().as_str(), TABLES | VIEWS | COLUMNS) } } diff --git a/datafusion/core/src/datasource/listing/url.rs b/datafusion/core/src/datasource/listing/url.rs index d1a527f23ac7..1c12309541b4 100644 --- a/datafusion/core/src/datasource/listing/url.rs +++ b/datafusion/core/src/datasource/listing/url.rs @@ -235,7 +235,7 @@ mod tests { let root = std::env::current_dir().unwrap(); let root = root.to_string_lossy(); - let url = ListingTableUrl::parse(&root).unwrap(); + let url = ListingTableUrl::parse(root).unwrap(); let child = url.prefix.child("partition").child("file"); let prefix: Vec<_> = url.strip_prefix(&child).unwrap().collect(); diff --git a/datafusion/core/src/datasource/mod.rs b/datafusion/core/src/datasource/mod.rs index 0a22fe804e9b..fc3e8f2d2913 100644 --- a/datafusion/core/src/datasource/mod.rs +++ b/datafusion/core/src/datasource/mod.rs @@ -125,8 +125,8 @@ pub async fn get_statistics_with_limit( }; let statistics = Statistics { - num_rows: Some(num_rows as usize), - total_byte_size: Some(total_byte_size as usize), + num_rows: Some(num_rows), + total_byte_size: Some(total_byte_size), column_statistics: column_stats, is_exact, }; @@ -167,7 +167,7 @@ fn get_col_stats( None => None, }; ColumnStatistics { - null_count: Some(null_counts[i] as usize), + null_count: Some(null_counts[i]), max_value, min_value, distinct_count: None, diff --git a/datafusion/core/src/execution/context.rs b/datafusion/core/src/execution/context.rs index d3989b5bd9f0..e0d2e06113fd 100644 --- a/datafusion/core/src/execution/context.rs +++ b/datafusion/core/src/execution/context.rs @@ -1630,7 +1630,7 @@ impl SessionState { Some(host) => format!("{}://{}", url.scheme(), host), None => format!("{}://", url.scheme()), }; - let path = &url.as_str()[authority.len() as usize..]; + let path = &url.as_str()[authority.len()..]; let path = object_store::path::Path::parse(path).expect("Can't parse path"); let store = ObjectStoreUrl::parse(authority.as_str()) .expect("Invalid default catalog url"); @@ -2649,7 +2649,7 @@ mod tests { // generate a partitioned file for partition in 0..partition_count { let filename = format!("partition-{}.{}", partition, file_extension); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = File::create(file_path)?; // generate some data diff --git a/datafusion/core/src/physical_plan/common.rs b/datafusion/core/src/physical_plan/common.rs index 4c6a624e7e93..c00c7421eaec 100644 --- a/datafusion/core/src/physical_plan/common.rs +++ b/datafusion/core/src/physical_plan/common.rs @@ -328,8 +328,8 @@ mod tests { RecordBatch::try_new( Arc::clone(&schema), vec![ - Arc::new(Float32Array::from_slice(&vec![i as f32; batch_size])), - Arc::new(Float64Array::from_slice(&vec![i as f64; batch_size])), + Arc::new(Float32Array::from_slice(vec![i as f32; batch_size])), + Arc::new(Float64Array::from_slice(vec![i as f64; batch_size])), ], ) .unwrap() diff --git a/datafusion/core/src/physical_plan/file_format/csv.rs b/datafusion/core/src/physical_plan/file_format/csv.rs index 51180c0f00a8..faa1401d2906 100644 --- a/datafusion/core/src/physical_plan/file_format/csv.rs +++ b/datafusion/core/src/physical_plan/file_format/csv.rs @@ -252,7 +252,7 @@ pub async fn plan_to_csv( for i in 0..plan.output_partitioning().partition_count() { let plan = plan.clone(); let filename = format!("part-{}.csv", i); - let path = fs_path.join(&filename); + let path = fs_path.join(filename); let file = fs::File::create(path)?; let mut writer = csv::Writer::new(file); let task_ctx = Arc::new(TaskContext::from(state)); @@ -539,7 +539,7 @@ mod tests { // generate a partitioned file for partition in 0..partition_count { let filename = format!("partition-{}.{}", partition, file_extension); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = File::create(file_path)?; // generate some data diff --git a/datafusion/core/src/physical_plan/file_format/json.rs b/datafusion/core/src/physical_plan/file_format/json.rs index ceb9e7958934..409c53639349 100644 --- a/datafusion/core/src/physical_plan/file_format/json.rs +++ b/datafusion/core/src/physical_plan/file_format/json.rs @@ -215,7 +215,7 @@ pub async fn plan_to_json( for i in 0..plan.output_partitioning().partition_count() { let plan = plan.clone(); let filename = format!("part-{}.json", i); - let path = fs_path.join(&filename); + let path = fs_path.join(filename); let file = fs::File::create(path)?; let mut writer = json::LineDelimitedWriter::new(file); let task_ctx = Arc::new(TaskContext::from(state)); diff --git a/datafusion/core/src/physical_plan/file_format/parquet.rs b/datafusion/core/src/physical_plan/file_format/parquet.rs index cf8763b07408..7a44749c85cc 100644 --- a/datafusion/core/src/physical_plan/file_format/parquet.rs +++ b/datafusion/core/src/physical_plan/file_format/parquet.rs @@ -601,7 +601,7 @@ pub async fn plan_to_parquet( for i in 0..plan.output_partitioning().partition_count() { let plan = plan.clone(); let filename = format!("part-{}.parquet", i); - let path = fs_path.join(&filename); + let path = fs_path.join(filename); let file = fs::File::create(path)?; let mut writer = ArrowWriter::try_new(file, plan.schema(), writer_properties.clone())?; @@ -1535,7 +1535,7 @@ mod tests { // generate a partitioned file for partition in 0..partition_count { let filename = format!("partition-{}.{}", partition, file_extension); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = File::create(file_path)?; // generate some data diff --git a/datafusion/core/src/physical_plan/file_format/parquet/row_filter.rs b/datafusion/core/src/physical_plan/file_format/parquet/row_filter.rs index e3cdc59199cc..353162c7f686 100644 --- a/datafusion/core/src/physical_plan/file_format/parquet/row_filter.rs +++ b/datafusion/core/src/physical_plan/file_format/parquet/row_filter.rs @@ -399,7 +399,7 @@ mod test { #[should_panic(expected = "building candidate failed")] fn test_filter_candidate_builder_ignore_projected_columns() { let testdata = crate::test_util::parquet_test_data(); - let file = std::fs::File::open(&format!("{}/alltypes_plain.parquet", testdata)) + let file = std::fs::File::open(format!("{}/alltypes_plain.parquet", testdata)) .expect("opening file"); let reader = SerializedFileReader::new(file).expect("creating reader"); @@ -423,7 +423,7 @@ mod test { #[test] fn test_filter_candidate_builder_ignore_complex_types() { let testdata = crate::test_util::parquet_test_data(); - let file = std::fs::File::open(&format!("{}/list_columns.parquet", testdata)) + let file = std::fs::File::open(format!("{}/list_columns.parquet", testdata)) .expect("opening file"); let reader = SerializedFileReader::new(file).expect("creating reader"); @@ -447,7 +447,7 @@ mod test { #[test] fn test_filter_candidate_builder_rewrite_missing_column() { let testdata = crate::test_util::parquet_test_data(); - let file = std::fs::File::open(&format!("{}/alltypes_plain.parquet", testdata)) + let file = std::fs::File::open(format!("{}/alltypes_plain.parquet", testdata)) .expect("opening file"); let reader = SerializedFileReader::new(file).expect("creating reader"); diff --git a/datafusion/core/src/test/mod.rs b/datafusion/core/src/test/mod.rs index f5c37eb053f0..be27b259608a 100644 --- a/datafusion/core/src/test/mod.rs +++ b/datafusion/core/src/test/mod.rs @@ -106,7 +106,7 @@ pub fn partitioned_file_groups( .get_ext_with_compression(file_compression_type.to_owned()) .unwrap() ); - let filename = tmp_dir.join(&filename); + let filename = tmp_dir.join(filename); let file = File::create(&filename).unwrap(); @@ -125,7 +125,7 @@ pub fn partitioned_file_groups( files.push(filename); } - let f = File::open(&path)?; + let f = File::open(path)?; let f = BufReader::new(f); for (i, line) in f.lines().enumerate() { let line = line.unwrap(); diff --git a/datafusion/core/tests/path_partition.rs b/datafusion/core/tests/path_partition.rs index fca9b9a43b1c..2d7e783f15ec 100644 --- a/datafusion/core/tests/path_partition.rs +++ b/datafusion/core/tests/path_partition.rs @@ -548,7 +548,7 @@ impl ObjectStore for MirroringObjectStore { ) -> object_store::Result { self.files.iter().find(|x| *x == location.as_ref()).unwrap(); let path = std::path::PathBuf::from(&self.mirrored_file); - let mut file = File::open(&path).unwrap(); + let mut file = File::open(path).unwrap(); file.seek(SeekFrom::Start(range.start as u64)).unwrap(); let to_read = range.end - range.start; diff --git a/datafusion/core/tests/provider_filter_pushdown.rs b/datafusion/core/tests/provider_filter_pushdown.rs index c1aa5ad7095c..84c8c50cd189 100644 --- a/datafusion/core/tests/provider_filter_pushdown.rs +++ b/datafusion/core/tests/provider_filter_pushdown.rs @@ -152,7 +152,7 @@ impl TableProvider for CustomProvider { Expr::Literal(ScalarValue::Int8(Some(i))) => *i as i64, Expr::Literal(ScalarValue::Int16(Some(i))) => *i as i64, Expr::Literal(ScalarValue::Int32(Some(i))) => *i as i64, - Expr::Literal(ScalarValue::Int64(Some(i))) => *i as i64, + Expr::Literal(ScalarValue::Int64(Some(i))) => *i, Expr::Cast(Cast { expr, data_type: _ }) => match expr.deref() { Expr::Literal(lit_value) => match lit_value { ScalarValue::Int8(Some(v)) => *v as i64, diff --git a/datafusion/core/tests/sql/aggregates.rs b/datafusion/core/tests/sql/aggregates.rs index 8f266a29f5cf..4b8a158fb4ff 100644 --- a/datafusion/core/tests/sql/aggregates.rs +++ b/datafusion/core/tests/sql/aggregates.rs @@ -2185,7 +2185,7 @@ async fn run_count_distinct_integers_aggregated_scenario( for (i, partition) in partitions.iter().enumerate() { let filename = format!("partition-{}.csv", i); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = File::create(file_path)?; for row in partition { let row_str = format!( diff --git a/datafusion/core/tests/sql/group_by.rs b/datafusion/core/tests/sql/group_by.rs index 057673dda5e7..56044862cdb9 100644 --- a/datafusion/core/tests/sql/group_by.rs +++ b/datafusion/core/tests/sql/group_by.rs @@ -501,7 +501,7 @@ async fn group_by_date_trunc() -> Result<()> { // generate a partitioned file for partition in 0..4 { let filename = format!("partition-{}.{}", partition, "csv"); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = File::create(file_path)?; // generate some data diff --git a/datafusion/core/tests/sql/mod.rs b/datafusion/core/tests/sql/mod.rs index c2d5726e259f..20d5371f1ce0 100644 --- a/datafusion/core/tests/sql/mod.rs +++ b/datafusion/core/tests/sql/mod.rs @@ -871,7 +871,7 @@ fn populate_csv_partitions( // generate a partitioned file for partition in 0..partition_count { let filename = format!("partition-{}.{}", partition, file_extension); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = File::create(file_path)?; // generate some data diff --git a/datafusion/core/tests/sql/parquet_schema.rs b/datafusion/core/tests/sql/parquet_schema.rs index 17bd88a24904..9c56ab1472c8 100644 --- a/datafusion/core/tests/sql/parquet_schema.rs +++ b/datafusion/core/tests/sql/parquet_schema.rs @@ -186,7 +186,7 @@ fn write_files(table_path: &Path, schemas: Vec) { for (i, schema) in schemas.into_iter().enumerate() { let schema = Arc::new(schema); let filename = format!("part-{}.parquet", i); - let path = table_path.join(&filename); + let path = table_path.join(filename); let file = fs::File::create(path).unwrap(); let mut writer = ArrowWriter::try_new(file, schema.clone(), None).unwrap(); diff --git a/datafusion/core/tests/sql/partitioned_csv.rs b/datafusion/core/tests/sql/partitioned_csv.rs index 9b87c2f09b67..efb63a296ec8 100644 --- a/datafusion/core/tests/sql/partitioned_csv.rs +++ b/datafusion/core/tests/sql/partitioned_csv.rs @@ -60,7 +60,7 @@ fn populate_csv_partitions( // generate a partitioned file for partition in 0..partition_count { let filename = format!("partition-{}.{}", partition, file_extension); - let file_path = tmp_dir.path().join(&filename); + let file_path = tmp_dir.path().join(filename); let mut file = std::fs::File::create(file_path)?; // generate some data diff --git a/datafusion/optimizer/src/common_subexpr_eliminate.rs b/datafusion/optimizer/src/common_subexpr_eliminate.rs index e8158e632469..8369c12c3c6a 100644 --- a/datafusion/optimizer/src/common_subexpr_eliminate.rs +++ b/datafusion/optimizer/src/common_subexpr_eliminate.rs @@ -543,7 +543,7 @@ impl ExprRewriter for CommonSubexprRewriter<'_> { // Alias this `Column` expr to it original "expr name", // `projection_push_down` optimizer use "expr name" to eliminate useless // projections. - Ok(col(id).alias(&expr_name)) + Ok(col(id).alias(expr_name)) } } diff --git a/datafusion/optimizer/src/filter_push_down.rs b/datafusion/optimizer/src/filter_push_down.rs index 674910cd1cc3..71de72732f6a 100644 --- a/datafusion/optimizer/src/filter_push_down.rs +++ b/datafusion/optimizer/src/filter_push_down.rs @@ -716,7 +716,7 @@ fn optimize(plan: &LogicalPlan, mut state: State) -> Result { // replace keys in join_cols_to_replace with values in resulting column // set .filter(|c| !join_cols_to_replace.contains_key(c)) - .chain(join_cols_to_replace.iter().map(|(_, v)| (*v).clone())) + .chain(join_cols_to_replace.values().map(|v| (*v).clone())) .collect(); Some(Ok((join_side_predicate, join_side_columns))) diff --git a/datafusion/physical-expr/src/aggregate/approx_percentile_cont.rs b/datafusion/physical-expr/src/aggregate/approx_percentile_cont.rs index e9f9f07212a6..7506757199e2 100644 --- a/datafusion/physical-expr/src/aggregate/approx_percentile_cont.rs +++ b/datafusion/physical-expr/src/aggregate/approx_percentile_cont.rs @@ -128,7 +128,7 @@ fn validate_input_percentile_expr(expr: &Arc) -> Result { .value(); let percentile = match lit { ScalarValue::Float32(Some(q)) => *q as f64, - ScalarValue::Float64(Some(q)) => *q as f64, + ScalarValue::Float64(Some(q)) => *q, got => return Err(DataFusionError::NotImplemented(format!( "Percentile value for 'APPROX_PERCENTILE_CONT' must be Float32 or Float64 literal (got data type {})", got.get_datatype() @@ -388,7 +388,7 @@ impl Accumulator for ApproxPercentileAccumulator { DataType::UInt32 => ScalarValue::UInt32(Some(q as u32)), DataType::UInt64 => ScalarValue::UInt64(Some(q as u64)), DataType::Float32 => ScalarValue::Float32(Some(q as f32)), - DataType::Float64 => ScalarValue::Float64(Some(q as f64)), + DataType::Float64 => ScalarValue::Float64(Some(q)), v => unreachable!("unexpected return type {:?}", v), }) } diff --git a/datafusion/physical-expr/src/aggregate/tdigest.rs b/datafusion/physical-expr/src/aggregate/tdigest.rs index e4112d848014..6457f7025fc7 100644 --- a/datafusion/physical-expr/src/aggregate/tdigest.rs +++ b/datafusion/physical-expr/src/aggregate/tdigest.rs @@ -420,8 +420,7 @@ impl TDigest { let mut compressed: Vec = Vec::with_capacity(max_size); let mut k_limit: f64 = 1.0; - let mut q_limit_times_count = - Self::k_to_q(k_limit, max_size as f64) * (count as f64); + let mut q_limit_times_count = Self::k_to_q(k_limit, max_size as f64) * (count); let mut iter_centroids = centroids.iter_mut(); let mut curr = iter_centroids.next().unwrap(); @@ -440,8 +439,7 @@ impl TDigest { sums_to_merge = 0_f64; weights_to_merge = 0_f64; compressed.push(curr.clone()); - q_limit_times_count = - Self::k_to_q(k_limit, max_size as f64) * (count as f64); + q_limit_times_count = Self::k_to_q(k_limit, max_size as f64) * (count); k_limit += 1.0; curr = centroid; } @@ -452,7 +450,7 @@ impl TDigest { compressed.shrink_to_fit(); compressed.sort(); - result.count = count as f64; + result.count = count; result.min = min; result.max = max; result.centroids = compressed; diff --git a/datafusion/physical-expr/src/expressions/binary.rs b/datafusion/physical-expr/src/expressions/binary.rs index 88dcaaccca04..a4904053086e 100644 --- a/datafusion/physical-expr/src/expressions/binary.rs +++ b/datafusion/physical-expr/src/expressions/binary.rs @@ -2437,10 +2437,10 @@ mod tests { let value: i128 = 123; let decimal_array = Arc::new(create_decimal_array( &[ - Some(value as i128), // 1.23 + Some(value), // 1.23 None, Some((value - 1) as i128), // 1.22 - Some((value + 1) as i128), // 1.24 + Some(value + 1), // 1.24 ], 10, 2, @@ -2563,7 +2563,7 @@ mod tests { &[ Some(value as i128), // 1.23 None, - Some((value - 1) as i128), // 1.22 + Some(value - 1), // 1.22 Some((value + 1) as i128), // 1.24 ], 10, @@ -2680,10 +2680,10 @@ mod tests { let value: i128 = 123; let decimal_array = Arc::new(create_decimal_array( &[ - Some(value as i128), // 1.23 + Some(value), // 1.23 None, Some((value - 1) as i128), // 1.22 - Some((value + 1) as i128), // 1.24 + Some(value + 1), // 1.24 ], 10, 2, diff --git a/datafusion/physical-expr/src/expressions/datetime.rs b/datafusion/physical-expr/src/expressions/datetime.rs index 9585e7ea0468..14e7072598f6 100644 --- a/datafusion/physical-expr/src/expressions/datetime.rs +++ b/datafusion/physical-expr/src/expressions/datetime.rs @@ -373,7 +373,7 @@ mod tests { match res { ColumnarValue::Scalar(ScalarValue::Date64(Some(d))) => { let epoch = NaiveDate::from_ymd_opt(1970, 1, 1).unwrap(); - let res = epoch.add(Duration::milliseconds(d as i64)); + let res = epoch.add(Duration::milliseconds(d)); assert_eq!(format!("{:?}", res).as_str(), "1969-12-16"); } _ => Err(DataFusionError::NotImplemented( diff --git a/datafusion/sql/src/planner.rs b/datafusion/sql/src/planner.rs index dacd4af87248..0fe8042e00b5 100644 --- a/datafusion/sql/src/planner.rs +++ b/datafusion/sql/src/planner.rs @@ -209,7 +209,7 @@ impl<'a, S: ContextProvider> SqlToRel<'a, S> { .project( plan.schema().fields().iter().zip(columns.into_iter()).map( |(field, ident)| { - col(field.name()).alias(&normalize_ident(&ident)) + col(field.name()).alias(normalize_ident(&ident)) }, ), ) @@ -717,7 +717,7 @@ impl<'a, S: ContextProvider> SqlToRel<'a, S> { JoinConstraint::Using(idents) => { let keys: Vec = idents .into_iter() - .map(|x| Column::from_name(&normalize_ident(&x))) + .map(|x| Column::from_name(normalize_ident(&x))) .collect(); LogicalPlanBuilder::from(left) .join_using(&right, join_type, keys)? @@ -840,7 +840,7 @@ impl<'a, S: ContextProvider> SqlToRel<'a, S> { Ok(LogicalPlanBuilder::from(plan.clone()) .project_with_alias( plan.schema().fields().iter().zip(columns_alias.iter()).map( - |(field, ident)| col(field.name()).alias(&normalize_ident(ident)), + |(field, ident)| col(field.name()).alias(normalize_ident(ident)), ), Some(normalize_ident(&alias.name)), )?