Skip to content
Closed
Show file tree
Hide file tree
Changes from 5 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 7 additions & 1 deletion ci/scripts/rust_fmt.sh
Original file line number Diff line number Diff line change
Expand Up @@ -18,4 +18,10 @@
# under the License.

set -ex
cargo fmt --all -- --check

# Install nightly toolchain (skips if already installed)
rustup toolchain install nightly --component rustfmt
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

would this update rust nightly as well?

what i mean, is there a chance we have new version of nightly installed every day without removing previous nightly installation

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

when I do rustup update it removes the existing version of nightly and replaces it with a new one

For example

andrewlamb@Andrews-MacBook-Pro-3:~$ rustup update
...
  stable-aarch64-apple-darwin unchanged - rustc 1.92.0 (ded5c06cf 2025-12-08)
   nightly-aarch64-apple-darwin updated - rustc 1.94.0-nightly (b68e16cfc 2026-01-11) (from rustc 1.94.0-nightly (8d670b93d 2025-12-31))
    1.74-aarch64-apple-darwin unchanged - rustc 1.74.1 (a28077b28 2023-12-04)
    1.81-aarch64-apple-darwin unchanged - rustc 1.81.0 (eeb90cda1 2024-09-04)
    1.88-aarch64-apple-darwin unchanged - rustc 1.88.0 (6b00bc388 2025-06-23)
    1.89-aarch64-apple-darwin unchanged - rustc 1.89.0 (29483883e 2025-08-04)
    1.90-aarch64-apple-darwin unchanged - rustc 1.90.0 (1159e78c4 2025-09-14)
    1.91-aarch64-apple-darwin unchanged - rustc 1.91.1 (ed61e7d7e 2025-11-07)

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

So it would install a new version every day, but would also remove the previous ones


# Use nightly rustfmt to check formatting including doc comments
# This requires nightly because format_code_in_doc_comments is an unstable feature
cargo +nightly fmt --all -- --check --config format_code_in_doc_comments=true
1 change: 0 additions & 1 deletion datafusion-examples/examples/data_io/remote_catalog.rs
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,6 @@
// under the License.

//! See `main.rs` for how to run it.
//!
/// This example shows how to implement the DataFusion [`CatalogProvider`] API
/// for catalogs that are remote (require network access) and/or offer only
/// asynchronous APIs such as [Polaris], [Unity], and [Hive].
Expand Down
1 change: 0 additions & 1 deletion datafusion-examples/examples/udf/simple_udaf.rs
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,6 @@
// under the License.

//! See `main.rs` for how to run it.
//!
/// In this example we will declare a single-type, single return type UDAF that computes the geometric mean.
/// The geometric mean is described here: https://en.wikipedia.org/wiki/Geometric_mean
use datafusion::arrow::{
Expand Down
8 changes: 4 additions & 4 deletions datafusion/common/src/scalar/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -2735,8 +2735,8 @@ impl ScalarValue {
/// ```
/// use arrow::array::{Int32Array, ListArray};
/// use arrow::datatypes::{DataType, Int32Type};
/// use datafusion_common::cast::as_list_array;
/// use datafusion_common::ScalarValue;
/// use datafusion_common::cast::as_list_array;
///
/// let scalars = vec![
/// ScalarValue::Int32(Some(1)),
Expand Down Expand Up @@ -2796,8 +2796,8 @@ impl ScalarValue {
/// ```
/// use arrow::array::{Int32Array, ListArray};
/// use arrow::datatypes::{DataType, Int32Type};
/// use datafusion_common::cast::as_list_array;
/// use datafusion_common::ScalarValue;
/// use datafusion_common::cast::as_list_array;
///
/// let scalars = vec![
/// ScalarValue::Int32(Some(1)),
Expand Down Expand Up @@ -2840,8 +2840,8 @@ impl ScalarValue {
/// ```
/// use arrow::array::{Int32Array, LargeListArray};
/// use arrow::datatypes::{DataType, Int32Type};
/// use datafusion_common::cast::as_large_list_array;
/// use datafusion_common::ScalarValue;
/// use datafusion_common::cast::as_large_list_array;
///
/// let scalars = vec![
/// ScalarValue::Int32(Some(1)),
Expand Down Expand Up @@ -3318,8 +3318,8 @@ impl ScalarValue {
/// ```
/// use arrow::array::ListArray;
/// use arrow::datatypes::{DataType, Int32Type};
/// use datafusion_common::utils::SingleRowListArrayBuilder;
/// use datafusion_common::ScalarValue;
/// use datafusion_common::utils::SingleRowListArrayBuilder;
/// use std::sync::Arc;
///
/// let list_arr = ListArray::from_iter_primitive::<Int32Type, _, _>(vec![
Expand Down
1 change: 1 addition & 0 deletions datafusion/core/src/dataframe/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1380,6 +1380,7 @@ impl DataFrame {
/// .read_csv("tests/data/example.csv", CsvReadOptions::new())
/// .await?;
/// let count = df.count().await?; // 1
///
/// # assert_eq!(count, 1);
/// # Ok(())
/// # }
Expand Down
9 changes: 5 additions & 4 deletions datafusion/core/src/execution/context/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -443,7 +443,7 @@ impl SessionContext {
/// # use datafusion::execution::SessionStateBuilder;
/// # use datafusion_optimizer::push_down_filter::PushDownFilter;
/// let my_rule = PushDownFilter {}; // pretend it is a new rule
/// // Create a new builder with a custom optimizer rule
/// // Create a new builder with a custom optimizer rule
/// let context: SessionContext = SessionStateBuilder::new()
/// .with_optimizer_rule(Arc::new(my_rule))
/// .build()
Expand Down Expand Up @@ -635,9 +635,10 @@ impl SessionContext {
/// .sql_with_options("CREATE TABLE foo (x INTEGER)", options)
/// .await
/// .unwrap_err();
/// assert!(err
/// .to_string()
/// .starts_with("Error during planning: DDL not supported: CreateMemoryTable"));
/// assert!(
/// err.to_string()
/// .starts_with("Error during planning: DDL not supported: CreateMemoryTable")
/// );
/// # Ok(())
/// # }
/// ```
Expand Down
2 changes: 1 addition & 1 deletion datafusion/core/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -491,7 +491,7 @@
//! consumes it immediately as well.
//!
//! ```text
//!
//!
//! Step 3: FilterExec calls next() Step 2: ProjectionExec calls
//! on input Stream next() on input Stream
//! ┌ ─ ─ ─ ─ ─ ─ ─ ─ ─ ┌ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ─ ┐
Expand Down
10 changes: 5 additions & 5 deletions datafusion/datasource-parquet/src/access_plan.rs
Original file line number Diff line number Diff line change
Expand Up @@ -42,11 +42,11 @@ use parquet::file::metadata::RowGroupMetaData;
/// // Use parquet reader RowSelector to specify scanning rows 100-200 and 350-400
/// // in a row group that has 1000 rows
/// let row_selection = RowSelection::from(vec![
/// RowSelector::skip(100),
/// RowSelector::select(100),
/// RowSelector::skip(150),
/// RowSelector::select(50),
/// RowSelector::skip(600), // skip last 600 rows
/// RowSelector::skip(100),
/// RowSelector::select(100),
/// RowSelector::skip(150),
/// RowSelector::select(50),
/// RowSelector::skip(600), // skip last 600 rows
/// ]);
/// access_plan.scan_selection(1, row_selection);
/// access_plan.skip(2); // skip row group 2
Expand Down
2 changes: 1 addition & 1 deletion datafusion/doc/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -196,7 +196,7 @@ impl Default for DocSection {
/// Example:
///
/// ```rust
///
///
/// # fn main() {
/// use datafusion_doc::{DocSection, Documentation};
/// let doc_section = DocSection {
Expand Down
2 changes: 1 addition & 1 deletion datafusion/expr-common/src/interval_arithmetic.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1317,7 +1317,7 @@ fn min_of_bounds(first: &ScalarValue, second: &ScalarValue) -> ScalarValue {
/// Example usage:
/// ```
/// use datafusion_common::DataFusionError;
/// use datafusion_expr_common::interval_arithmetic::{satisfy_greater, Interval};
/// use datafusion_expr_common::interval_arithmetic::{Interval, satisfy_greater};
///
/// let left = Interval::make(Some(-1000.0_f32), Some(1000.0_f32))?;
/// let right = Interval::make(Some(500.0_f32), Some(2000.0_f32))?;
Expand Down
2 changes: 1 addition & 1 deletion datafusion/expr-common/src/signature.rs
Original file line number Diff line number Diff line change
Expand Up @@ -967,7 +967,7 @@ fn get_data_types(native_type: &NativeType) -> Vec<DataType> {
/// # Examples
///
/// ```
/// use datafusion_common::types::{logical_binary, logical_string, NativeType};
/// use datafusion_common::types::{NativeType, logical_binary, logical_string};
/// use datafusion_expr_common::signature::{Coercion, TypeSignatureClass};
///
/// // Exact coercion that only accepts timestamp types
Expand Down
2 changes: 1 addition & 1 deletion datafusion/expr/src/expr.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1851,7 +1851,7 @@ impl Expr {
/// Example
/// ```
/// # use datafusion_common::Column;
/// use datafusion_expr::{col, Expr};
/// use datafusion_expr::{Expr, col};
/// let expr = col("foo");
/// assert_eq!(expr.try_as_col(), Some(&Column::from("foo")));
///
Expand Down
5 changes: 3 additions & 2 deletions datafusion/expr/src/logical_plan/ddl.rs
Original file line number Diff line number Diff line change
Expand Up @@ -255,8 +255,9 @@ impl CreateExternalTable {
/// TableReference::bare("my_table"),
/// "/path/to/data",
/// "parquet",
/// Arc::new(DFSchema::empty())
/// ).build();
/// Arc::new(DFSchema::empty()),
/// )
/// .build();
/// ```
pub fn builder(
name: impl Into<TableReference>,
Expand Down
8 changes: 4 additions & 4 deletions datafusion/expr/src/logical_plan/plan.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1587,7 +1587,7 @@ impl LogicalPlan {
///
/// ```
/// use arrow::datatypes::{DataType, Field, Schema};
/// use datafusion_expr::{col, lit, logical_plan::table_scan, LogicalPlanBuilder};
/// use datafusion_expr::{LogicalPlanBuilder, col, lit, logical_plan::table_scan};
/// let schema = Schema::new(vec![Field::new("id", DataType::Int32, false)]);
/// let plan = table_scan(Some("t1"), &schema, None)
/// .unwrap()
Expand Down Expand Up @@ -1629,7 +1629,7 @@ impl LogicalPlan {
///
/// ```
/// use arrow::datatypes::{DataType, Field, Schema};
/// use datafusion_expr::{col, lit, logical_plan::table_scan, LogicalPlanBuilder};
/// use datafusion_expr::{LogicalPlanBuilder, col, lit, logical_plan::table_scan};
/// let schema = Schema::new(vec![Field::new("id", DataType::Int32, false)]);
/// let plan = table_scan(Some("t1"), &schema, None)
/// .unwrap()
Expand Down Expand Up @@ -1694,7 +1694,7 @@ impl LogicalPlan {
///
/// ```
/// use arrow::datatypes::{DataType, Field, Schema};
/// use datafusion_expr::{col, lit, logical_plan::table_scan, LogicalPlanBuilder};
/// use datafusion_expr::{LogicalPlanBuilder, col, lit, logical_plan::table_scan};
/// let schema = Schema::new(vec![Field::new("id", DataType::Int32, false)]);
/// let plan = table_scan(Some("t1"), &schema, None)
/// .unwrap()
Expand Down Expand Up @@ -1752,7 +1752,7 @@ impl LogicalPlan {
/// ```
/// ```
/// use arrow::datatypes::{DataType, Field, Schema};
/// use datafusion_expr::{col, lit, logical_plan::table_scan, LogicalPlanBuilder};
/// use datafusion_expr::{LogicalPlanBuilder, col, lit, logical_plan::table_scan};
/// let schema = Schema::new(vec![Field::new("id", DataType::Int32, false)]);
/// let plan = table_scan(Some("t1"), &schema, None)
/// .unwrap()
Expand Down
1 change: 0 additions & 1 deletion datafusion/expr/src/predicate_bounds.rs
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,6 @@ use datafusion_expr_common::operator::Operator;
///
/// The function returns a [NullableInterval] that describes the possible boolean values the
/// predicate can evaluate to.
///
pub(super) fn evaluate_bounds(
predicate: &Expr,
certainly_null_expr: Option<&Expr>,
Expand Down
2 changes: 1 addition & 1 deletion datafusion/expr/src/udf.rs
Original file line number Diff line number Diff line change
Expand Up @@ -142,7 +142,7 @@ impl ScalarUDF {
///
/// # Example
/// ```no_run
/// use datafusion_expr::{col, lit, ScalarUDF};
/// use datafusion_expr::{ScalarUDF, col, lit};
/// # fn my_udf() -> ScalarUDF { unimplemented!() }
/// let my_func: ScalarUDF = my_udf();
/// // Create an expr for `my_func(a, 12.3)`
Expand Down
2 changes: 1 addition & 1 deletion datafusion/functions-aggregate-common/src/tdigest.rs
Original file line number Diff line number Diff line change
Expand Up @@ -523,7 +523,7 @@ impl TDigest {
/// [`ScalarValue::Float64`]:
///
/// ```text
///
///
/// ┌────────┬────────┬────────┬───────┬────────┬────────┐
/// │max_size│ sum │ count │ max │ min │centroid│
/// └────────┴────────┴────────┴───────┴────────┴────────┘
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -148,7 +148,7 @@ impl<S: SimplifyInfo> ExprSimplifier<S> {
/// use datafusion_expr::execution_props::ExecutionProps;
/// use datafusion_expr::simplify::SimplifyContext;
/// use datafusion_expr::simplify::SimplifyInfo;
/// use datafusion_expr::{col, lit, Expr};
/// use datafusion_expr::{Expr, col, lit};
/// use datafusion_optimizer::simplify_expressions::ExprSimplifier;
/// use std::sync::Arc;
///
Expand Down Expand Up @@ -290,7 +290,7 @@ impl<S: SimplifyInfo> ExprSimplifier<S> {
/// use datafusion_expr::execution_props::ExecutionProps;
/// use datafusion_expr::interval_arithmetic::{Interval, NullableInterval};
/// use datafusion_expr::simplify::SimplifyContext;
/// use datafusion_expr::{col, lit, Expr};
/// use datafusion_expr::{Expr, col, lit};
/// use datafusion_optimizer::simplify_expressions::ExprSimplifier;
///
/// let schema = Schema::new(vec![
Expand Down Expand Up @@ -352,7 +352,7 @@ impl<S: SimplifyInfo> ExprSimplifier<S> {
/// use datafusion_expr::execution_props::ExecutionProps;
/// use datafusion_expr::interval_arithmetic::{Interval, NullableInterval};
/// use datafusion_expr::simplify::SimplifyContext;
/// use datafusion_expr::{col, lit, Expr};
/// use datafusion_expr::{Expr, col, lit};
/// use datafusion_optimizer::simplify_expressions::ExprSimplifier;
///
/// let schema = Schema::new(vec![
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,6 @@ use std::fmt::Debug;
/// ELSE <optional-fallback_literal>
/// END
/// ```
///
#[derive(Debug)]
pub(in super::super) struct LiteralLookupTable {
/// The lookup table to use for evaluating the CASE expression
Expand Down Expand Up @@ -252,7 +251,6 @@ pub(super) trait WhenLiteralIndexMap: Debug + Send + Sync {
///
/// the returned vector will be:
/// - `[0, 2, else_index, 1, 0]`
///
fn map_to_when_indices(
&self,
array: &ArrayRef,
Expand Down
4 changes: 2 additions & 2 deletions datafusion/physical-expr/src/intervals/cp_solver.rs
Original file line number Diff line number Diff line change
Expand Up @@ -562,11 +562,11 @@ impl ExprIntervalGraph {
/// use arrow::datatypes::Field;
/// use arrow::datatypes::Schema;
/// use datafusion_common::ScalarValue;
/// use datafusion_expr::interval_arithmetic::Interval;
/// use datafusion_expr::Operator;
/// use datafusion_expr::interval_arithmetic::Interval;
/// use datafusion_physical_expr::PhysicalExpr;
/// use datafusion_physical_expr::expressions::{BinaryExpr, Column, Literal};
/// use datafusion_physical_expr::intervals::cp_solver::ExprIntervalGraph;
/// use datafusion_physical_expr::PhysicalExpr;
/// use std::sync::Arc;
///
/// let expr = Arc::new(BinaryExpr::new(
Expand Down
8 changes: 4 additions & 4 deletions datafusion/physical-expr/src/projection.rs
Original file line number Diff line number Diff line change
Expand Up @@ -260,12 +260,12 @@ impl ProjectionExprs {
/// # Example
///
/// ```rust
/// use std::sync::Arc;
/// use arrow::datatypes::{DataType, Field, Schema};
/// use datafusion_common::Result;
/// use datafusion_physical_expr::PhysicalExpr;
/// use datafusion_physical_expr::expressions::Column;
/// use datafusion_physical_expr::projection::ProjectionExprs;
/// use datafusion_physical_expr::PhysicalExpr;
/// use std::sync::Arc;
///
/// // Create a schema and projection
/// let schema = Arc::new(Schema::new(vec![
Expand Down Expand Up @@ -516,10 +516,10 @@ impl ProjectionExprs {
///
/// ```rust
/// use arrow::datatypes::{DataType, Field, Schema};
/// use datafusion_common::stats::{ColumnStatistics, Precision, Statistics};
/// use datafusion_physical_expr::projection::ProjectionExprs;
/// use datafusion_common::Result;
/// use datafusion_common::ScalarValue;
/// use datafusion_common::stats::{ColumnStatistics, Precision, Statistics};
/// use datafusion_physical_expr::projection::ProjectionExprs;
/// use std::sync::Arc;
///
/// fn main() -> Result<()> {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -180,7 +180,6 @@ pub struct GroupValuesColumn<const STREAMING: bool> {
/// We don't really store the actual `group values` in `hashtable`,
/// instead we store the `group indices` pointing to values in `GroupValues`.
/// And we use [`GroupIndexView`] to represent such `group indices` in table.
///
map: HashTable<(u64, GroupIndexView)>,

/// The size of `map` in bytes
Expand Down
2 changes: 1 addition & 1 deletion datafusion/physical-plan/src/aggregates/row_hash.rs
Original file line number Diff line number Diff line change
Expand Up @@ -234,7 +234,7 @@ enum OutOfMemoryMode {
/// # Architecture
///
/// ```text
///
///
/// Assigns a consecutive group internally stores aggregate values
/// index for each unique set for all groups
/// of group values
Expand Down
2 changes: 1 addition & 1 deletion datafusion/physical-plan/src/joins/hash_join/exec.rs
Original file line number Diff line number Diff line change
Expand Up @@ -211,7 +211,7 @@ impl JoinLeftData {
///
///
/// ```text
///
///
/// Original build-side data Inserting build-side values into hashmap Concatenated build-side batch
/// ┌───────────────────────────┐
/// hashmap.insert(row-hash, row-idx + offset) │ idx │
Expand Down
2 changes: 1 addition & 1 deletion datafusion/physical-plan/src/joins/hash_join/stream.rs
Original file line number Diff line number Diff line change
Expand Up @@ -110,7 +110,7 @@ impl BuildSide {
/// Expected state transitions performed by HashJoinStream are:
///
/// ```text
///
///
/// WaitBuildSide
/// │
/// ▼
Expand Down
15 changes: 12 additions & 3 deletions datafusion/physical-plan/src/sort_pushdown.rs
Original file line number Diff line number Diff line change
Expand Up @@ -100,15 +100,24 @@ impl<T> SortOrderPushdownResult<T> {
/// # use datafusion_physical_plan::SortOrderPushdownResult;
/// let exact = SortOrderPushdownResult::Exact { inner: 42 };
/// let inexact = exact.into_inexact();
/// assert!(matches!(inexact, SortOrderPushdownResult::Inexact { inner: 42 }));
/// assert!(matches!(
/// inexact,
/// SortOrderPushdownResult::Inexact { inner: 42 }
/// ));
///
/// let already_inexact = SortOrderPushdownResult::Inexact { inner: 42 };
/// let still_inexact = already_inexact.into_inexact();
/// assert!(matches!(still_inexact, SortOrderPushdownResult::Inexact { inner: 42 }));
/// assert!(matches!(
/// still_inexact,
/// SortOrderPushdownResult::Inexact { inner: 42 }
/// ));
///
/// let unsupported = SortOrderPushdownResult::<i32>::Unsupported;
/// let still_unsupported = unsupported.into_inexact();
/// assert!(matches!(still_unsupported, SortOrderPushdownResult::Unsupported));
/// assert!(matches!(
/// still_unsupported,
/// SortOrderPushdownResult::Unsupported
/// ));
/// ```
pub fn into_inexact(self) -> Self {
match self {
Expand Down
2 changes: 1 addition & 1 deletion datafusion/physical-plan/src/sorts/cursor.rs
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ pub trait CursorValues {
/// [`CursorValues`]
///
/// ```text
///
///
/// ┌───────────────────────┐
/// │ │ ┌──────────────────────┐
/// │ ┌─────────┐ ┌─────┐ │ ─ ─ ─ ─│ Cursor<T> │
Expand Down
Loading
Loading