[ https://issues.apache.org/jira/browse/ARROW-5044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Rok Mihevc updated ARROW-5044: ------------------------------ External issue URL: https://github.com/apache/arrow/issues/21539 > [Release][Rust] Format error in verification script > --------------------------------------------------- > > Key: ARROW-5044 > URL: https://issues.apache.org/jira/browse/ARROW-5044 > Project: Apache Arrow > Issue Type: Bug > Components: Packaging, Rust > Reporter: Kouhei Sutou > Assignee: Kouhei Sutou > Priority: Major > Labels: pull-request-available > Fix For: 0.13.0 > > Time Spent: 40m > Remaining Estimate: 0h > > {noformat} > + cargo fmt --all -- --check > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/arrow/src/csv/writer.rs at > line 53: > //! let batch = RecordBatch::try_new( > //! Arc::new(schema), > //! vec![Arc::new(c1), Arc::new(c2), Arc::new(c3), Arc::new(c4)], > -//! ).unwrap(); > +//! ) > +//! .unwrap(); > //! > //! let file = get_temp_file("out.csv", &[]); > //! > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/datasource/datasource.rs > at line 24: > > use crate::error::Result; > > -/// Returned by implementors of `Table#scan`, this `RecordBatchIterator` is > wrapped with an `Arc` > -/// and `Mutex` so that it can be shared across threads as it is used. > +/// Returned by implementors of `Table#scan`, this `RecordBatchIterator` is > wrapped with > +/// an `Arc` and `Mutex` so that it can be shared across threads as it is > used. > pub type ScanResult = Arc<Mutex<RecordBatchIterator>>; > > /// Source table > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/datasource/datasource.rs > at line 33: > /// Get a reference to the schema for this table > fn schema(&self) -> &Arc<Schema>; > > - /// Perform a scan of a table and return a sequence of iterators over > the data (one iterator per partition) > + /// Perform a scan of a table and return a sequence of iterators over > the data (one > + /// iterator per partition) > fn scan( > &self, > projection: &Option<Vec<usize>>, > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/datasource/memory.rs > at line 16: > // under the License. > > //! In-memory data source for presenting a Vec<RecordBatch> as a data source > that can be > -//! queried by DataFusion. This allows data to be pre-loaded into memory and > then repeatedly > -//! queried without incurring additional file I/O overhead. > +//! queried by DataFusion. This allows data to be pre-loaded into memory and > then > +//! repeatedly queried without incurring additional file I/O overhead. > > use std::sync::{Arc, Mutex}; > > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/context.rs > at line 15: > // specific language governing permissions and limitations > // under the License. > > -//! ExecutionContext contains methods for registering data sources and > executing SQL queries > +//! ExecutionContext contains methods for registering data sources and > executing SQL > +//! queries > > use std::cell::RefCell; > use std::collections::HashMap; > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/context.rs > at line 139: > Ok(plan) > } > > - /// Execute a logical plan and produce a Relation (a schema-aware > iterator over a series > - /// of RecordBatch instances) > + /// Execute a logical plan and produce a Relation (a schema-aware > iterator over a > + /// series of RecordBatch instances) > pub fn execute( > &mut self, > plan: &LogicalPlan, > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/filter.rs > at line 15: > // specific language governing permissions and limitations > // under the License. > > -//! Execution of a filter (predicate) relation. The SQL clause `WHERE expr` > represents a filter. > +//! Execution of a filter (predicate) relation. The SQL clause `WHERE expr` > represents a > +//! filter. > > use std::cell::RefCell; > use std::rc::Rc; > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/filter.rs > at line 32: > > /// Implementation of a filter relation > pub(super) struct FilterRelation { > - /// The schema for the filter relation. This is always the same as the > schema of the input relation. > + /// The schema for the filter relation. This is always the same as the > schema of the > + /// input relation. > schema: Arc<Schema>, > /// Relation that is being filtered > input: Rc<RefCell<Relation>>, > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/limit.rs > at line 33: > pub(super) struct LimitRelation { > /// The relation which the limit is being applied to > input: Rc<RefCell<Relation>>, > - /// The schema for the limit relation, which is always the same as the > schema of the input relation > + /// The schema for the limit relation, which is always the same as the > schema of the > + /// input relation > schema: Arc<Schema>, > /// The number of rows returned by this relation > limit: usize, > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/projection.rs > at line 15: > // specific language governing permissions and limitations > // under the License. > > -//! Defines the projection relation. A projection determines which columns > or expressions are > -//! returned from a query. The SQL statement `SELECT a, b, a+b FROM t1` is > an example of a > -//! projection on table `t1` where the expressions `a`, `b`, and `a+b` are > the projection > -//! expressions. > +//! Defines the projection relation. A projection determines which columns > or expressions > +//! are returned from a query. The SQL statement `SELECT a, b, a+b FROM t1` > is an example > +//! of a projection on table `t1` where the expressions `a`, `b`, and `a+b` > are the > +//! projection expressions. > > use std::cell::RefCell; > use std::rc::Rc; > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/execution/relation.rs > at line 16: > // under the License. > > //! A relation is a representation of a set of tuples. A database table is a > -//! type of relation. During query execution, each operation on a relation > (such as projection, > -//! selection, aggregation) results in a new relation. > +//! type of relation. During query execution, each operation on a relation > (such as > +//! projection, selection, aggregation) results in a new relation. > > use std::sync::{Arc, Mutex}; > > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/optimizer/optimizer.rs > at line 21: > use crate::logicalplan::LogicalPlan; > use std::sync::Arc; > > -/// An optimizer rules performs a transformation on a logical plan to > produce an optimized logical plan. > +/// An optimizer rules performs a transformation on a logical plan to > produce an optimized > +/// logical plan. > pub trait OptimizerRule { > /// Perform optimizations on the plan > fn optimize(&mut self, plan: &LogicalPlan) -> Result<Arc<LogicalPlan>>; > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/optimizer/projection_push_down.rs > at line 142: > schema, > .. > } => { > - // once we reach the table scan, we can use the accumulated > set of column indexes as > - // the projection in the table scan > + // once we reach the table scan, we can use the accumulated > set of column > + // indexes as the projection in the table scan > let mut projection: Vec<usize> = > Vec::with_capacity(accum.len()); > accum.iter().for_each(|i| projection.push(*i)); > > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/optimizer/projection_push_down.rs > at line 158: > } > let projected_schema = Schema::new(projected_fields); > > - // now that the table scan is returning a different schema > we need to create a > - // mapping from the original column index to the new column > index so that we > - // can rewrite expressions as we walk back up the tree > + // now that the table scan is returning a different schema > we need to > + // create a mapping from the original column index to the > + // new column index so that we can rewrite expressions as > + // we walk back up the tree > > if mapping.len() != 0 { > return Err(ExecutionError::InternalError( > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/optimizer/type_coercion.rs > at line 17: > > //! The type_coercion optimizer rule ensures that all binary operators are > operating on > //! compatible types by adding explicit cast operations to expressions. For > example, > -//! the operation `c_float + c_int` would be rewritten as `c_float + > CAST(c_int AS float)`. > -//! This keeps the runtime query execution code much simpler. > +//! the operation `c_float + c_int` would be rewritten as `c_float + > CAST(c_int AS > +//! float)`. This keeps the runtime query execution code much simpler. > > use std::sync::Arc; > > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/optimizer/utils.rs > at line 24: > use crate::error::{ExecutionError, Result}; > use crate::logicalplan::Expr; > > -/// Recursively walk a list of expression trees, collecting the unique set > of column indexes > -/// referenced in the expression > +/// Recursively walk a list of expression trees, collecting the unique set > of column > +/// indexes referenced in the expression > pub fn exprlist_to_column_indices(expr: &Vec<Expr>, accum: &mut > HashSet<usize>) { > expr.iter().for_each(|e| expr_to_column_indices(e, accum)); > } > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/src/table.rs at > line 15: > // specific language governing permissions and limitations > // under the License. > > -//! Table API for building a logical query plan. This is similar to the > Table API in Ibis and > -//! the DataFrame API in Apache Spark > +//! Table API for building a logical query plan. This is similar to the > Table API in Ibis > +//! and the DataFrame API in Apache Spark > > use crate::error::Result; > use crate::logicalplan::LogicalPlan; > Diff in > /tmp/arrow-0.13.0.tW4Dz/apache-arrow-0.13.0/rust/datafusion/tests/sql.rs at > line 129: > assert_eq!(expected, actual); > } > > -//TODO Uncomment the following test when ORDER BY is implemented to be able > to test ORDER BY + LIMIT > +//TODO Uncomment the following test when ORDER BY is implemented to be able > to test ORDER > +// BY + LIMIT > /* > #[test] > fn csv_query_limit_with_order_by() { > {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)