Github user dongjoon-hyun commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22313#discussion_r214778954
  
    --- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/orc/OrcFilters.scala
 ---
    @@ -71,12 +71,24 @@ private[orc] object OrcFilters {
     
         for {
           // Combines all convertible filters using `And` to produce a single 
conjunction
    -      conjunction <- 
convertibleFilters.reduceOption(org.apache.spark.sql.sources.And)
    +      conjunction <- buildTree(convertibleFilters)
    --- End diff --
    
    BTW, Parquet has another issue here due to `.reduceOption(FilterApi.and)`. 
When I make a benchmark, Parquet seems to be unable to handle 1000 filters, 
@cloud-fan .


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to