[jira] [Deleted] (SPARK-30709) Spark 2.3 to Spark 2.4 Upgrade. Problems reading HIVE partitioned tables.

2024-04-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-30709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen deleted SPARK-30709:
-


> Spark 2.3 to Spark 2.4 Upgrade. Problems reading HIVE partitioned tables.
> -
>
> Key: SPARK-30709
> URL: https://issues.apache.org/jira/browse/SPARK-30709
> Project: Spark
>  Issue Type: Question
> Environment: PRE- Production
>Reporter: Carlos Mario
>Priority: Major
>  Labels: SQL, Spark
>
> Hello
> We recently updated our preproduction environment from Spark 2.3 to Spark 
> 2.4.0
> Along time we have created a big amount of tables in Hive Metastore, 
> partitioned by 2 fields one of them String and the other one BigInt.
> We were reading this tables with Spark 2.3 with no problem, but after 
> upgrading to Spark 2.4 we get the following log every time we run our SW:
> 
> log_filterBIGINT.out:
>  Caused by: MetaException(message:Filtering is supported only on partition 
> keys of type string) Caused by: MetaException(message:Filtering is supported 
> only on partition keys of type string) Caused by: 
> MetaException(message:Filtering is supported only on partition keys of type 
> string)
>  
> hadoop-cmf-hive-HIVEMETASTORE-isblcsmsttc0001.scisb.isban.corp.log.out.1:
>  
> 2020-01-10 09:36:05,781 ERROR 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler: [pool-5-thread-138]: 
> MetaException(message:Filtering is supported only on partition keys of type 
> string)
> 2020-01-10 11:19:19,208 ERROR 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler: [pool-5-thread-187]: 
> MetaException(message:Filtering is supported only on partition keys of type 
> string)
> 2020-01-10 11:19:54,780 ERROR 
> org.apache.hadoop.hive.metastore.RetryingHMSHandler: [pool-5-thread-167]: 
> MetaException(message:Filtering is supported only on partition keys of type 
> string)
>  
>  
> We know the best practice from Spark point of view is to use 'STRING' type 
> for partition columns, but we need to explore a solution we'll be able to 
> deploy with ease, due to the big amount of tables created with a bigiint type 
> column partition.
>  
> As a first solution we tried to set the  
> spark.sql.hive.manageFilesourcePartitions parameter to false in the Spark 
> Submmit, but after reruning the SW the error stood still.
>  
> Is there anyone in the community who experienced the same problem? What was 
> the solution for it? 
>  
> Kind Regards and thanks in advance.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-47172) Upgrade Transport block cipher mode to GCM

2024-02-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-47172?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-47172:
-
Shepherd:   (was: Sean R. Owen)

> Upgrade Transport block cipher mode to GCM
> --
>
> Key: SPARK-47172
> URL: https://issues.apache.org/jira/browse/SPARK-47172
> Project: Spark
>  Issue Type: Improvement
>  Components: Security
>Affects Versions: 3.4.2, 3.5.0
>Reporter: Steve Weis
>Priority: Minor
>
> The cipher transformation currently used for encrypting RPC calls is an 
> unauthenticated mode (AES/CTR/NoPadding). This needs to be upgraded to an 
> authenticated mode (AES/GCM/NoPadding) to prevent ciphertext from being 
> modified in transit.
> The relevant line is here: 
> [https://github.com/apache/spark/blob/a939a7d0fd9c6b23c879cbee05275c6fbc939e38/common/network-common/src/main/java/org/apache/spark/network/util/TransportConf.java#L220]
> GCM is relatively more computationally expensive than CTR and adds a 16-byte 
> block of authentication tag data to each payload. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46760) Make the document of spark.sql.adaptive.coalescePartitions.parallelismFirst clearer

2024-02-03 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46760?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46760.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44787
[https://github.com/apache/spark/pull/44787]

> Make the document of spark.sql.adaptive.coalescePartitions.parallelismFirst 
> clearer
> ---
>
> Key: SPARK-46760
> URL: https://issues.apache.org/jira/browse/SPARK-46760
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46760) Make the document of spark.sql.adaptive.coalescePartitions.parallelismFirst clearer

2024-02-03 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46760?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46760:
-
Priority: Trivial  (was: Major)

> Make the document of spark.sql.adaptive.coalescePartitions.parallelismFirst 
> clearer
> ---
>
> Key: SPARK-46760
> URL: https://issues.apache.org/jira/browse/SPARK-46760
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Trivial
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45110) Upgrade rocksdbjni to 8.8.1

2024-02-01 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45110:
-
Summary: Upgrade rocksdbjni to 8.8.1  (was: Upgrade rocksdbjni to 8.6.7)

> Upgrade rocksdbjni to 8.8.1
> ---
>
> Key: SPARK-45110
> URL: https://issues.apache.org/jira/browse/SPARK-45110
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 3.5.0, 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45110) Upgrade rocksdbjni to 8.6.7

2024-02-01 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45110.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43924
[https://github.com/apache/spark/pull/43924]

> Upgrade rocksdbjni to 8.6.7
> ---
>
> Key: SPARK-45110
> URL: https://issues.apache.org/jira/browse/SPARK-45110
> Project: Spark
>  Issue Type: Bug
>  Components: Build
>Affects Versions: 3.5.0, 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45110) Upgrade rocksdbjni to 8.8.1

2024-02-01 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45110?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45110:
-
Issue Type: Task  (was: Bug)

> Upgrade rocksdbjni to 8.8.1
> ---
>
> Key: SPARK-45110
> URL: https://issues.apache.org/jira/browse/SPARK-45110
> Project: Spark
>  Issue Type: Task
>  Components: Build
>Affects Versions: 3.5.0, 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46473) Reuse `getPartitionedFile` method

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46473.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44437
[https://github.com/apache/spark/pull/44437]

> Reuse `getPartitionedFile` method
> -
>
> Key: SPARK-46473
> URL: https://issues.apache.org/jira/browse/SPARK-46473
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.5.0
>Reporter: xiaoping.huang
>Assignee: xiaoping.huang
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-46473) Reuse `getPartitionedFile` method

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-46473:


Assignee: xiaoping.huang

> Reuse `getPartitionedFile` method
> -
>
> Key: SPARK-46473
> URL: https://issues.apache.org/jira/browse/SPARK-46473
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.5.0
>Reporter: xiaoping.huang
>Assignee: xiaoping.huang
>Priority: Trivial
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46929) Use ThreadUtils.shutdown to close thread pools

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46929.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44962
[https://github.com/apache/spark/pull/44962]

> Use ThreadUtils.shutdown to close thread pools
> --
>
> Key: SPARK-46929
> URL: https://issues.apache.org/jira/browse/SPARK-46929
> Project: Spark
>  Issue Type: Improvement
>  Components: Connect, Spark Core, SS
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46929) Use ThreadUtils.shutdown to close thread pools

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46929?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46929:
-
Priority: Trivial  (was: Major)

> Use ThreadUtils.shutdown to close thread pools
> --
>
> Key: SPARK-46929
> URL: https://issues.apache.org/jira/browse/SPARK-46929
> Project: Spark
>  Issue Type: Improvement
>  Components: Connect, Spark Core, SS
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Trivial
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46400) When there are corrupted files in the local maven repo, retry to skip this cache

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46400.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44343
[https://github.com/apache/spark/pull/44343]

> When there are corrupted files in the local maven repo, retry to skip this 
> cache
> 
>
> Key: SPARK-46400
> URL: https://issues.apache.org/jira/browse/SPARK-46400
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-46400) When there are corrupted files in the local maven repo, retry to skip this cache

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46400?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-46400:


Assignee: BingKun Pan

> When there are corrupted files in the local maven repo, retry to skip this 
> cache
> 
>
> Key: SPARK-46400
> URL: https://issues.apache.org/jira/browse/SPARK-46400
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45522) Migrate jetty 9 to jetty 12

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45522:


Assignee: HiuFung

> Migrate jetty 9 to jetty 12
> ---
>
> Key: SPARK-45522
> URL: https://issues.apache.org/jira/browse/SPARK-45522
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: HiuFung
>Priority: Minor
>  Labels: pull-request-available
>
> Jetty 12 supports JakartaEE 8/JakartaEE 9/JakartaEE 10 simultaneously. But 
> the version span is quite large, need to read the documentation in detail, 
> not sure if it can be completed within the 4.0 cycle, so it's set to low 
> priority.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45522) Migrate jetty 9 to jetty 12

2024-01-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45522.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43765
[https://github.com/apache/spark/pull/43765]

> Migrate jetty 9 to jetty 12
> ---
>
> Key: SPARK-45522
> URL: https://issues.apache.org/jira/browse/SPARK-45522
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: HiuFung
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> Jetty 12 supports JakartaEE 8/JakartaEE 9/JakartaEE 10 simultaneously. But 
> the version span is quite large, need to read the documentation in detail, 
> not sure if it can be completed within the 4.0 cycle, so it's set to low 
> priority.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46473) Reuse `getPartitionedFile` method

2024-01-30 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46473:
-
Priority: Trivial  (was: Minor)

> Reuse `getPartitionedFile` method
> -
>
> Key: SPARK-46473
> URL: https://issues.apache.org/jira/browse/SPARK-46473
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.5.0
>Reporter: xiaoping.huang
>Priority: Trivial
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-31794) Incorrect distribution with repartitionByRange and repartition column expression

2024-01-24 Thread Sean R. Owen (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-31794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17810476#comment-17810476
 ] 

Sean R. Owen commented on SPARK-31794:
--

Not that it helps, but I observe the same behavior, and it must be a bug. The 
result is exactly as expected, except that the first partition is 2x the size 
of the others, and the last partition is empty. I tried tweaks to identify the 
problem (i.e. what if I increased the desired partition count by 1? added 1 to 
my indices?) but didn't help.

I don't have any fix or further insight, just adding that yeah this seems to be 
a problem

> Incorrect distribution with repartitionByRange and repartition column 
> expression
> 
>
> Key: SPARK-31794
> URL: https://issues.apache.org/jira/browse/SPARK-31794
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 2.3.2, 2.4.5, 3.0.1
> Environment: Sample code for obtaining the above test results.
> import java.io.File 
> import java.io.PrintWriter 
> val logfile="/tmp/sparkdftest.log"
> val writer = new PrintWriter(logfile) 
> writer.println("Spark Version " + sc.version)
> val df= Range(1, 1002).toDF("val")
> writer.println("Default Partition Length:" + df.rdd.partitions.length)
> writer.println("Default Partition getNumPartitions:" + 
> df.rdd.getNumPartitions)
> writer.println("Default Partition groupBy spark_partition_id:" + 
> df.groupBy(spark_partition_id).count().rdd.partitions.length)
> val dfcount=df.mapPartitions\{part => Iterator(part.size)}
> writer.println("Default Partition:" + dfcount.collect().toList)
> val numparts=24
> val dfparts_range=df.withColumn("partid", $"val" % 
> numparts).repartitionByRange(numparts, $"partid")
> writer.println("repartitionByRange Length:" + 
> dfparts_range.rdd.partitions.length)
> writer.println("repartitionByRange getNumPartitions:" + 
> dfparts_range.rdd.getNumPartitions)
> writer.println("repartitionByRange groupBy spark_partition_id:" + 
> dfparts_range.groupBy(spark_partition_id).count().rdd.partitions.length)
> val dfpartscount=dfparts_range.mapPartitions\{part => Iterator(part.size)}
> writer.println("repartitionByRange: " + dfpartscount.collect().toList)
> val dfparts_expr=df.withColumn("partid", $"val" % 
> numparts).repartition(numparts, $"partid")
> writer.println("repartition by column expr Length:" + 
> dfparts_expr.rdd.partitions.length)
> writer.println("repartition by column expr getNumPartitions:" + 
> dfparts_expr.rdd.getNumPartitions)
> writer.println("repartition by column expr groupBy spark_partitoin_id:" + 
> dfparts_expr.groupBy(spark_partition_id).count().rdd.partitions.length)
> val dfpartscount=dfparts_expr.mapPartitions\{part => Iterator(part.size)}
> writer.println("repartition by column expr:" + dfpartscount.collect().toList)
> writer.close()
>Reporter: Ramesha Bhatta
>Priority: Major
>  Labels: performance
>
> Both repartitionByRange and  repartition(, )  resulting in wrong 
> distribution within the resulting partition.  
>  
> In the Range partition one of the partition has 2x volume and last one with 
> zero.  In repartition this is more problematic with some partition with 4x, 
> 2x the avg and many partitions with zero volume.  
>  
> This distribution imbalance can cause performance problem in a concurrent 
> environment.
> Details from testing in 3 different versions.
> |Verion 2.3.2|Version 2.4.5|Versoin 3.0 Preview2|
> |Spark Version 2.3.2.3.1.4.0-315|Spark Version 2.4.5|Spark Version 
> 3.0.0-preview2|
> |Default Partition Length:2|Default Partition Length:2|Default Partition 
> Length:80|
> |Default Partition getNumPartitions:2|Default Partition 
> getNumPartitions:2|Default Partition getNumPartitions:80|
> |Default Partition groupBy spark_partition_id:200|Default Partition groupBy 
> spark_partition_id:200|Default Partition groupBy spark_partition_id:200|
> |repartitionByRange Length:24|repartitionByRange Length:24|repartitionByRange 
> Length:24|
> |repartitionByRange getNumPartitions:24|repartitionByRange 
> getNumPartitions:24|repartitionByRange getNumPartitions:24|
> |repartitionByRange groupBy spark_partition_id:200|repartitionByRange groupBy 
> spark_partition_id:200|repartitionByRange groupBy spark_partition_id:200|
> |repartitionByRange: List(83, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 
> 42, 42, 42, 42, 41, 41, 41, 41, 41, 41, 0)|repartitionByRange: List(83, 42, 
> 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 42, 41, 41, 41, 41, 
> 41, 41, 0)|repartitionByRange: List(83, 42, 42, 42, 42, 42, 42, 42, 42, 42, 
> 42, 42, 42, 42, 42, 42, 42, 41, 41, 41, 41, 41, 41, 0)|
> |repartition by column expr Length:24|repartition by column expr 
> Length:24|repartition by column expr Length:24|
> |repartition by column 

[jira] [Updated] (SPARK-46637) Enhancing the Visual Appeal of Spark doc website

2024-01-09 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46637?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46637:
-
Priority: Minor  (was: Major)

> Enhancing the Visual Appeal of Spark doc website
> 
>
> Key: SPARK-46637
> URL: https://issues.apache.org/jira/browse/SPARK-46637
> Project: Spark
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 4.0.0, 3.5.1
>Reporter: Gengliang Wang
>Assignee: Gengliang Wang
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46124) Replace explicit `ArrayOps#toSeq` with `s.c.immutable.ArraySeq.unsafeWrapArray`

2023-11-29 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46124?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46124.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44041
[https://github.com/apache/spark/pull/44041]

> Replace explicit  `ArrayOps#toSeq` with 
> `s.c.immutable.ArraySeq.unsafeWrapArray`
> 
>
> Key: SPARK-46124
> URL: https://issues.apache.org/jira/browse/SPARK-46124
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Kubernetes, ML, MLlib, Spark Core, SQL, 
> Structured Streaming, YARN
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> There is a behavioral difference between Scala 2.13 and 2.12 for explicit 
> `ArrayOps.toSeq` calls, similar to the implicit conversion from `Array` to 
> `Seq`.
> In Scala 2.12, it returns a `mutable.WrappedArray`, which does not involve a 
> collection copy.
> ```scala
> Welcome to Scala 2.12.18 (OpenJDK 64-Bit Server VM, Java 17.0.9).
> Type in expressions for evaluation. Or try :help.
> scala> Array(1,2,3).toSeq
> res0: Seq[Int] = WrappedArray(1, 2, 3)
> ```
> However, in Scala 2.13, it returns an `immutable.ArraySeq` that with 
> collection copy.
> Since we have always used the non-collection copy behavior for this explicit 
> conversion in the era of Scala 2.12, it is safe to assume that no collection 
> copy is needed for Scala 2.13.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46098) Reduce stack depth by replace (string|array).size with (string|array).length

2023-11-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46098:
-
Priority: Minor  (was: Major)

> Reduce stack depth by replace (string|array).size with (string|array).length
> 
>
> Key: SPARK-46098
> URL: https://issues.apache.org/jira/browse/SPARK-46098
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Minor
> Fix For: 4.0.0
>
>
> There are a lot of (string|array).size called.
> In fact, the size calls the underlying length, this behavior increase the 
> stack length.
> We should call (string|array).length directly.
> We also get the compile waring Replace .size with .length on arrays and 
> strings



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46098) Reduce stack depth by replace (string|array).size with (string|array).length

2023-11-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46098.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

> Reduce stack depth by replace (string|array).size with (string|array).length
> 
>
> Key: SPARK-46098
> URL: https://issues.apache.org/jira/browse/SPARK-46098
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Major
> Fix For: 4.0.0
>
>
> There are a lot of (string|array).size called.
> In fact, the size calls the underlying length, this behavior increase the 
> stack length.
> We should call (string|array).length directly.
> We also get the compile waring Replace .size with .length on arrays and 
> strings



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46101) Replace (string|array).size with (string|array).length in all the modules

2023-11-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46101.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44016
[https://github.com/apache/spark/pull/44016]

> Replace (string|array).size with (string|array).length in all the modules
> -
>
> Key: SPARK-46101
> URL: https://issues.apache.org/jira/browse/SPARK-46101
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-46100) Replace (string|array).size with (string|array).length in module core

2023-11-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-46100.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 44011
[https://github.com/apache/spark/pull/44011]

> Replace (string|array).size with (string|array).length in module core
> -
>
> Key: SPARK-46100
> URL: https://issues.apache.org/jira/browse/SPARK-46100
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46101) Replace (string|array).size with (string|array).length in module SQL

2023-11-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46101:
-
Priority: Minor  (was: Major)
 Summary: Replace (string|array).size with (string|array).length in module 
SQL  (was: Fix these issue in module sql)

> Replace (string|array).size with (string|array).length in module SQL
> 
>
> Key: SPARK-46101
> URL: https://issues.apache.org/jira/browse/SPARK-46101
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46100) Replace (string|array).size with (string|array).length in module core

2023-11-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46100:
-
Summary: Replace (string|array).size with (string|array).length in module 
core  (was: Fix these issue in module core)

> Replace (string|array).size with (string|array).length in module core
> -
>
> Key: SPARK-46100
> URL: https://issues.apache.org/jira/browse/SPARK-46100
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-46100) Replace (string|array).size with (string|array).length in module core

2023-11-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-46100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-46100:
-
Priority: Minor  (was: Major)

> Replace (string|array).size with (string|array).length in module core
> -
>
> Key: SPARK-46100
> URL: https://issues.apache.org/jira/browse/SPARK-46100
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Jiaan Geng
>Assignee: Jiaan Geng
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45686) Fix `method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is deprecated`

2023-11-10 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45686?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45686:


Assignee: Yang Jie

> Fix `method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated`
> 
>
> Key: SPARK-45686
> URL: https://issues.apache.org/jira/browse/SPARK-45686
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:57:31:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(s1.indices, s1.values, s2.indices, 
> s2.values)
> [error]                               ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:57:54:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(s1.indices, s1.values, s2.indices, 
> s2.values)
> [error]                                                      ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:59:31:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(s1.indices, s1.values, 0 until d1.size, 
> d1.values)
> [error]                               ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:61:59:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(0 until d1.size, d1.values, s1.indices, 
> s1.values)
> [error]  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45686) Fix `method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is deprecated`

2023-11-10 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45686?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45686.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43670
[https://github.com/apache/spark/pull/43670]

> Fix `method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated`
> 
>
> Key: SPARK-45686
> URL: https://issues.apache.org/jira/browse/SPARK-45686
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:57:31:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(s1.indices, s1.values, s2.indices, 
> s2.values)
> [error]                               ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:57:54:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(s1.indices, s1.values, s2.indices, 
> s2.values)
> [error]                                                      ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:59:31:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(s1.indices, s1.values, 0 until d1.size, 
> d1.values)
> [error]                               ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/mllib-local/src/main/scala/org/apache/spark/ml/linalg/Vectors.scala:61:59:
>  method copyArrayToImmutableIndexedSeq in class LowPriorityImplicits2 is 
> deprecated (since 2.13.0): implicit conversions from Array to 
> immutable.IndexedSeq are implemented by copying; use `toIndexedSeq` 
> explicitly if you want to copy, or use the more efficient non-copying 
> ArraySeq.unsafeWrapArray
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=deprecation, 
> site=org.apache.spark.ml.linalg.Vector.equals, 
> origin=scala.LowPriorityImplicits2.copyArrayToImmutableIndexedSeq, 
> version=2.13.0
> [error]             Vectors.equals(0 until d1.size, d1.values, s1.indices, 
> s1.values)
> [error]  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45687) Fix `Passing an explicit array value to a Scala varargs method is deprecated`

2023-11-10 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45687.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43642
[https://github.com/apache/spark/pull/43642]

> Fix `Passing an explicit array value to a Scala varargs method is deprecated`
> -
>
> Key: SPARK-45687
> URL: https://issues.apache.org/jira/browse/SPARK-45687
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Tengfei Huang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
>  
> {code:java}
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/AggregationQuerySuite.scala:945:21:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.AggregationQuerySuite, version=2.13.0
> [warn]         df.agg(udaf(allColumns: _*)),
> [warn]                     ^
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ObjectHashAggregateSuite.scala:156:48:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.ObjectHashAggregateSuite, 
> version=2.13.0
> [warn]         df.agg(aggFunctions.head, aggFunctions.tail: _*),
> [warn]                                                ^
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ObjectHashAggregateSuite.scala:161:76:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.ObjectHashAggregateSuite, 
> version=2.13.0
> [warn]         df.groupBy($"id" % 4 as "mod").agg(aggFunctions.head, 
> aggFunctions.tail: _*),
> [warn]                                                                        
>     ^
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ObjectHashAggregateSuite.scala:171:50:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.ObjectHashAggregateSuite, 
> version=2.13.0
> [warn]           df.agg(aggFunctions.head, aggFunctions.tail: _*),
> [warn]  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45687) Fix `Passing an explicit array value to a Scala varargs method is deprecated`

2023-11-10 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45687:


Assignee: Tengfei Huang

> Fix `Passing an explicit array value to a Scala varargs method is deprecated`
> -
>
> Key: SPARK-45687
> URL: https://issues.apache.org/jira/browse/SPARK-45687
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Tengfei Huang
>Priority: Major
>  Labels: pull-request-available
>
> Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
>  
> {code:java}
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/AggregationQuerySuite.scala:945:21:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.AggregationQuerySuite, version=2.13.0
> [warn]         df.agg(udaf(allColumns: _*)),
> [warn]                     ^
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ObjectHashAggregateSuite.scala:156:48:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.ObjectHashAggregateSuite, 
> version=2.13.0
> [warn]         df.agg(aggFunctions.head, aggFunctions.tail: _*),
> [warn]                                                ^
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ObjectHashAggregateSuite.scala:161:76:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.ObjectHashAggregateSuite, 
> version=2.13.0
> [warn]         df.groupBy($"id" % 4 as "mod").agg(aggFunctions.head, 
> aggFunctions.tail: _*),
> [warn]                                                                        
>     ^
> [warn] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/ObjectHashAggregateSuite.scala:171:50:
>  Passing an explicit array value to a Scala varargs method is deprecated 
> (since 2.13.0) and will result in a defensive copy; Use the more efficient 
> non-copying ArraySeq.unsafeWrapArray or an explicit toIndexedSeq call
> [warn] Applicable -Wconf / @nowarn filters for this warning: msg= message>, cat=deprecation, 
> site=org.apache.spark.sql.hive.execution.ObjectHashAggregateSuite, 
> version=2.13.0
> [warn]           df.agg(aggFunctions.head, aggFunctions.tail: _*),
> [warn]  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-42821) Remove unused parameters in splitFiles methods

2023-11-08 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-42821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-42821.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 40454
[https://github.com/apache/spark/pull/40454]

> Remove unused parameters in splitFiles methods
> --
>
> Key: SPARK-42821
> URL: https://issues.apache.org/jira/browse/SPARK-42821
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-42821) Remove unused parameters in splitFiles methods

2023-11-08 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-42821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-42821:


Assignee: BingKun Pan

> Remove unused parameters in splitFiles methods
> --
>
> Key: SPARK-42821
> URL: https://issues.apache.org/jira/browse/SPARK-42821
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45368) Remove scala2.12 compatibility logic for DoubleType, FloatType, Decimal

2023-10-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45368.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43456
[https://github.com/apache/spark/pull/43456]

> Remove scala2.12 compatibility logic for DoubleType, FloatType, Decimal
> ---
>
> Key: SPARK-45368
> URL: https://issues.apache.org/jira/browse/SPARK-45368
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: tangjiafu
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45368) Remove scala2.12 compatibility logic for DoubleType, FloatType, Decimal

2023-10-31 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45368:


Assignee: tangjiafu

> Remove scala2.12 compatibility logic for DoubleType, FloatType, Decimal
> ---
>
> Key: SPARK-45368
> URL: https://issues.apache.org/jira/browse/SPARK-45368
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: tangjiafu
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45605) Replace `s.c.MapOps.mapValues` with `s.c.MapOps.view.mapValues`

2023-10-29 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45605:
-
Priority: Minor  (was: Major)

>Replace `s.c.MapOps.mapValues` with `s.c.MapOps.view.mapValues`
> --
>
> Key: SPARK-45605
> URL: https://issues.apache.org/jira/browse/SPARK-45605
> Project: Spark
>  Issue Type: Sub-task
>  Components: Connect, DStreams, Examples, MLlib, Spark Core, SS
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> @deprecated("Use .view.mapValues(f). A future version will include a strict 
> version of this method (for now, .view.mapValues(f).toMap).", "2.13.0")
> def mapValues[W](f: V => W): MapView[K, W] = new MapView.MapValues(this, f) 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45605) Replace `s.c.MapOps.mapValues` with `s.c.MapOps.view.mapValues`

2023-10-29 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45605.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43448
[https://github.com/apache/spark/pull/43448]

>Replace `s.c.MapOps.mapValues` with `s.c.MapOps.view.mapValues`
> --
>
> Key: SPARK-45605
> URL: https://issues.apache.org/jira/browse/SPARK-45605
> Project: Spark
>  Issue Type: Sub-task
>  Components: Connect, DStreams, Examples, MLlib, Spark Core, SS
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> @deprecated("Use .view.mapValues(f). A future version will include a strict 
> version of this method (for now, .view.mapValues(f).toMap).", "2.13.0")
> def mapValues[W](f: V => W): MapView[K, W] = new MapView.MapValues(this, f) 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45636) Upgrade jersey to 2.41

2023-10-29 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45636.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43490
[https://github.com/apache/spark/pull/43490]

> Upgrade jersey to 2.41
> --
>
> Key: SPARK-45636
> URL: https://issues.apache.org/jira/browse/SPARK-45636
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> https://github.com/eclipse-ee4j/jersey/releases/tag/2.41



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45636) Upgrade jersey to 2.41

2023-10-29 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45636:


Assignee: Yang Jie

> Upgrade jersey to 2.41
> --
>
> Key: SPARK-45636
> URL: https://issues.apache.org/jira/browse/SPARK-45636
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> https://github.com/eclipse-ee4j/jersey/releases/tag/2.41



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45636) Upgrade jersey to 2.41

2023-10-29 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45636:
-
Priority: Minor  (was: Major)

> Upgrade jersey to 2.41
> --
>
> Key: SPARK-45636
> URL: https://issues.apache.org/jira/browse/SPARK-45636
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> https://github.com/eclipse-ee4j/jersey/releases/tag/2.41



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-40154) PySpark: DataFrame.cache docstring gives wrong storage level

2023-10-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-40154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-40154:
-
Priority: Trivial  (was: Minor)

> PySpark: DataFrame.cache docstring gives wrong storage level
> 
>
> Key: SPARK-40154
> URL: https://issues.apache.org/jira/browse/SPARK-40154
> Project: Spark
>  Issue Type: Bug
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Paul Staab
>Assignee: Paul Staab
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 3.4.2, 4.0.0, 3.5.1
>
>
> The docstring of the `DataFrame.cache()` method currently states that it uses 
> a serialized storage level
> {code:java}
> Persists the :class:`DataFrame` with the default storage level 
> (`MEMORY_AND_DISK`).
> [...]
> -The default storage level has changed to `MEMORY_AND_DISK` to match 
> Scala in 2.0.{code}
> while `DataFrame.persist()` states that it uses a deserialized storage level
> {code:java}
> If no storage level is specified defaults to (`MEMORY_AND_DISK_DESER`)
> [...]
> The default storage level has changed to `MEMORY_AND_DISK_DESER` to match 
> Scala in 3.0.{code}
>  
> However, in practice both `.cache()` and `.persist()` use deserialized 
> storage levels:
> {code:java}
> import pyspark
> from pyspark.sql import SparkSession
> from pyspark import StorageLevel
> print(pyspark.__version__)
> # 3.3.0
> spark = SparkSession.builder.master("local[2]").getOrCreate()
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.cache()
> df.count()
> # Storage level in Spark UI: "Disk Memory Deserialized 1x Replicated"
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.persist()
> df.count()
> # Storage level in Spark UI: "Disk Memory Deserialized 1x Replicated"
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.persist(StorageLevel.MEMORY_AND_DISK)
> df.count()
> # Storage level in Spark UI: "Disk Memory Serialized 1x Replicated"{code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-40154) PySpark: DataFrame.cache docstring gives wrong storage level

2023-10-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-40154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-40154.
--
Fix Version/s: 3.5.1
   4.0.0
   3.4.2
   Resolution: Fixed

Issue resolved by pull request 43229
[https://github.com/apache/spark/pull/43229]

> PySpark: DataFrame.cache docstring gives wrong storage level
> 
>
> Key: SPARK-40154
> URL: https://issues.apache.org/jira/browse/SPARK-40154
> Project: Spark
>  Issue Type: Bug
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Paul Staab
>Assignee: Paul Staab
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.5.1, 4.0.0, 3.4.2
>
>
> The docstring of the `DataFrame.cache()` method currently states that it uses 
> a serialized storage level
> {code:java}
> Persists the :class:`DataFrame` with the default storage level 
> (`MEMORY_AND_DISK`).
> [...]
> -The default storage level has changed to `MEMORY_AND_DISK` to match 
> Scala in 2.0.{code}
> while `DataFrame.persist()` states that it uses a deserialized storage level
> {code:java}
> If no storage level is specified defaults to (`MEMORY_AND_DISK_DESER`)
> [...]
> The default storage level has changed to `MEMORY_AND_DISK_DESER` to match 
> Scala in 3.0.{code}
>  
> However, in practice both `.cache()` and `.persist()` use deserialized 
> storage levels:
> {code:java}
> import pyspark
> from pyspark.sql import SparkSession
> from pyspark import StorageLevel
> print(pyspark.__version__)
> # 3.3.0
> spark = SparkSession.builder.master("local[2]").getOrCreate()
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.cache()
> df.count()
> # Storage level in Spark UI: "Disk Memory Deserialized 1x Replicated"
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.persist()
> df.count()
> # Storage level in Spark UI: "Disk Memory Deserialized 1x Replicated"
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.persist(StorageLevel.MEMORY_AND_DISK)
> df.count()
> # Storage level in Spark UI: "Disk Memory Serialized 1x Replicated"{code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-40154) PySpark: DataFrame.cache docstring gives wrong storage level

2023-10-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-40154?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-40154:


Assignee: Paul Staab

> PySpark: DataFrame.cache docstring gives wrong storage level
> 
>
> Key: SPARK-40154
> URL: https://issues.apache.org/jira/browse/SPARK-40154
> Project: Spark
>  Issue Type: Bug
>  Components: PySpark
>Affects Versions: 3.3.0
>Reporter: Paul Staab
>Assignee: Paul Staab
>Priority: Minor
>  Labels: pull-request-available
>
> The docstring of the `DataFrame.cache()` method currently states that it uses 
> a serialized storage level
> {code:java}
> Persists the :class:`DataFrame` with the default storage level 
> (`MEMORY_AND_DISK`).
> [...]
> -The default storage level has changed to `MEMORY_AND_DISK` to match 
> Scala in 2.0.{code}
> while `DataFrame.persist()` states that it uses a deserialized storage level
> {code:java}
> If no storage level is specified defaults to (`MEMORY_AND_DISK_DESER`)
> [...]
> The default storage level has changed to `MEMORY_AND_DISK_DESER` to match 
> Scala in 3.0.{code}
>  
> However, in practice both `.cache()` and `.persist()` use deserialized 
> storage levels:
> {code:java}
> import pyspark
> from pyspark.sql import SparkSession
> from pyspark import StorageLevel
> print(pyspark.__version__)
> # 3.3.0
> spark = SparkSession.builder.master("local[2]").getOrCreate()
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.cache()
> df.count()
> # Storage level in Spark UI: "Disk Memory Deserialized 1x Replicated"
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.persist()
> df.count()
> # Storage level in Spark UI: "Disk Memory Deserialized 1x Replicated"
> df = spark.createDataFrame(zip(["A"] * 1000, ["B"] * 1000), ["col_a", 
> "col_b"])
> df = df.persist(StorageLevel.MEMORY_AND_DISK)
> df.count()
> # Storage level in Spark UI: "Disk Memory Serialized 1x Replicated"{code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45610) Fix "Auto-application to `()` is deprecated."

2023-10-22 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45610.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43472
[https://github.com/apache/spark/pull/43472]

> Fix "Auto-application to `()` is deprecated."
> -
>
> Key: SPARK-45610
> URL: https://issues.apache.org/jira/browse/SPARK-45610
> Project: Spark
>  Issue Type: Sub-task
>  Components: GraphX, MLlib, Spark Core, SQL, Structured Streaming
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> For the following case, a compile warning will be issued in Scala 2.13:
>  
> {code:java}
> Welcome to Scala 2.13.12 (OpenJDK 64-Bit Server VM, Java 17.0.8).
> Type in expressions for evaluation. Or try :help.
> scala> class Foo {
>      |     def isEmpty(): Boolean = true
>      |     def isTrue(x: Boolean): Boolean = x
>      |   }
> class Foo
> scala> val foo = new Foo
> val foo: Foo = Foo@7061622
> scala> val ret = foo.isEmpty
>                      ^
>        warning: Auto-application to `()` is deprecated. Supply the empty 
> argument list `()` explicitly to invoke method isEmpty,
>        or remove the empty argument list from its definition (Java-defined 
> methods are exempt).
>        In Scala 3, an unapplied method like this will be eta-expanded into a 
> function. [quickfixable]
> val ret: Boolean = true {code}
> But for Scala 3, it is a compile error:
> {code:java}
> Welcome to Scala 3.3.1 (17.0.8, Java OpenJDK 64-Bit Server VM).
> Type in expressions for evaluation. Or try :help.
>                                                                               
>                                                                               
>                                                                               
>            
> scala> class Foo {
>      |     def isEmpty(): Boolean = true
>      |     def isTrue(x: Boolean): Boolean = x
>      |   }
> // defined class Foo
>                                                                               
>                                                                               
>                                                                               
>            
> scala> val foo = new Foo
> val foo: Foo = Foo@591f6f83
>                                                                               
>                                                                               
>                                                                               
>            
> scala> val ret = foo.isEmpty
> -- [E100] Syntax Error: 
> 
> 1 |val ret = foo.isEmpty
>   |          ^^^
>   |          method isEmpty in class Foo must be called with () argument
>   |
>   | longer explanation available when compiling with `-explain`
> 1 error found {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45610) Fix "Auto-application to `()` is deprecated."

2023-10-22 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45610:


Assignee: Yang Jie

> Fix "Auto-application to `()` is deprecated."
> -
>
> Key: SPARK-45610
> URL: https://issues.apache.org/jira/browse/SPARK-45610
> Project: Spark
>  Issue Type: Sub-task
>  Components: GraphX, MLlib, Spark Core, SQL, Structured Streaming
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> For the following case, a compile warning will be issued in Scala 2.13:
>  
> {code:java}
> Welcome to Scala 2.13.12 (OpenJDK 64-Bit Server VM, Java 17.0.8).
> Type in expressions for evaluation. Or try :help.
> scala> class Foo {
>      |     def isEmpty(): Boolean = true
>      |     def isTrue(x: Boolean): Boolean = x
>      |   }
> class Foo
> scala> val foo = new Foo
> val foo: Foo = Foo@7061622
> scala> val ret = foo.isEmpty
>                      ^
>        warning: Auto-application to `()` is deprecated. Supply the empty 
> argument list `()` explicitly to invoke method isEmpty,
>        or remove the empty argument list from its definition (Java-defined 
> methods are exempt).
>        In Scala 3, an unapplied method like this will be eta-expanded into a 
> function. [quickfixable]
> val ret: Boolean = true {code}
> But for Scala 3, it is a compile error:
> {code:java}
> Welcome to Scala 3.3.1 (17.0.8, Java OpenJDK 64-Bit Server VM).
> Type in expressions for evaluation. Or try :help.
>                                                                               
>                                                                               
>                                                                               
>            
> scala> class Foo {
>      |     def isEmpty(): Boolean = true
>      |     def isTrue(x: Boolean): Boolean = x
>      |   }
> // defined class Foo
>                                                                               
>                                                                               
>                                                                               
>            
> scala> val foo = new Foo
> val foo: Foo = Foo@591f6f83
>                                                                               
>                                                                               
>                                                                               
>            
> scala> val ret = foo.isEmpty
> -- [E100] Syntax Error: 
> 
> 1 |val ret = foo.isEmpty
>   |          ^^^
>   |          method isEmpty in class Foo must be called with () argument
>   |
>   | longer explanation available when compiling with `-explain`
> 1 error found {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-45610) Handle "Auto-application to `()` is deprecated."

2023-10-19 Thread Sean R. Owen (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-45610?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1405#comment-1405
 ] 

Sean R. Owen commented on SPARK-45610:
--

I think it's better to make big changes at major version boundaries. I'd expect 
we support Scala 3 at some point for Spark 4.x. Therefore I think it'd be OK to 
proceed with these changes now for 4.0.

> Handle "Auto-application to `()` is deprecated."
> 
>
> Key: SPARK-45610
> URL: https://issues.apache.org/jira/browse/SPARK-45610
> Project: Spark
>  Issue Type: Sub-task
>  Components: GraphX, MLlib, Spark Core, SQL, Structured Streaming
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Priority: Major
>
> For the following case, a compile warning will be issued in Scala 2.13:
>  
> {code:java}
> Welcome to Scala 2.13.12 (OpenJDK 64-Bit Server VM, Java 17.0.8).
> Type in expressions for evaluation. Or try :help.
> scala> class Foo {
>      |     def isEmpty(): Boolean = true
>      |     def isTrue(x: Boolean): Boolean = x
>      |   }
> class Foo
> scala> val foo = new Foo
> val foo: Foo = Foo@7061622
> scala> val ret = foo.isEmpty
>                      ^
>        warning: Auto-application to `()` is deprecated. Supply the empty 
> argument list `()` explicitly to invoke method isEmpty,
>        or remove the empty argument list from its definition (Java-defined 
> methods are exempt).
>        In Scala 3, an unapplied method like this will be eta-expanded into a 
> function. [quickfixable]
> val ret: Boolean = true {code}
> But for Scala 3, it is a compile error:
> {code:java}
> Welcome to Scala 3.3.1 (17.0.8, Java OpenJDK 64-Bit Server VM).
> Type in expressions for evaluation. Or try :help.
>                                                                               
>                                                                               
>                                                                               
>            
> scala> class Foo {
>      |     def isEmpty(): Boolean = true
>      |     def isTrue(x: Boolean): Boolean = x
>      |   }
> // defined class Foo
>                                                                               
>                                                                               
>                                                                               
>            
> scala> val foo = new Foo
> val foo: Foo = Foo@591f6f83
>                                                                               
>                                                                               
>                                                                               
>            
> scala> val ret = foo.isEmpty
> -- [E100] Syntax Error: 
> 
> 1 |val ret = foo.isEmpty
>   |          ^^^
>   |          method isEmpty in class Foo must be called with () argument
>   |
>   | longer explanation available when compiling with `-explain`
> 1 error found {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45588) Minor scaladoc improvement in StreamingForeachBatchHelper

2023-10-18 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45588:
-
Issue Type: Improvement  (was: Bug)
  Priority: Trivial  (was: Major)

> Minor scaladoc improvement in StreamingForeachBatchHelper
> -
>
> Key: SPARK-45588
> URL: https://issues.apache.org/jira/browse/SPARK-45588
> Project: Spark
>  Issue Type: Improvement
>  Components: Connect, Structured Streaming
>Affects Versions: 4.0.0
>Reporter: Raghu Angadi
>Priority: Trivial
>  Labels: pull-request-available
>
> Document RunnerCleaner in StreamingForeachBatchHelper.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45549) Remove unused `numExistingExecutors` in `CoarseGrainedSchedulerBackend`

2023-10-18 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45549:
-
Priority: Trivial  (was: Minor)

> Remove unused `numExistingExecutors` in `CoarseGrainedSchedulerBackend`
> ---
>
> Key: SPARK-45549
> URL: https://issues.apache.org/jira/browse/SPARK-45549
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: xiaoping.huang
>Priority: Trivial
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45564) Simplify 'DataFrameStatFunctions.bloomFilter' function with 'BloomFilterAggregate' expression

2023-10-17 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45564:
-
Priority: Minor  (was: Major)

> Simplify 'DataFrameStatFunctions.bloomFilter' function with 
> 'BloomFilterAggregate' expression
> -
>
> Key: SPARK-45564
> URL: https://issues.apache.org/jira/browse/SPARK-45564
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Ruifeng Zheng
>Assignee: Ruifeng Zheng
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45564) Simplify 'DataFrameStatFunctions.bloomFilter' function with 'BloomFilterAggregate' expression

2023-10-17 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45564.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43391
[https://github.com/apache/spark/pull/43391]

> Simplify 'DataFrameStatFunctions.bloomFilter' function with 
> 'BloomFilterAggregate' expression
> -
>
> Key: SPARK-45564
> URL: https://issues.apache.org/jira/browse/SPARK-45564
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Ruifeng Zheng
>Assignee: Ruifeng Zheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45564) Simplify 'DataFrameStatFunctions.bloomFilter' function with 'BloomFilterAggregate' expression

2023-10-17 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45564?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45564:


Assignee: Ruifeng Zheng

> Simplify 'DataFrameStatFunctions.bloomFilter' function with 
> 'BloomFilterAggregate' expression
> -
>
> Key: SPARK-45564
> URL: https://issues.apache.org/jira/browse/SPARK-45564
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 4.0.0
>Reporter: Ruifeng Zheng
>Assignee: Ruifeng Zheng
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45512) Fix compilation warnings related to other-nullary-override

2023-10-17 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45512.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43332
[https://github.com/apache/spark/pull/43332]

> Fix compilation warnings related to other-nullary-override
> --
>
> Key: SPARK-45512
> URL: https://issues.apache.org/jira/browse/SPARK-45512
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala:36:16:
>  method with a single empty parameter list overrides method hasNext in trait 
> Iterator defined without a parameter list [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.WrappedCloseableIterator
> [error]   override def hasNext(): Boolean = innerIterator.hasNext
> [error]                ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala:136:16:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.ExecutePlanResponseReattachableIterator
> [error]   override def hasNext: Boolean = synchronized {
> [error]                ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcExceptionConverter.scala:73:20:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcExceptionConverter.convertIterator
> [error]       override def hasNext: Boolean = {
> [error]                    ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:77:18:
>  method without a parameter list overrides method next in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
> [error]     override def next: U = {
> [error]                  ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:81:18:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
> [error]     override def hasNext: Boolean = {
> [error]                  ^
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45512) Fix compilation warnings related to other-nullary-override

2023-10-17 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45512:
-
Priority: Minor  (was: Major)

> Fix compilation warnings related to other-nullary-override
> --
>
> Key: SPARK-45512
> URL: https://issues.apache.org/jira/browse/SPARK-45512
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala:36:16:
>  method with a single empty parameter list overrides method hasNext in trait 
> Iterator defined without a parameter list [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.WrappedCloseableIterator
> [error]   override def hasNext(): Boolean = innerIterator.hasNext
> [error]                ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala:136:16:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.ExecutePlanResponseReattachableIterator
> [error]   override def hasNext: Boolean = synchronized {
> [error]                ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcExceptionConverter.scala:73:20:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcExceptionConverter.convertIterator
> [error]       override def hasNext: Boolean = {
> [error]                    ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:77:18:
>  method without a parameter list overrides method next in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
> [error]     override def next: U = {
> [error]                  ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:81:18:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
> [error]     override def hasNext: Boolean = {
> [error]                  ^
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45512) Fix compilation warnings related to other-nullary-override

2023-10-17 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45512:


Assignee: Yang Jie

> Fix compilation warnings related to other-nullary-override
> --
>
> Key: SPARK-45512
> URL: https://issues.apache.org/jira/browse/SPARK-45512
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/CloseableIterator.scala:36:16:
>  method with a single empty parameter list overrides method hasNext in trait 
> Iterator defined without a parameter list [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.WrappedCloseableIterator
> [error]   override def hasNext(): Boolean = innerIterator.hasNext
> [error]                ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/ExecutePlanResponseReattachableIterator.scala:136:16:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.ExecutePlanResponseReattachableIterator
> [error]   override def hasNext: Boolean = synchronized {
> [error]                ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcExceptionConverter.scala:73:20:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcExceptionConverter.convertIterator
> [error]       override def hasNext: Boolean = {
> [error]                    ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:77:18:
>  method without a parameter list overrides method next in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
> [error]     override def next: U = {
> [error]                  ^
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/connector/connect/common/src/main/scala/org/apache/spark/sql/connect/client/GrpcRetryHandler.scala:81:18:
>  method without a parameter list overrides method hasNext in class 
> WrappedCloseableIterator defined with a single empty parameter list 
> [quickfixable]
> [error] Applicable -Wconf / @nowarn filters for this fatal warning: msg= of the message>, cat=other-nullary-override, 
> site=org.apache.spark.sql.connect.client.GrpcRetryHandler.RetryIterator
> [error]     override def hasNext: Boolean = {
> [error]                  ^
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45273) Http header Attack【HttpSecurityFilter】

2023-10-16 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45273.
--
Resolution: Invalid

> Http header Attack【HttpSecurityFilter】
> --
>
> Key: SPARK-45273
> URL: https://issues.apache.org/jira/browse/SPARK-45273
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: chenyu
>Priority: Major
>  Labels: pull-request-available
>
> There is an HTTP host header attack vulnerability in the target URL



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45467) Replace `Proxy.getProxyClass()` with `Proxy.newProxyInstance().getClass`

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45467.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43291
[https://github.com/apache/spark/pull/43291]

> Replace `Proxy.getProxyClass()` with `Proxy.newProxyInstance().getClass`
> 
>
> Key: SPARK-45467
> URL: https://issues.apache.org/jira/browse/SPARK-45467
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
>  * @deprecated Proxy classes generated in a named module are encapsulated
>  *  and not accessible to code outside its module.
>  *  {@link Constructor#newInstance(Object...) Constructor.newInstance}
>  *  will throw {@code IllegalAccessException} when it is called on
>  *  an inaccessible proxy class.
>  *  Use {@link #newProxyInstance(ClassLoader, Class[], InvocationHandler)}
>  *  to create a proxy instance instead.
>  *
>  * @see Package and Module Membership of Proxy Class
>  * @revised 9
>  */
> @Deprecated
> @CallerSensitive
> public static Class getProxyClass(ClassLoader loader,
>  Class... interfaces)
> throws IllegalArgumentException {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45467) Replace `Proxy.getProxyClass()` with `Proxy.newProxyInstance().getClass`

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45467:


Assignee: Yang Jie

> Replace `Proxy.getProxyClass()` with `Proxy.newProxyInstance().getClass`
> 
>
> Key: SPARK-45467
> URL: https://issues.apache.org/jira/browse/SPARK-45467
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
>  * @deprecated Proxy classes generated in a named module are encapsulated
>  *  and not accessible to code outside its module.
>  *  {@link Constructor#newInstance(Object...) Constructor.newInstance}
>  *  will throw {@code IllegalAccessException} when it is called on
>  *  an inaccessible proxy class.
>  *  Use {@link #newProxyInstance(ClassLoader, Class[], InvocationHandler)}
>  *  to create a proxy instance instead.
>  *
>  * @see Package and Module Membership of Proxy Class
>  * @revised 9
>  */
> @Deprecated
> @CallerSensitive
> public static Class getProxyClass(ClassLoader loader,
>  Class... interfaces)
> throws IllegalArgumentException {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45467) Replace `Proxy.getProxyClass()` with `Proxy.newProxyInstance().getClass`

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45467:
-
Priority: Minor  (was: Major)

> Replace `Proxy.getProxyClass()` with `Proxy.newProxyInstance().getClass`
> 
>
> Key: SPARK-45467
> URL: https://issues.apache.org/jira/browse/SPARK-45467
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
>  * @deprecated Proxy classes generated in a named module are encapsulated
>  *  and not accessible to code outside its module.
>  *  {@link Constructor#newInstance(Object...) Constructor.newInstance}
>  *  will throw {@code IllegalAccessException} when it is called on
>  *  an inaccessible proxy class.
>  *  Use {@link #newProxyInstance(ClassLoader, Class[], InvocationHandler)}
>  *  to create a proxy instance instead.
>  *
>  * @see Package and Module Membership of Proxy Class
>  * @revised 9
>  */
> @Deprecated
> @CallerSensitive
> public static Class getProxyClass(ClassLoader loader,
>  Class... interfaces)
> throws IllegalArgumentException {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45496) Fix the compilation warning related to other-pure-statement

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45496.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43312
[https://github.com/apache/spark/pull/43312]

> Fix the compilation warning related to other-pure-statement
> ---
>
> Key: SPARK-45496
> URL: https://issues.apache.org/jira/browse/SPARK-45496
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Spark Core
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> "-Wconf:cat=other-match-analysis=org.apache.spark.sql.catalyst.catalog.SessionCatalog.lookupFunction.catalogFunction:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.streaming.util.FileBasedWriteAheadLog.readAll.readFile:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.scheduler.OutputCommitCoordinatorSuite:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.sql.streaming.sources.StreamingDataSourceV2Suite.testPositiveCase.\\$anonfun:wv",
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45496) Fix the compilation warning related to other-pure-statement

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45496:
-
Priority: Minor  (was: Major)

> Fix the compilation warning related to other-pure-statement
> ---
>
> Key: SPARK-45496
> URL: https://issues.apache.org/jira/browse/SPARK-45496
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Spark Core
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
>
> {code:java}
> "-Wconf:cat=other-match-analysis=org.apache.spark.sql.catalyst.catalog.SessionCatalog.lookupFunction.catalogFunction:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.streaming.util.FileBasedWriteAheadLog.readAll.readFile:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.scheduler.OutputCommitCoordinatorSuite:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.sql.streaming.sources.StreamingDataSourceV2Suite.testPositiveCase.\\$anonfun:wv",
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-43828) Add config to control whether close idle connection

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-43828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-43828.
--
Resolution: Won't Fix

> Add config to control whether close idle connection
> ---
>
> Key: SPARK-43828
> URL: https://issues.apache.org/jira/browse/SPARK-43828
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 3.4.0
>Reporter: Zhongwei Zhu
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45496) Fix the compilation warning related to other-pure-statement

2023-10-11 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45496?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45496:


Assignee: Yang Jie

> Fix the compilation warning related to other-pure-statement
> ---
>
> Key: SPARK-45496
> URL: https://issues.apache.org/jira/browse/SPARK-45496
> Project: Spark
>  Issue Type: Sub-task
>  Components: DStreams, Spark Core
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> "-Wconf:cat=other-match-analysis=org.apache.spark.sql.catalyst.catalog.SessionCatalog.lookupFunction.catalogFunction:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.streaming.util.FileBasedWriteAheadLog.readAll.readFile:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.scheduler.OutputCommitCoordinatorSuite:wv",
> "-Wconf:cat=other-pure-statement=org.apache.spark.sql.streaming.sources.StreamingDataSourceV2Suite.testPositiveCase.\\$anonfun:wv",
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45377) [CORE] Handle InputStream in NettyLogger

2023-10-02 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45377:
-
Priority: Minor  (was: Major)

> [CORE] Handle InputStream in NettyLogger
> 
>
> Key: SPARK-45377
> URL: https://issues.apache.org/jira/browse/SPARK-45377
> Project: Spark
>  Issue Type: Task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Hasnain Lakhani
>Assignee: Hasnain Lakhani
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> Allow NettyLogger to also print the size of InputStreams which aids debugging 
> for SSL functionality



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45377) [CORE] Handle InputStream in NettyLogger

2023-10-02 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45377.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43165
[https://github.com/apache/spark/pull/43165]

> [CORE] Handle InputStream in NettyLogger
> 
>
> Key: SPARK-45377
> URL: https://issues.apache.org/jira/browse/SPARK-45377
> Project: Spark
>  Issue Type: Task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Hasnain Lakhani
>Assignee: Hasnain Lakhani
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> Allow NettyLogger to also print the size of InputStreams which aids debugging 
> for SSL functionality



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45377) [CORE] Handle InputStream in NettyLogger

2023-10-02 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45377?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45377:


Assignee: Hasnain Lakhani

> [CORE] Handle InputStream in NettyLogger
> 
>
> Key: SPARK-45377
> URL: https://issues.apache.org/jira/browse/SPARK-45377
> Project: Spark
>  Issue Type: Task
>  Components: Spark Core
>Affects Versions: 4.0.0
>Reporter: Hasnain Lakhani
>Assignee: Hasnain Lakhani
>Priority: Major
>  Labels: pull-request-available
>
> Allow NettyLogger to also print the size of InputStreams which aids debugging 
> for SSL functionality



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45389) Correct MetaException matching rule on getting partition metadata

2023-10-02 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45389.
--
Fix Version/s: 4.0.0
 Assignee: Cheng Pan
   Resolution: Fixed

https://github.com/apache/spark/pull/43191

> Correct MetaException matching rule on getting partition metadata
> -
>
> Key: SPARK-45389
> URL: https://issues.apache.org/jira/browse/SPARK-45389
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.3.3
>Reporter: Cheng Pan
>Assignee: Cheng Pan
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-44895) Considering 'daemon', 'priority' from higher JDKs for ThreadStackTrace class

2023-09-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-44895:
-
Priority: Minor  (was: Major)

> Considering 'daemon', 'priority' from higher JDKs for ThreadStackTrace class
> 
>
> Key: SPARK-44895
> URL: https://issues.apache.org/jira/browse/SPARK-44895
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 4.0.0
>Reporter: Kent Yao
>Assignee: Kent Yao
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> jshell> var t = java.lang.management.ManagementFactory.getThreadMXBean()t ==> 
> com.sun.management.internal.HotSpotThreadImpl@7daf6ecc
> jshell> var tt = t.dumpAllThreads(true, true)tt ==> ThreadInfo[10] { "main" 
> prio=5 Id=1 RUNNABLE  at  ... k$NonfairSync@27fa135a
>  }
> jshell> for (java.lang.management.ThreadInfo t1: tt) 
> {System.out.println(t1.toString());}"main" prio=5 Id=1 RUNNABLE  at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpThreads0(Native Method) 
> at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpAllThreads(ThreadImpl.java:540)
>  at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpAllThreads(ThreadImpl.java:527)
>  at REPL.$JShell$12.do_it$Aux($JShell$12.java:7) at 
> REPL.$JShell$12.do_it$($JShell$12.java:11)   at 
> java.base@20.0.1/java.lang.invoke.DirectMethodHandle$Holder.invokeStatic(DirectMethodHandle$Holder)
>   at 
> java.base@20.0.1/java.lang.invoke.LambdaForm$MH/0x007001008c00.invoke(LambdaForm$MH)
>  at 
> java.base@20.0.1/java.lang.invoke.Invokers$Holder.invokeExact_MT(Invokers$Holder)
> ...
> "Reference Handler" daemon prio=10 Id=8 RUNNABLE  at 
> java.base@20.0.1/java.lang.ref.Reference.waitForReferencePendingList(Native 
> Method)  at 
> java.base@20.0.1/java.lang.ref.Reference.processPendingReferences(Reference.java:246)
> at 
> java.base@20.0.1/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:208)
>  {code}
> the `daemon prio=10` is not available for ThreadInfo of jdk8
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-44895) Considering 'daemon', 'priority' from higher JDKs for ThreadStackTrace class

2023-09-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-44895.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43095
[https://github.com/apache/spark/pull/43095]

> Considering 'daemon', 'priority' from higher JDKs for ThreadStackTrace class
> 
>
> Key: SPARK-44895
> URL: https://issues.apache.org/jira/browse/SPARK-44895
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 4.0.0
>Reporter: Kent Yao
>Assignee: Kent Yao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> jshell> var t = java.lang.management.ManagementFactory.getThreadMXBean()t ==> 
> com.sun.management.internal.HotSpotThreadImpl@7daf6ecc
> jshell> var tt = t.dumpAllThreads(true, true)tt ==> ThreadInfo[10] { "main" 
> prio=5 Id=1 RUNNABLE  at  ... k$NonfairSync@27fa135a
>  }
> jshell> for (java.lang.management.ThreadInfo t1: tt) 
> {System.out.println(t1.toString());}"main" prio=5 Id=1 RUNNABLE  at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpThreads0(Native Method) 
> at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpAllThreads(ThreadImpl.java:540)
>  at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpAllThreads(ThreadImpl.java:527)
>  at REPL.$JShell$12.do_it$Aux($JShell$12.java:7) at 
> REPL.$JShell$12.do_it$($JShell$12.java:11)   at 
> java.base@20.0.1/java.lang.invoke.DirectMethodHandle$Holder.invokeStatic(DirectMethodHandle$Holder)
>   at 
> java.base@20.0.1/java.lang.invoke.LambdaForm$MH/0x007001008c00.invoke(LambdaForm$MH)
>  at 
> java.base@20.0.1/java.lang.invoke.Invokers$Holder.invokeExact_MT(Invokers$Holder)
> ...
> "Reference Handler" daemon prio=10 Id=8 RUNNABLE  at 
> java.base@20.0.1/java.lang.ref.Reference.waitForReferencePendingList(Native 
> Method)  at 
> java.base@20.0.1/java.lang.ref.Reference.processPendingReferences(Reference.java:246)
> at 
> java.base@20.0.1/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:208)
>  {code}
> the `daemon prio=10` is not available for ThreadInfo of jdk8
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-44895) Considering 'daemon', 'priority' from higher JDKs for ThreadStackTrace class

2023-09-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44895?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-44895:


Assignee: Kent Yao

> Considering 'daemon', 'priority' from higher JDKs for ThreadStackTrace class
> 
>
> Key: SPARK-44895
> URL: https://issues.apache.org/jira/browse/SPARK-44895
> Project: Spark
>  Issue Type: Sub-task
>  Components: Web UI
>Affects Versions: 4.0.0
>Reporter: Kent Yao
>Assignee: Kent Yao
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> jshell> var t = java.lang.management.ManagementFactory.getThreadMXBean()t ==> 
> com.sun.management.internal.HotSpotThreadImpl@7daf6ecc
> jshell> var tt = t.dumpAllThreads(true, true)tt ==> ThreadInfo[10] { "main" 
> prio=5 Id=1 RUNNABLE  at  ... k$NonfairSync@27fa135a
>  }
> jshell> for (java.lang.management.ThreadInfo t1: tt) 
> {System.out.println(t1.toString());}"main" prio=5 Id=1 RUNNABLE  at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpThreads0(Native Method) 
> at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpAllThreads(ThreadImpl.java:540)
>  at 
> java.management@20.0.1/sun.management.ThreadImpl.dumpAllThreads(ThreadImpl.java:527)
>  at REPL.$JShell$12.do_it$Aux($JShell$12.java:7) at 
> REPL.$JShell$12.do_it$($JShell$12.java:11)   at 
> java.base@20.0.1/java.lang.invoke.DirectMethodHandle$Holder.invokeStatic(DirectMethodHandle$Holder)
>   at 
> java.base@20.0.1/java.lang.invoke.LambdaForm$MH/0x007001008c00.invoke(LambdaForm$MH)
>  at 
> java.base@20.0.1/java.lang.invoke.Invokers$Holder.invokeExact_MT(Invokers$Holder)
> ...
> "Reference Handler" daemon prio=10 Id=8 RUNNABLE  at 
> java.base@20.0.1/java.lang.ref.Reference.waitForReferencePendingList(Native 
> Method)  at 
> java.base@20.0.1/java.lang.ref.Reference.processPendingReferences(Reference.java:246)
> at 
> java.base@20.0.1/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:208)
>  {code}
> the `daemon prio=10` is not available for ThreadInfo of jdk8
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45364) Clean up the unnecessary Scala 2.12 logical in SparkBuild

2023-09-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45364:


Assignee: BingKun Pan

> Clean up the unnecessary Scala 2.12 logical in SparkBuild
> -
>
> Key: SPARK-45364
> URL: https://issues.apache.org/jira/browse/SPARK-45364
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build, Project Infra
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45364) Clean up the unnecessary Scala 2.12 logical in SparkBuild

2023-09-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45364:
-
Priority: Trivial  (was: Minor)

> Clean up the unnecessary Scala 2.12 logical in SparkBuild
> -
>
> Key: SPARK-45364
> URL: https://issues.apache.org/jira/browse/SPARK-45364
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build, Project Infra
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45364) Clean up the unnecessary Scala 2.12 logical in SparkBuild

2023-09-28 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45364?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45364.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43158
[https://github.com/apache/spark/pull/43158]

> Clean up the unnecessary Scala 2.12 logical in SparkBuild
> -
>
> Key: SPARK-45364
> URL: https://issues.apache.org/jira/browse/SPARK-45364
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build, Project Infra
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-44442) Drop mesos support

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-2?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-2:


Assignee: Sean R. Owen

> Drop mesos support
> --
>
> Key: SPARK-2
> URL: https://issues.apache.org/jira/browse/SPARK-2
> Project: Spark
>  Issue Type: Sub-task
>  Components: Mesos
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Sean R. Owen
>Priority: Major
>  Labels: pull-request-available
>
> [https://spark.apache.org/docs/latest/running-on-mesos.html]
>  
> {_}Note{_}: Apache Mesos support is deprecated as of Apache Spark 3.2.0. It 
> will be removed in a future version.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-44442) Drop mesos support

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-2?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-2.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43135
[https://github.com/apache/spark/pull/43135]

> Drop mesos support
> --
>
> Key: SPARK-2
> URL: https://issues.apache.org/jira/browse/SPARK-2
> Project: Spark
>  Issue Type: Sub-task
>  Components: Mesos
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Sean R. Owen
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> [https://spark.apache.org/docs/latest/running-on-mesos.html]
>  
> {_}Note{_}: Apache Mesos support is deprecated as of Apache Spark 3.2.0. It 
> will be removed in a future version.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-44539) Upgrade RoaringBitmap to 1.0.0

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-44539:
-
Priority: Minor  (was: Trivial)

>  Upgrade RoaringBitmap to 1.0.0
> ---
>
> Key: SPARK-44539
> URL: https://issues.apache.org/jira/browse/SPARK-44539
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-44539) Upgrade RoaringBitmap to 1.0.0

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-44539:


Assignee: BingKun Pan

>  Upgrade RoaringBitmap to 1.0.0
> ---
>
> Key: SPARK-44539
> URL: https://issues.apache.org/jira/browse/SPARK-44539
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Trivial
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-44539) Upgrade RoaringBitmap to 1.0.0

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-44539.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 42143
[https://github.com/apache/spark/pull/42143]

>  Upgrade RoaringBitmap to 1.0.0
> ---
>
> Key: SPARK-44539
> URL: https://issues.apache.org/jira/browse/SPARK-44539
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: BingKun Pan
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45343) CSV multiLine documentation is confusing

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45343:


Assignee: Bill Schneider

> CSV multiLine documentation is confusing
> 
>
> Key: SPARK-45343
> URL: https://issues.apache.org/jira/browse/SPARK-45343
> Project: Spark
>  Issue Type: Documentation
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: Bill Schneider
>Assignee: Bill Schneider
>Priority: Trivial
>  Labels: pull-request-available
>
> This is confusing, maybe copy-paste from JSON:
> |Parse one record, which may span multiple lines, per file. CSV built-in 
> functions ignore this option.|
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45343) CSV multiLine documentation is confusing

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45343.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43132
[https://github.com/apache/spark/pull/43132]

> CSV multiLine documentation is confusing
> 
>
> Key: SPARK-45343
> URL: https://issues.apache.org/jira/browse/SPARK-45343
> Project: Spark
>  Issue Type: Documentation
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: Bill Schneider
>Assignee: Bill Schneider
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> This is confusing, maybe copy-paste from JSON:
> |Parse one record, which may span multiple lines, per file. CSV built-in 
> functions ignore this option.|
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-44170) Migrating Junit4 to Junit5

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-44170:


Assignee: Yang Jie

> Migrating Junit4 to Junit5
> --
>
> Key: SPARK-44170
> URL: https://issues.apache.org/jira/browse/SPARK-44170
> Project: Spark
>  Issue Type: Sub-task
>  Components: Tests
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> JUnit5 is a powerful and flexible update to the JUnit framework, and it 
> provides a variety of improvements and new features to organize and
> describe test cases, as well as help in understanding test results:
>  # JUnit 5 leverages features from Java 8 or later, such as lambda functions, 
> making tests more powerful and easier to maintain, but Junit 4 still a Java 7 
> compatible version
>  # JUnit 5 has added some useful new features for describing, organizing, and 
> executing tests. For examples: [Parameterized 
> Tests|https://junit.org/junit5/docs/current/user-guide/#writing-tests-parameterized-tests]
>  and [Conditional Test 
> Execution|https://junit.org/junit5/docs/current/user-guide/#extensions-conditions]
>  may make our test code look simpler, [Parallel 
> Execution|https://junit.org/junit5/docs/current/user-guide/#writing-tests-parallel-execution]
>  may make our test faster
>  
> More importantly, Junit4 is currently an inactive project, which has not 
> released a new version for more than two years
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-44170) Migrating Junit4 to Junit5

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-44170.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43074
[https://github.com/apache/spark/pull/43074]

> Migrating Junit4 to Junit5
> --
>
> Key: SPARK-44170
> URL: https://issues.apache.org/jira/browse/SPARK-44170
> Project: Spark
>  Issue Type: Sub-task
>  Components: Tests
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> JUnit5 is a powerful and flexible update to the JUnit framework, and it 
> provides a variety of improvements and new features to organize and
> describe test cases, as well as help in understanding test results:
>  # JUnit 5 leverages features from Java 8 or later, such as lambda functions, 
> making tests more powerful and easier to maintain, but Junit 4 still a Java 7 
> compatible version
>  # JUnit 5 has added some useful new features for describing, organizing, and 
> executing tests. For examples: [Parameterized 
> Tests|https://junit.org/junit5/docs/current/user-guide/#writing-tests-parameterized-tests]
>  and [Conditional Test 
> Execution|https://junit.org/junit5/docs/current/user-guide/#extensions-conditions]
>  may make our test code look simpler, [Parallel 
> Execution|https://junit.org/junit5/docs/current/user-guide/#writing-tests-parallel-execution]
>  may make our test faster
>  
> More importantly, Junit4 is currently an inactive project, which has not 
> released a new version for more than two years
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45338) Remove scala.collection.JavaConverters

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45338?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45338:


Assignee: Jia Fan

> Remove scala.collection.JavaConverters
> --
>
> Key: SPARK-45338
> URL: https://issues.apache.org/jira/browse/SPARK-45338
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Jia Fan
>Assignee: Jia Fan
>Priority: Major
>  Labels: pull-request-available
>
> Remove deprecated scala.collection.JavaConverters, replaced by 
> scala.jdk.CollectionConverters



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45338) Remove scala.collection.JavaConverters

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45338?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45338.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43126
[https://github.com/apache/spark/pull/43126]

> Remove scala.collection.JavaConverters
> --
>
> Key: SPARK-45338
> URL: https://issues.apache.org/jira/browse/SPARK-45338
> Project: Spark
>  Issue Type: Sub-task
>  Components: Spark Core, SQL
>Affects Versions: 4.0.0
>Reporter: Jia Fan
>Assignee: Jia Fan
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> Remove deprecated scala.collection.JavaConverters, replaced by 
> scala.jdk.CollectionConverters



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45341) Make the sbt doc command execute successfully with Java 17

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45341:
-
Priority: Minor  (was: Major)

> Make the sbt doc command execute successfully with Java 17
> --
>
> Key: SPARK-45341
> URL: https://issues.apache.org/jira/browse/SPARK-45341
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> [error] /Users/yangjie01/SourceCode/git/spark-mine-sbt/Picked up 
> JAVA_TOOL_OPTIONS:-Duser.language=en
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDBTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/ArrayWrappers.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVIndex.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/InMemoryStore.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDBIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDB.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDBTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/UnsupportedStoreVersionException.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDB.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStore.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreView.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDBIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreSerializer.java...
> [error] Constructing Javadoc information...
> [error] Building index for all the packages and classes...
> [error] Standard Doclet version 17.0.8+7-LTS
> [error] Building tree for all the packages and classes...
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStore.java:32:1:
>   error: heading used out of sequence: , compared to implicit preceding 
> heading: 
> [error]  * Serialization
> [error]    ^Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/InMemoryStore.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVIndex.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStore.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreIterator.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreSerializer.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreView.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVTypeInfo.html...
> [error] Generating 
> 

[jira] [Resolved] (SPARK-45341) Make the sbt doc command execute successfully with Java 17

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45341.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43131
[https://github.com/apache/spark/pull/43131]

> Make the sbt doc command execute successfully with Java 17
> --
>
> Key: SPARK-45341
> URL: https://issues.apache.org/jira/browse/SPARK-45341
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> {code:java}
> [error] /Users/yangjie01/SourceCode/git/spark-mine-sbt/Picked up 
> JAVA_TOOL_OPTIONS:-Duser.language=en
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDBTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/ArrayWrappers.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVIndex.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/InMemoryStore.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDBIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDB.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDBTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/UnsupportedStoreVersionException.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDB.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStore.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreView.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDBIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreSerializer.java...
> [error] Constructing Javadoc information...
> [error] Building index for all the packages and classes...
> [error] Standard Doclet version 17.0.8+7-LTS
> [error] Building tree for all the packages and classes...
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStore.java:32:1:
>   error: heading used out of sequence: , compared to implicit preceding 
> heading: 
> [error]  * Serialization
> [error]    ^Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/InMemoryStore.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVIndex.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStore.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreIterator.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreSerializer.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreView.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVTypeInfo.html...
> [error] Generating 

[jira] [Assigned] (SPARK-45341) Make the sbt doc command execute successfully with Java 17

2023-09-27 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45341?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45341:


Assignee: Yang Jie

> Make the sbt doc command execute successfully with Java 17
> --
>
> Key: SPARK-45341
> URL: https://issues.apache.org/jira/browse/SPARK-45341
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Major
>  Labels: pull-request-available
>
> {code:java}
> [error] /Users/yangjie01/SourceCode/git/spark-mine-sbt/Picked up 
> JAVA_TOOL_OPTIONS:-Duser.language=en
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDBTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/ArrayWrappers.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVIndex.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/InMemoryStore.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDBIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDB.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDBTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/UnsupportedStoreVersionException.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/LevelDB.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStore.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreView.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVTypeInfo.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/RocksDBIterator.java...
> [error] Loading source file 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStoreSerializer.java...
> [error] Constructing Javadoc information...
> [error] Building index for all the packages and classes...
> [error] Standard Doclet version 17.0.8+7-LTS
> [error] Building tree for all the packages and classes...
> [error] 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/src/main/java/org/apache/spark/util/kvstore/KVStore.java:32:1:
>   error: heading used out of sequence: , compared to implicit preceding 
> heading: 
> [error]  * Serialization
> [error]    ^Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/InMemoryStore.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVIndex.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStore.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreIterator.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreSerializer.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVStoreView.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/KVTypeInfo.html...
> [error] Generating 
> /Users/yangjie01/SourceCode/git/spark-mine-sbt/common/kvstore/target/scala-2.13/api/org/apache/spark/util/kvstore/LevelDB.html...
> 

[jira] [Assigned] (SPARK-45334) Remove misleading comment in parquetSchemaConverter

2023-09-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45334:


Assignee: Mengran Lan

> Remove misleading comment in parquetSchemaConverter
> ---
>
> Key: SPARK-45334
> URL: https://issues.apache.org/jira/browse/SPARK-45334
> Project: Spark
>  Issue Type: Documentation
>  Components: SQL
>Affects Versions: 3.5.0
>Reporter: Mengran Lan
>Assignee: Mengran Lan
>Priority: Trivial
>  Labels: pull-request-available
>
> I'm debugging a parquet issue and reading spark code as references. Happened 
> to find a misleading comment which remains in the latest version as well.
> {code:java}
> Types
>   .buildGroup(repetition).as(LogicalTypeAnnotation.listType())
>   .addField(Types
> .buildGroup(REPEATED)
> // "array" is the name chosen by parquet-hive (1.7.0 and prior version)
> .addField(convertField(StructField("array", elementType, nullable)))
> .named("bag"))
>   .named(field.name) {code}
> the comment above is misleading since Hive always uses "array_element" as the 
> name.
> It is imported by this PR [https://github.com/apache/spark/pull/14399] and 
> relates to this issue https://issues.apache.org/jira/browse/SPARK-16777
> Furthermore, the parquet-hive module has been removed from the parquet-mr 
> project https://issues.apache.org/jira/browse/PARQUET-1676 
> I suggest removing this piece of comment and will submit a PR later.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45334) Remove misleading comment in parquetSchemaConverter

2023-09-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45334.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43119
[https://github.com/apache/spark/pull/43119]

> Remove misleading comment in parquetSchemaConverter
> ---
>
> Key: SPARK-45334
> URL: https://issues.apache.org/jira/browse/SPARK-45334
> Project: Spark
>  Issue Type: Documentation
>  Components: SQL
>Affects Versions: 3.5.0
>Reporter: Mengran Lan
>Assignee: Mengran Lan
>Priority: Trivial
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> I'm debugging a parquet issue and reading spark code as references. Happened 
> to find a misleading comment which remains in the latest version as well.
> {code:java}
> Types
>   .buildGroup(repetition).as(LogicalTypeAnnotation.listType())
>   .addField(Types
> .buildGroup(REPEATED)
> // "array" is the name chosen by parquet-hive (1.7.0 and prior version)
> .addField(convertField(StructField("array", elementType, nullable)))
> .named("bag"))
>   .named(field.name) {code}
> the comment above is misleading since Hive always uses "array_element" as the 
> name.
> It is imported by this PR [https://github.com/apache/spark/pull/14399] and 
> relates to this issue https://issues.apache.org/jira/browse/SPARK-16777
> Furthermore, the parquet-hive module has been removed from the parquet-mr 
> project https://issues.apache.org/jira/browse/PARQUET-1676 
> I suggest removing this piece of comment and will submit a PR later.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-44366) Migrate antlr4 from 4.9 to 4.10+

2023-09-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-44366.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43075
[https://github.com/apache/spark/pull/43075]

> Migrate antlr4 from 4.9 to 4.10+
> 
>
> Key: SPARK-44366
> URL: https://issues.apache.org/jira/browse/SPARK-44366
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-44366) Migrate antlr4 from 4.9 to 4.10+

2023-09-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-44366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-44366:


Assignee: Yang Jie

> Migrate antlr4 from 4.9 to 4.10+
> 
>
> Key: SPARK-44366
> URL: https://issues.apache.org/jira/browse/SPARK-44366
> Project: Spark
>  Issue Type: Sub-task
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: BingKun Pan
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45343) CSV multiLine documentation is confusing

2023-09-26 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45343:
-
Priority: Trivial  (was: Major)

> CSV multiLine documentation is confusing
> 
>
> Key: SPARK-45343
> URL: https://issues.apache.org/jira/browse/SPARK-45343
> Project: Spark
>  Issue Type: Documentation
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: Bill Schneider
>Priority: Trivial
>  Labels: pull-request-available
>
> This is confusing, maybe copy-paste from JSON:
> |Parse one record, which may span multiple lines, per file. CSV built-in 
> functions ignore this option.|
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45248) Slow HTTP Denial of Service Attack【jetty server in spark】

2023-09-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45248:
-
Issue Type: Improvement  (was: Bug)
  Priority: Minor  (was: Major)

> Slow HTTP Denial of Service Attack【jetty server in spark】
> -
>
> Key: SPARK-45248
> URL: https://issues.apache.org/jira/browse/SPARK-45248
> Project: Spark
>  Issue Type: Improvement
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: chenyu
>Assignee: chenyu
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> The Jetty server may be slow HTTP denial of service attacks in spark ui.
> The default value of connector IdleTimeout is 3.
> This could trigger such an attack



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45248) Slow HTTP Denial of Service Attack【jetty server in spark】

2023-09-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45248.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 43078
[https://github.com/apache/spark/pull/43078]

> Slow HTTP Denial of Service Attack【jetty server in spark】
> -
>
> Key: SPARK-45248
> URL: https://issues.apache.org/jira/browse/SPARK-45248
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: chenyu
>Assignee: chenyu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>
> The Jetty server may be slow HTTP denial of service attacks in spark ui.
> The default value of connector IdleTimeout is 3.
> This could trigger such an attack



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45248) Slow HTTP Denial of Service Attack【jetty server in spark】

2023-09-25 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45248?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45248:


Assignee: chenyu

> Slow HTTP Denial of Service Attack【jetty server in spark】
> -
>
> Key: SPARK-45248
> URL: https://issues.apache.org/jira/browse/SPARK-45248
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: chenyu
>Assignee: chenyu
>Priority: Major
>  Labels: pull-request-available
>
> The Jetty server may be slow HTTP denial of service attacks in spark ui.
> The default value of connector IdleTimeout is 3.
> This could trigger such an attack



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45286) Add back Matomo analytics to release docs

2023-09-24 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45286.
--
Fix Version/s: 3.3.4
   3.5.1
   4.0.0
   3.4.2
   Resolution: Fixed

Issue resolved by pull request 43063
[https://github.com/apache/spark/pull/43063]

> Add back Matomo analytics to release docs
> -
>
> Key: SPARK-45286
> URL: https://issues.apache.org/jira/browse/SPARK-45286
> Project: Spark
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 4.0.0
>Reporter: Sean R. Owen
>Assignee: Sean R. Owen
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.3.4, 3.5.1, 4.0.0, 3.4.2
>
>
> We had previously removed Google Analytics from the website and release docs, 
> per ASF policy: https://github.com/apache/spark/pull/36310
> We just restored analytics using the ASF-hosted Matomo service on the website:
> https://github.com/apache/spark-website/commit/a1548627b48a62c2e51870d1488ca3e09397bd30
> This change would put the same new tracking code back into the release docs.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45286) Add back Matomo analytics to release docs

2023-09-22 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45286?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45286:
-
Target Version/s: 3.4.2, 4.0.0, 3.5.1  (was: 4.0.0)

> Add back Matomo analytics to release docs
> -
>
> Key: SPARK-45286
> URL: https://issues.apache.org/jira/browse/SPARK-45286
> Project: Spark
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 4.0.0
>Reporter: Sean R. Owen
>Assignee: Sean R. Owen
>Priority: Minor
>  Labels: pull-request-available
>
> We had previously removed Google Analytics from the website and release docs, 
> per ASF policy: https://github.com/apache/spark/pull/36310
> We just restored analytics using the ASF-hosted Matomo service on the website:
> https://github.com/apache/spark-website/commit/a1548627b48a62c2e51870d1488ca3e09397bd30
> This change would put the same new tracking code back into the release docs.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-45273) Http header Attack【HttpSecurityFilter】

2023-09-22 Thread Sean R. Owen (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-45273?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17768144#comment-17768144
 ] 

Sean R. Owen commented on SPARK-45273:
--

Yep we typically evaluate security reports on priv...@spark.apache.org first, 
not here

> Http header Attack【HttpSecurityFilter】
> --
>
> Key: SPARK-45273
> URL: https://issues.apache.org/jira/browse/SPARK-45273
> Project: Spark
>  Issue Type: Bug
>  Components: Spark Core
>Affects Versions: 3.5.0
>Reporter: chenyu
>Priority: Major
>
> There is an HTTP host header attack vulnerability in the target URL



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-45286) Add back Matomo analytics to release docs

2023-09-22 Thread Sean R. Owen (Jira)
Sean R. Owen created SPARK-45286:


 Summary: Add back Matomo analytics to release docs
 Key: SPARK-45286
 URL: https://issues.apache.org/jira/browse/SPARK-45286
 Project: Spark
  Issue Type: Improvement
  Components: Documentation
Affects Versions: 4.0.0
Reporter: Sean R. Owen
Assignee: Sean R. Owen


We had previously removed Google Analytics from the website and release docs, 
per ASF policy: https://github.com/apache/spark/pull/36310

We just restored analytics using the ASF-hosted Matomo service on the website:
https://github.com/apache/spark-website/commit/a1548627b48a62c2e51870d1488ca3e09397bd30

This change would put the same new tracking code back into the release docs.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-45148) Upgrade ScalaTest to the 3.2.17 series.

2023-09-18 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen reassigned SPARK-45148:


Assignee: Yang Jie

> Upgrade ScalaTest to the 3.2.17 series.
> ---
>
> Key: SPARK-45148
> URL: https://issues.apache.org/jira/browse/SPARK-45148
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-45148) Upgrade ScalaTest to the 3.2.17 series.

2023-09-18 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen resolved SPARK-45148.
--
Fix Version/s: 4.0.0
   Resolution: Fixed

Issue resolved by pull request 42906
[https://github.com/apache/spark/pull/42906]

> Upgrade ScalaTest to the 3.2.17 series.
> ---
>
> Key: SPARK-45148
> URL: https://issues.apache.org/jira/browse/SPARK-45148
> Project: Spark
>  Issue Type: Improvement
>  Components: Build
>Affects Versions: 4.0.0
>Reporter: Yang Jie
>Assignee: Yang Jie
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 4.0.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-45195) Update Example with docker official image

2023-09-18 Thread Sean R. Owen (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-45195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean R. Owen updated SPARK-45195:
-
Priority: Minor  (was: Major)

> Update Example with docker official image
> -
>
> Key: SPARK-45195
> URL: https://issues.apache.org/jira/browse/SPARK-45195
> Project: Spark
>  Issue Type: Sub-task
>  Components: Documentation
>Affects Versions: 4.0.0
>Reporter: Ruifeng Zheng
>Assignee: Ruifeng Zheng
>Priority: Minor
> Fix For: 4.0.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



  1   2   3   4   5   6   7   8   9   10   >