[spark] branch master updated (7d8a721 -> dc0fa1e)

2021-12-12 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 7d8a721  [SPARK-37569][SQL] Don't mark nested view fields as nullable
 add dc0fa1e  [SPARK-37577][SQL] Fix ClassCastException: ArrayType cannot 
be cast to StructType for Generate Pruning

No new revisions were added by this update.

Summary of changes:
 .../catalyst/optimizer/NestedColumnAliasing.scala  | 26 +-
 .../execution/datasources/SchemaPruningSuite.scala | 21 -
 2 files changed, 40 insertions(+), 7 deletions(-)

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (414771d -> 7d8a721)

2021-12-12 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 414771d  [SPARK-37590][SQL][TEST] Unify v1 and v2 ALTER TABLE .. SET 
PROPERTIES tests
 add 7d8a721  [SPARK-37569][SQL] Don't mark nested view fields as nullable

No new revisions were added by this update.

Summary of changes:
 .../org/apache/spark/sql/catalyst/analysis/Analyzer.scala  |  2 +-
 .../org/apache/spark/sql/execution/SQLViewTestSuite.scala  | 10 ++
 2 files changed, 11 insertions(+), 1 deletion(-)

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (1cc1459 -> 414771d)

2021-12-12 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 1cc1459  [SPARK-36850][SQL][FOLLOWUP] Simplify exception code and fix 
wrong condition for CTAS and RTAS
 add 414771d  [SPARK-37590][SQL][TEST] Unify v1 and v2 ALTER TABLE .. SET 
PROPERTIES tests

No new revisions were added by this update.

Summary of changes:
 .../sql/catalyst/catalog/InMemoryCatalog.scala |   6 +-
 .../spark/sql/catalyst/parser/DDLParserSuite.scala |  32 --
 .../spark/sql/connector/DataSourceV2SQLSuite.scala |  48 -
 .../AlterNamespaceSetPropertiesParserSuite.scala   |  49 +
 .../AlterNamespaceSetPropertiesSuiteBase.scala | 117 +
 .../sql/execution/command/DDLParserSuite.scala |   6 --
 .../spark/sql/execution/command/DDLSuite.scala |  30 +-
 ...cala => AlterNamespaceSetPropertiesSuite.scala} |  18 ++--
 ...cala => AlterNamespaceSetPropertiesSuite.scala} |   8 +-
 ...cala => AlterNamespaceSetPropertiesSuite.scala} |   9 +-
 10 files changed, 187 insertions(+), 136 deletions(-)
 create mode 100644 
sql/core/src/test/scala/org/apache/spark/sql/execution/command/AlterNamespaceSetPropertiesParserSuite.scala
 create mode 100644 
sql/core/src/test/scala/org/apache/spark/sql/execution/command/AlterNamespaceSetPropertiesSuiteBase.scala
 copy 
sql/core/src/test/scala/org/apache/spark/sql/execution/command/v1/{AlterNamespaceSetLocationSuite.scala
 => AlterNamespaceSetPropertiesSuite.scala} (76%)
 copy 
sql/core/src/test/scala/org/apache/spark/sql/execution/command/v2/{AlterNamespaceSetLocationSuite.scala
 => AlterNamespaceSetPropertiesSuite.scala} (81%)
 copy 
sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/command/{AlterTableRecoverPartitionsSuite.scala
 => AlterNamespaceSetPropertiesSuite.scala} (75%)

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [SPARK-36850][SQL][FOLLOWUP] Simplify exception code and fix wrong condition for CTAS and RTAS

2021-12-12 Thread wenchen
This is an automated email from the ASF dual-hosted git repository.

wenchen pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 1cc1459  [SPARK-36850][SQL][FOLLOWUP] Simplify exception code and fix 
wrong condition for CTAS and RTAS
1cc1459 is described below

commit 1cc14598c7487887016793c0723624ceb4f13a8c
Author: Huaxin Gao 
AuthorDate: Mon Dec 13 12:34:27 2021 +0800

[SPARK-36850][SQL][FOLLOWUP] Simplify exception code and fix wrong 
condition for CTAS and RTAS

### What changes were proposed in this pull request?
fixed a few problems:
1. addressed this 
[comment](https://github.com/apache/spark/pull/34060#discussion_r765992537)
2. combined several `xxxOnlySupportedWithV2TableError`
3. in CTAS and RTAS, the `if isSessionCatalog(catalog)` should not be on 
the pattern, it should be `if (isSessionCatalog(catalog) && 
!isV2Provider(provider))`. Otherwise, `c.partitioning ++ 
c.tableSpec.bucketSpec.map(_.asTransform)` is not done for non SessionCatalog 
case.
I tried this  `c.partitioning ++ c.tableSpec.bucketSpec.map(_.asTransform)` 
inside `AstBuilder` but it failed [here]( 
https://github.com/apache/spark/blob/master/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala#L850)
 so I kept this in `ResolveSessionCatalog`

### Why are the changes needed?
code cleaning up and bug fixing

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
Existing tests

Closes #34857 from huaxingao/followup.

Lead-authored-by: Huaxin Gao 
Co-authored-by: Huaxin Gao 
Signed-off-by: Wenchen Fan 
---
 .../spark/sql/errors/QueryCompilationErrors.scala  | 28 +
 .../catalyst/analysis/ResolveSessionCatalog.scala  | 73 --
 .../spark/sql/streaming/DataStreamWriter.scala |  4 +-
 .../spark/sql/connector/DataSourceV2SQLSuite.scala | 38 +++
 4 files changed, 82 insertions(+), 61 deletions(-)

diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala
index 920a748..4843051 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryCompilationErrors.scala
@@ -479,48 +479,24 @@ object QueryCompilationErrors {
 new AnalysisException("ADD COLUMN with v1 tables cannot specify NOT NULL.")
   }
 
-  def replaceColumnsOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("REPLACE COLUMNS is only supported with v2 tables.")
-  }
-
-  def alterQualifiedColumnOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("ALTER COLUMN with qualified column is only 
supported with v2 tables.")
+  def operationOnlySupportedWithV2TableError(operation: String): Throwable = {
+new AnalysisException(s"$operation is only supported with v2 tables.")
   }
 
   def alterColumnWithV1TableCannotSpecifyNotNullError(): Throwable = {
 new AnalysisException("ALTER COLUMN with v1 tables cannot specify NOT 
NULL.")
   }
 
-  def alterOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("ALTER COLUMN ... FIRST | ALTER is only supported 
with v2 tables.")
-  }
-
   def alterColumnCannotFindColumnInV1TableError(colName: String, v1Table: 
V1Table): Throwable = {
 new AnalysisException(
   s"ALTER COLUMN cannot find column $colName in v1 table. " +
 s"Available: ${v1Table.schema.fieldNames.mkString(", ")}")
   }
 
-  def renameColumnOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("RENAME COLUMN is only supported with v2 tables.")
-  }
-
-  def dropColumnOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("DROP COLUMN is only supported with v2 tables.")
-  }
-
   def invalidDatabaseNameError(quoted: String): Throwable = {
 new AnalysisException(s"The database name is not valid: $quoted")
   }
 
-  def replaceTableOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("REPLACE TABLE is only supported with v2 tables.")
-  }
-
-  def replaceTableAsSelectOnlySupportedWithV2TableError(): Throwable = {
-new AnalysisException("REPLACE TABLE AS SELECT is only supported with v2 
tables.")
-  }
-
   def cannotDropViewWithDropTableError(): Throwable = {
 new AnalysisException("Cannot drop a view with DROP TABLE. Please use DROP 
VIEW instead")
   }
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveSessionCatalog.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveSessionCatalog.scala
index d5f99bf..15798e0 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/catalyst/analysis/ResolveSessionCatalog.scala
+++ 

[spark] branch master updated (5880df4 -> 7390eb0)

2021-12-12 Thread wuyi
This is an automated email from the ASF dual-hosted git repository.

wuyi pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 5880df4  [SPARK-37481][CORE][WEBUI] Fix disappearance of skipped 
stages after they retry
 add 7390eb0  [SPARK-37300][CORE] TaskSchedulerImpl should ignore task 
finished eveā€¦

No new revisions were added by this update.

Summary of changes:
 .../apache/spark/scheduler/TaskSetManager.scala|  9 ++-
 .../spark/scheduler/TaskSchedulerImplSuite.scala   | 94 +-
 2 files changed, 101 insertions(+), 2 deletions(-)

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch branch-3.2 updated: [SPARK-37481][CORE][WEBUI] Fix disappearance of skipped stages after they retry

2021-12-12 Thread wuyi
This is an automated email from the ASF dual-hosted git repository.

wuyi pushed a commit to branch branch-3.2
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.2 by this push:
 new f20c408  [SPARK-37481][CORE][WEBUI] Fix disappearance of skipped 
stages after they retry
f20c408 is described below

commit f20c408be0cd2f53815b4b26a88aa8c2e3545a5c
Author: Kent Yao 
AuthorDate: Mon Dec 13 10:40:42 2021 +0800

[SPARK-37481][CORE][WEBUI] Fix disappearance of skipped stages after they 
retry

### What changes were proposed in this pull request?

When skipped stages retry, their skipped info will be lost on the UI, and 
then we may see a stage with 200 tasks indeed, shows that it only has 3 tasks 
but its `retry 1` has 15 tasks and completely different inputs/outputs.

A simple way to reproduce,

```
bin/spark-sql --packages com.github.yaooqinn:itachi_2.12:0.3.0
```

and run

```
select * from (select v from (values (1), (2), (3) t(v))) t1 join (select 
stage_id_with_retry(3) from (select v from values (1), (2), (3) t(v) group by 
v)) t2;
```

Also, Detailed in the Gist here - 
https://gist.github.com/yaooqinn/6acb7b74b343a6a6dffe8401f6b7b45c

In this PR, we increase the nextAttempIds of these skipped stages once they 
get visited.

### Why are the changes needed?

fix problems when we have skipped stage retries.

### Does this PR introduce _any_ user-facing change?

Yes, the UI will keep the skipped stages info


![image](https://user-images.githubusercontent.com/8326978/144010378-02a688ce-0ead-4c41-ab9b-bc5fce4f8b90.png)

### How was this patch tested?

manually as recorded in 
https://gist.github.com/yaooqinn/6acb7b74b343a6a6dffe8401f6b7b45c

existing tests

Closes #34735 from yaooqinn/SPARK-37481.

Authored-by: Kent Yao 
Signed-off-by: yi.wu 
(cherry picked from commit 5880df41f50210f2df44f25640437d99b8979d70)
Signed-off-by: yi.wu 
---
 core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala  | 6 ++
 core/src/main/scala/org/apache/spark/scheduler/Stage.scala | 7 +++
 .../test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala  | 6 +++---
 3 files changed, 16 insertions(+), 3 deletions(-)

diff --git a/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala 
b/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
index bb77a58d..9a27d9c 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
@@ -716,6 +716,12 @@ private[spark] class DAGScheduler(
 // read from merged output as the MergeStatuses are not 
available.
 if (!mapStage.isAvailable || 
!mapStage.shuffleDep.shuffleMergeFinalized) {
   missing += mapStage
+} else {
+  // Forward the nextAttemptId if skipped and get visited for 
the first time.
+  // Otherwise, once it gets retried,
+  // 1) the stuffs in stage info become distorting, e.g. task 
num, input byte, e.t.c
+  // 2) the first attempt starts from 0-idx, it will not be 
marked as a retry
+  mapStage.increaseAttemptIdOnFirstSkip()
 }
   case narrowDep: NarrowDependency[_] =>
 waitingForVisit.prepend(narrowDep.rdd)
diff --git a/core/src/main/scala/org/apache/spark/scheduler/Stage.scala 
b/core/src/main/scala/org/apache/spark/scheduler/Stage.scala
index ae7924d..9707211 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/Stage.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/Stage.scala
@@ -107,6 +107,13 @@ private[scheduler] abstract class Stage(
 nextAttemptId += 1
   }
 
+  /** Forward the nextAttemptId if skipped and get visited for the first time. 
*/
+  def increaseAttemptIdOnFirstSkip(): Unit = {
+if (nextAttemptId == 0) {
+  nextAttemptId = 1
+}
+  }
+
   /** Returns the StageInfo for the most recent attempt for this stage. */
   def latestInfo: StageInfo = _latestInfo
 
diff --git 
a/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala 
b/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
index 4cb64ed..afea912 100644
--- a/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
@@ -1622,7 +1622,7 @@ class DAGSchedulerSuite extends SparkFunSuite with 
TempLocalSparkContext with Ti
 // the shuffle map output is still available from stage 0); make sure 
we've still got internal
 // accumulators setup
 assert(scheduler.stageIdToStage(2).latestInfo.taskMetrics != null)
-completeShuffleMapStageSuccessfully(2, 0, 2)
+

[spark] branch master updated (a30bec1 -> 5880df4)

2021-12-12 Thread wuyi
This is an automated email from the ASF dual-hosted git repository.

wuyi pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from a30bec1  [SPARK-37369][SQL] Avoid redundant ColumnarToRow transistion 
on InMemoryTableScan
 add 5880df4  [SPARK-37481][CORE][WEBUI] Fix disappearance of skipped 
stages after they retry

No new revisions were added by this update.

Summary of changes:
 core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala  | 6 ++
 core/src/main/scala/org/apache/spark/scheduler/Stage.scala | 7 +++
 .../test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala  | 6 +++---
 3 files changed, 16 insertions(+), 3 deletions(-)

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [SPARK-37369][SQL] Avoid redundant ColumnarToRow transistion on InMemoryTableScan

2021-12-12 Thread viirya
This is an automated email from the ASF dual-hosted git repository.

viirya pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new a30bec1  [SPARK-37369][SQL] Avoid redundant ColumnarToRow transistion 
on InMemoryTableScan
a30bec1 is described below

commit a30bec17d6187ef6dfe276c67dbd4e023062b92b
Author: Liang-Chi Hsieh 
AuthorDate: Sun Dec 12 17:49:58 2021 -0800

[SPARK-37369][SQL] Avoid redundant ColumnarToRow transistion on 
InMemoryTableScan

### What changes were proposed in this pull request?

This patch proposes to let `InMemoryTableScanExec` produces row output 
directly, if its parent query plan only accepts rows instead of columnar 
output. In particular, this change adds a new method in `SparkPlan` called 
`supportsRowBased`, alongside with the existing `supportsColumnar`.

### Why are the changes needed?

We currently have `supportsColumnar` indicating if a physical node can 
produce columnar output. The current columnar transition rule seems taking an 
assumption that one node can only produce columnar output but not row-based one 
if `supportsColumnar` returns true. But actually one node can possibly produce 
both format, i.e. columnar and row-based. For such node, if we require 
row-based output, the columnar transition rule will add additional 
columnar-to-row after it due to the wrong a [...]

So this change introduces `supportsRowBased` which is used to indicates if 
the node can produce row-based output. The rule can check this method when 
deciding if a columnar-to-row transition is necessary or not.

For example, `InMemoryTableScanExec` can produce columnar output. So if its 
parent plan isn't columnar, the rule adds a `ColumnarToRow` between them, e.g.,

```
+- Union
:- ColumnarToRow
: +- InMemoryTableScan i#8, j#9
: +- InMemoryRelation i#8, j#9, StorageLevel(disk, memory, deserialized, 1 
replicas)
```

But `InMemoryTableScanExec` is capable of row-based output too. After this 
change, for such case, we can simply ask `InMemoryTableScanExec` to produce row 
output instead of a redundant conversion.

```


Int In-memory



OpenJDK 64-Bit Server VM 1.8.0_265-b01 on Mac OS X 10.16
Intel(R) Core(TM) i7-9750H CPU  2.60GHz
Int In-Memory scan: Best Time(ms)   Avg Time(ms)   
Stdev(ms)Rate(M/s)   Per Row(ns)   Relative

--
columnar deserialization + columnar-to-row228245
  15  4.4 227.7   1.0X
row-based deserialization 179187
  10  5.6 179.4   1.3X
```

### Does this PR introduce _any_ user-facing change?

No

### How was this patch tested?

Existing tests.

Closes #34642 from viirya/SPARK-37369.

Authored-by: Liang-Chi Hsieh 
Signed-off-by: Liang-Chi Hsieh 
---
 .../InMemoryColumnarBenchmark-results.txt  | 12 
 .../org/apache/spark/sql/execution/Columnar.scala  |  4 +-
 .../org/apache/spark/sql/execution/SparkPlan.scala | 11 +++-
 .../sql/execution/columnar/InMemoryRelation.scala  |  3 +-
 .../execution/columnar/InMemoryTableScanExec.scala |  2 +
 .../org/apache/spark/sql/CachedTableSuite.scala| 21 ++-
 .../spark/sql/DataFrameSetOperationsSuite.scala|  7 +--
 .../sql/execution/WholeStageCodegenSuite.scala |  2 +-
 .../columnar/CachedBatchSerializerSuite.scala  |  5 +-
 .../columnar/InMemoryColumnarBenchmark.scala   | 66 ++
 .../columnar/InMemoryColumnarQuerySuite.scala  |  6 +-
 .../spark/sql/execution/debug/DebuggingSuite.scala |  5 +-
 .../sql/execution/metric/SQLMetricsSuite.scala |  6 +-
 13 files changed, 128 insertions(+), 22 deletions(-)

diff --git a/sql/core/benchmarks/InMemoryColumnarBenchmark-results.txt 
b/sql/core/benchmarks/InMemoryColumnarBenchmark-results.txt
new file mode 100644
index 000..2998d8b
--- /dev/null
+++ b/sql/core/benchmarks/InMemoryColumnarBenchmark-results.txt
@@ -0,0 +1,12 @@
+
+Int In-memory
+
+
+OpenJDK 64-Bit Server VM 1.8.0_312-b07 on Linux 5.11.0-1021-azure
+Intel(R) Xeon(R) Platinum 8272CL CPU @ 2.60GHz
+Int In-Memory scan: Best Time(ms)   Avg Time(ms)   
Stdev(ms)Rate(M/s)   Per Row(ns)   Relative

[spark] branch master updated: [SPARK-36038][CORE] Speculation metrics summary at stage level

2021-12-12 Thread sarutak
This is an automated email from the ASF dual-hosted git repository.

sarutak pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 4158d35  [SPARK-36038][CORE] Speculation metrics summary at stage level
4158d35 is described below

commit 4158d3544030058a23c7741d4283213a56ae7cc7
Author: Thejdeep Gudivada 
AuthorDate: Mon Dec 13 09:58:27 2021 +0900

[SPARK-36038][CORE] Speculation metrics summary at stage level

### What changes were proposed in this pull request?
Currently there are no speculation metrics available for Spark either at 
application/job/stage level. This PR is to add some basic speculation metrics 
for a stage when speculation execution is enabled.

This is similar to the existing stage level metrics tracking numTotal 
(total number of speculated tasks), numCompleted (total number of successful 
speculated tasks), numFailed (total number of failed speculated tasks), 
numKilled (total number of killed speculated tasks) etc.

With this new set of metrics, it helps further understanding speculative 
execution feature in the context of the application and also helps in further 
tuning the speculative execution config knobs.

 ### Why are the changes needed?
Additional metrics for speculative execution.

 ### Does this PR introduce _any_ user-facing change?
Yes, Stages Page in SHS UI will have an additional table for speculation 
metrics, if present.

 ### How was this patch tested?
Unit tests added and also tested on our internal platform.

Absence of speculation metrics :
![Screen Shot 2021-11-15 at 10 12 23 
AM](https://user-images.githubusercontent.com/1708757/141836847-b0768265-35eb-4963-b83a-8b785318e82d.png)

Presence of speculation metrics :
![Screen Shot 2021-11-15 at 10 11 52 
AM](https://user-images.githubusercontent.com/1708757/141836945-1078bfcb-1c07-44db-9f33-5dbd09ece6f2.png)

Closes #34607 from thejdeep/SPARK-36038.

Lead-authored-by: Thejdeep Gudivada 
Co-authored-by: Ron Hu 
Co-authored-by: Venkata krishnan Sowrirajan 
Signed-off-by: Kousuke Saruta 
---
 .../org/apache/spark/ui/static/stagepage.js|  32 ++
 .../spark/ui/static/stagespage-template.html   |  15 +
 .../resources/org/apache/spark/ui/static/webui.css |  10 +
 .../apache/spark/status/AppStatusListener.scala|  14 +
 .../org/apache/spark/status/AppStatusStore.scala   |  11 +
 .../scala/org/apache/spark/status/LiveEntity.scala |  26 ++
 .../scala/org/apache/spark/status/api/v1/api.scala |   8 +
 .../scala/org/apache/spark/status/storeTypes.scala |  12 +
 .../scala/org/apache/spark/ui/jobs/JobPage.scala   |   1 +
 .../application_list_json_expectation.json |  15 +
 .../completed_app_list_json_expectation.json   |  15 +
 .../limit_app_list_json_expectation.json   |  30 +-
 .../minDate_app_list_json_expectation.json |  15 +
 .../minEndDate_app_list_json_expectation.json  |  15 +
 ...stage_with_speculation_summary_expectation.json | 507 +
 .../spark-events/application_1628109047826_1317105 |  52 +++
 .../spark/deploy/history/HistoryServerSuite.scala  |   5 +-
 .../spark/status/AppStatusListenerSuite.scala  |  10 +
 .../apache/spark/status/AppStatusStoreSuite.scala  |  51 ++-
 .../scala/org/apache/spark/ui/StagePageSuite.scala |   1 +
 dev/.rat-excludes  |   3 +-
 21 files changed, 830 insertions(+), 18 deletions(-)

diff --git a/core/src/main/resources/org/apache/spark/ui/static/stagepage.js 
b/core/src/main/resources/org/apache/spark/ui/static/stagepage.js
index 584e1a7..a5955f3 100644
--- a/core/src/main/resources/org/apache/spark/ui/static/stagepage.js
+++ b/core/src/main/resources/org/apache/spark/ui/static/stagepage.js
@@ -652,6 +652,38 @@ $(document).ready(function () {
 
executorSummaryTableSelector.column(14).visible(dataToShow.showBytesSpilledData);
   });
 
+// Prepare data for speculation metrics
+$("#speculationSummaryTitle").hide();
+$("#speculationSummary").hide();
+var speculationSummaryInfo = responseBody.speculationSummary;
+var speculationData;
+if(speculationSummaryInfo) {
+  speculationData = [[
+speculationSummaryInfo.numTasks,
+speculationSummaryInfo.numActiveTasks,
+speculationSummaryInfo.numCompletedTasks,
+speculationSummaryInfo.numFailedTasks,
+speculationSummaryInfo.numKilledTasks
+  ]];
+  if (speculationSummaryInfo.numTasks > 0) {
+// Show speculationSummary if there is atleast one speculated task 
that ran
+$("#speculationSummaryTitle").show();
+$("#speculationSummary").show();
+  }
+}
+var speculationMetricsTableConf = {
+  "data": 

[spark] branch master updated: [SPARK-37619][BUILD][TEST-MAVEN] Upgrade Maven to 3.8.4

2021-12-12 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 8477d82  [SPARK-37619][BUILD][TEST-MAVEN] Upgrade Maven to 3.8.4
8477d82 is described below

commit 8477d824d36b03222c310ebb2f1230e3e2466d08
Author: Dongjoon Hyun 
AuthorDate: Mon Dec 13 09:53:28 2021 +0900

[SPARK-37619][BUILD][TEST-MAVEN] Upgrade Maven to 3.8.4

### What changes were proposed in this pull request?

This PR aims to upgrade Maven to 3.8.4 from 3.6.3.

### Why are the changes needed?

https://maven.apache.org/docs/3.8.4/release-notes.html
```
Upgrade to Jansi 2.4.0 which supports macOS on ARM natively
```

Note that MNG-6843 `Parallel build fails due to missing JAR artifacts in 
compilePath` exists in both 3.6.3 and 3.8.4 because MNG-6843 was merged at 
3.8.2 and reverted at 3.8.4.

https://issues.apache.org/jira/browse/MNG-6843

### Does this PR introduce _any_ user-facing change?

No. This is a dev-only change.

### How was this patch tested?

Pass the Jenkins with Maven. The current Jenkins run is triggered via 
`Maven 3.8.4`

```

Building Spark

[info] Building Spark using Maven with these arguments:  -Phadoop-3 -Pmesos 
-Phadoop-cloud -Pdocker-integration-tests -Phive-thriftserver -Phive 
-Pkubernetes -Pyarn -Pspark-ganglia-lgpl -Pkinesis-asl clean package -DskipTests
exec: curl --silent --show-error -L 
https://downloads.lightbend.com/scala/2.12.15/scala-2.12.15.tgz
exec: curl --silent --show-error -L 
https://www.apache.org/dyn/closer.lua/maven/maven-3/3.8.4/binaries/apache-maven-3.8.4-bin.tar.gz?action=download
exec: curl --silent --show-error -L 
https://archive.apache.org/dist/maven/maven-3/3.8.4/binaries/apache-maven-3.8.4-bin.tar.gz.sha512
```

Closes #34873 from dongjoon-hyun/SPARK-37619.

Authored-by: Dongjoon Hyun 
Signed-off-by: Hyukjin Kwon 
---
 dev/appveyor-install-dependencies.ps1 | 2 +-
 docs/building-spark.md| 2 +-
 pom.xml   | 2 +-
 3 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/dev/appveyor-install-dependencies.ps1 
b/dev/appveyor-install-dependencies.ps1
index fdc1e5e..5f239b2 100644
--- a/dev/appveyor-install-dependencies.ps1
+++ b/dev/appveyor-install-dependencies.ps1
@@ -81,7 +81,7 @@ if (!(Test-Path $tools)) {
 # == Maven
 # Push-Location $tools
 #
-# $mavenVer = "3.6.3"
+# $mavenVer = "3.8.4"
 # Start-FileDownload 
"https://archive.apache.org/dist/maven/maven-3/$mavenVer/binaries/apache-maven-$mavenVer-bin.zip;
 "maven.zip"
 #
 # # extract
diff --git a/docs/building-spark.md b/docs/building-spark.md
index 5428f77..4e77509 100644
--- a/docs/building-spark.md
+++ b/docs/building-spark.md
@@ -27,7 +27,7 @@ license: |
 ## Apache Maven
 
 The Maven-based build is the build of reference for Apache Spark.
-Building Spark using Maven requires Maven 3.6.3 and Java 8.
+Building Spark using Maven requires Maven 3.8.4 and Java 8.
 Spark requires Scala 2.12/2.13; support for Scala 2.11 was removed in Spark 
3.0.0.
 
 ### Setting up Maven's Memory Usage
diff --git a/pom.xml b/pom.xml
index 4ea8b38..4c04c70 100644
--- a/pom.xml
+++ b/pom.xml
@@ -115,7 +115,7 @@
 1.8
 ${java.version}
 ${java.version}
-3.6.3
+3.8.4
 1.6.0
 spark
 1.7.30

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (01414e3 -> 79068fb)

2021-12-12 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 01414e3  [SPARK-37152][PYTHON] Inline type hints for 
python/pyspark/context.py
 add 79068fb  [SPARK-37598][PYTHON] Adding support for ShortWriables to 
pyspark's newAPIHadoopRDD method

No new revisions were added by this update.

Summary of changes:
 .../apache/spark/api/python/PythonHadoopUtil.scala |  4 ++
 .../spark/api/python/PythonHadoopUtilSuite.scala   | 80 ++
 2 files changed, 84 insertions(+)
 create mode 100644 
core/src/test/scala/org/apache/spark/api/python/PythonHadoopUtilSuite.scala

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (7692773 -> 01414e3)

2021-12-12 Thread zero323
This is an automated email from the ASF dual-hosted git repository.

zero323 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git.


from 7692773  [SPARK-37615][BUILD][FOLLOWUP] Upgrade SBT to 1.5.6 in 
AppVeyor
 add 01414e3  [SPARK-37152][PYTHON] Inline type hints for 
python/pyspark/context.py

No new revisions were added by this update.

Summary of changes:
 python/pyspark/context.py  | 375 ++---
 python/pyspark/context.pyi | 195 ---
 python/pyspark/util.py |   4 +
 3 files changed, 225 insertions(+), 349 deletions(-)
 delete mode 100644 python/pyspark/context.pyi

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org