manuzhang commented on PR #36698:
URL: https://github.com/apache/spark/pull/36698#issuecomment-1153543507
@cloud-fan do we plan to back-port it to branch-3.1?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL abo
dongjoon-hyun commented on PR #36852:
URL: https://github.com/apache/spark/pull/36852#issuecomment-1153540713
Thank you for updates.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific c
MaxGekk commented on code in PR #36811:
URL: https://github.com/apache/spark/pull/36811#discussion_r895359514
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Cast.scala:
##
@@ -108,8 +108,9 @@ object Cast {
case (TimestampType, TimestampNTZType) =>
MaxGekk commented on code in PR #36811:
URL: https://github.com/apache/spark/pull/36811#discussion_r895359204
##
sql/core/src/test/resources/sql-tests/inputs/cast.sql:
##
@@ -104,3 +104,15 @@ select cast('a' as timestamp_ntz);
select cast(cast('inf' as double) as timestamp);
cloud-fan closed pull request #36698: [SPARK-39316][SQL] Merge PromotePrecision
and CheckOverflow into decimal binary arithmetic
URL: https://github.com/apache/spark/pull/36698
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
cloud-fan commented on PR #36698:
URL: https://github.com/apache/spark/pull/36698#issuecomment-1153504963
thanks, merging to master!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific c
mridulm commented on code in PR #36162:
URL: https://github.com/apache/spark/pull/36162#discussion_r895335273
##
core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala:
##
@@ -1218,6 +1249,71 @@ private[spark] class TaskSetManager(
def executorAdded(): Unit = {
HyukjinKwon commented on code in PR #36683:
URL: https://github.com/apache/spark/pull/36683#discussion_r894062038
##
python/pyspark/sql/pandas/conversion.py:
##
@@ -596,7 +596,7 @@ def _create_from_pandas_with_arrow(
]
# Slice the DataFrame to be batched
HyukjinKwon commented on PR #36683:
URL: https://github.com/apache/spark/pull/36683#issuecomment-1153477826
Rebased
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsub
cxzl25 commented on PR #36740:
URL: https://github.com/apache/spark/pull/36740#issuecomment-1153476515
Gentle ping @sarutak @cloud-fan @dongjoon-hyun
This should be a bug, hope to help review.
--
This is an automated message from the Apache Git Service.
To respond to the message, please
pan3793 commented on PR #36789:
URL: https://github.com/apache/spark/pull/36789#issuecomment-1153466782
@HyukjinKwon would you please take another look if you have time?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
mridulm commented on PR #35683:
URL: https://github.com/apache/spark/pull/35683#issuecomment-1153436041
Can you update the description to reflect the changes made to the PR
@abhishekd0907 ?
Specifically - this is only related to decomissioning and we do not handle
shuffle ?
The c
mcdull-zhang commented on PR #36831:
URL: https://github.com/apache/spark/pull/36831#issuecomment-1153431584
@cloud-fan @ulysses-you friendly ping
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
pan3793 commented on code in PR #36832:
URL: https://github.com/apache/spark/pull/36832#discussion_r895303204
##
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala:
##
@@ -747,6 +747,15 @@ private[history] class FsHistoryProvider(conf: SparkConf,
clock:
mridulm commented on code in PR #35906:
URL: https://github.com/apache/spark/pull/35906#discussion_r895292260
##
common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/RemoteBlockPushResolver.java:
##
@@ -317,22 +353,24 @@ public void applicationRemoved(String app
mridulm commented on code in PR #35906:
URL: https://github.com/apache/spark/pull/35906#discussion_r895282920
##
common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/RemoteBlockPushResolver.java:
##
@@ -350,15 +415,27 @@ void closeAndDeletePartitionFilesIfNeeded
mridulm commented on code in PR #35906:
URL: https://github.com/apache/spark/pull/35906#discussion_r895282269
##
common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/RemoteBlockPushResolver.java:
##
@@ -342,6 +380,33 @@ void closeAndDeletePartitionFilesIfNeeded(
weixiuli commented on code in PR #36162:
URL: https://github.com/apache/spark/pull/36162#discussion_r895280301
##
core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala:
##
@@ -863,6 +872,29 @@ private[spark] class TaskSchedulerImpl(
executorUpdates)
}
weixiuli commented on code in PR #36162:
URL: https://github.com/apache/spark/pull/36162#discussion_r895280301
##
core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala:
##
@@ -863,6 +872,29 @@ private[spark] class TaskSchedulerImpl(
executorUpdates)
}
beliefer commented on code in PR #36830:
URL: https://github.com/apache/spark/pull/36830#discussion_r895270044
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java:
##
@@ -97,6 +97,10 @@ public String build(Expression expr) {
r
beliefer commented on code in PR #36830:
URL: https://github.com/apache/spark/pull/36830#discussion_r895270044
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java:
##
@@ -97,6 +97,10 @@ public String build(Expression expr) {
r
dongjoon-hyun commented on PR #36847:
URL: https://github.com/apache/spark/pull/36847#issuecomment-1153369445
I checked Apache Spark 3.3.0 RC6 and added `3.3.0` to the Affected Version
of the JIRA, @wangyum .
```
scala> spark.version
val res0: String = 3.3.0
scala> sql("set
dongjoon-hyun commented on code in PR #36847:
URL: https://github.com/apache/spark/pull/36847#discussion_r895266088
##
sql/core/src/main/scala/org/apache/spark/sql/execution/SparkOptimizer.scala:
##
@@ -87,7 +87,8 @@ class SparkOptimizer(
GroupBasedRowLevelOperationScanPlan
dongjoon-hyun commented on code in PR #36847:
URL: https://github.com/apache/spark/pull/36847#discussion_r895266012
##
sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala:
##
@@ -4456,6 +4456,20 @@ class SQLQuerySuite extends QueryTest with
SharedSparkSession with
dongjoon-hyun commented on code in PR #36847:
URL: https://github.com/apache/spark/pull/36847#discussion_r895265570
##
sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala:
##
@@ -4456,6 +4456,20 @@ class SQLQuerySuite extends QueryTest with
SharedSparkSession with
HyukjinKwon commented on code in PR #36830:
URL: https://github.com/apache/spark/pull/36830#discussion_r895264380
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/util/V2ExpressionSQLBuilder.java:
##
@@ -97,6 +97,10 @@ public String build(Expression expr) {
weixiuli commented on PR #36162:
URL: https://github.com/apache/spark/pull/36162#issuecomment-1153357410
ping @Ngone51 @mridulm Thanks.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specifi
dongjoon-hyun commented on PR #36832:
URL: https://github.com/apache/spark/pull/36832#issuecomment-1153354775
Thank you for your update, @pan3793 .
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
dongjoon-hyun commented on code in PR #36832:
URL: https://github.com/apache/spark/pull/36832#discussion_r895262820
##
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala:
##
@@ -747,6 +747,15 @@ private[history] class FsHistoryProvider(conf: SparkConf,
dongjoon-hyun commented on code in PR #36832:
URL: https://github.com/apache/spark/pull/36832#discussion_r895262820
##
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala:
##
@@ -747,6 +747,15 @@ private[history] class FsHistoryProvider(conf: SparkConf,
gengliangwang commented on code in PR #36811:
URL: https://github.com/apache/spark/pull/36811#discussion_r895262282
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Cast.scala:
##
@@ -108,8 +108,9 @@ object Cast {
case (TimestampType, TimestampNTZTyp
dongjoon-hyun commented on code in PR #36832:
URL: https://github.com/apache/spark/pull/36832#discussion_r895261688
##
core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala:
##
@@ -747,6 +747,15 @@ private[history] class FsHistoryProvider(conf: SparkConf,
gengliangwang commented on code in PR #36811:
URL: https://github.com/apache/spark/pull/36811#discussion_r895261353
##
sql/core/src/test/resources/sql-tests/inputs/cast.sql:
##
@@ -104,3 +104,15 @@ select cast('a' as timestamp_ntz);
select cast(cast('inf' as double) as timest
dongjoon-hyun commented on code in PR #36852:
URL: https://github.com/apache/spark/pull/36852#discussion_r895261028
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -28,14 +28,30 @@ import com.fasterxml.jackson.module.scala.DefaultScalaModule
import org.apache.s
dongjoon-hyun commented on code in PR #36852:
URL: https://github.com/apache/spark/pull/36852#discussion_r895260896
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -28,14 +28,30 @@ import com.fasterxml.jackson.module.scala.DefaultScalaModule
import org.apache.s
dongjoon-hyun commented on code in PR #36852:
URL: https://github.com/apache/spark/pull/36852#discussion_r895260558
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -28,14 +28,30 @@ import com.fasterxml.jackson.module.scala.DefaultScalaModule
import org.apache.s
HyukjinKwon commented on PR #36829:
URL: https://github.com/apache/spark/pull/36829#issuecomment-1153343979
cc @peter-toth @allisonwang-db @maryannxue FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above t
HyukjinKwon commented on PR #36811:
URL: https://github.com/apache/spark/pull/36811#issuecomment-1153343457
cc @gengliangwang FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comm
HyukjinKwon commented on PR #36841:
URL: https://github.com/apache/spark/pull/36841#issuecomment-1153342860
cc @maryannxue and @allisonwang-db FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
HyukjinKwon commented on PR #36842:
URL: https://github.com/apache/spark/pull/36842#issuecomment-1153340903
cc @hvanhovell FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment
HyukjinKwon commented on PR #36845:
URL: https://github.com/apache/spark/pull/36845#issuecomment-1153340424
cc @maryannxue FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment
HyukjinKwon commented on code in PR #36844:
URL: https://github.com/apache/spark/pull/36844#discussion_r895257421
##
repl/src/main/scala/org/apache/spark/repl/ExecutorClassLoader.scala:
##
@@ -54,7 +54,7 @@ class ExecutorClassLoader(
classUri: String,
parent: ClassLoad
wangyum commented on code in PR #36847:
URL: https://github.com/apache/spark/pull/36847#discussion_r895254709
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala:
##
@@ -270,7 +270,8 @@ abstract class Optimizer(catalogManager: CatalogManager)
HyukjinKwon commented on code in PR #36847:
URL: https://github.com/apache/spark/pull/36847#discussion_r895254287
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala:
##
@@ -270,7 +270,8 @@ abstract class Optimizer(catalogManager: CatalogManage
github-actions[bot] commented on PR #35256:
URL: https://github.com/apache/spark/pull/35256#issuecomment-1153332516
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
github-actions[bot] commented on PR #35719:
URL: https://github.com/apache/spark/pull/35719#issuecomment-1153332508
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
HyukjinKwon commented on PR #36848:
URL: https://github.com/apache/spark/pull/36848#issuecomment-1153331778
cc @hvanhovell FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment
HyukjinKwon commented on code in PR #36850:
URL: https://github.com/apache/spark/pull/36850#discussion_r895252969
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala:
##
@@ -1501,19 +1501,42 @@ object EliminateSorts extends Rule[LogicalPlan] {
HyukjinKwon closed pull request #36793: [SPARK-39406][PYTHON] Accept NumPy
array in createDataFrame
URL: https://github.com/apache/spark/pull/36793
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to th
HyukjinKwon commented on PR #36793:
URL: https://github.com/apache/spark/pull/36793#issuecomment-1153328439
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
HyukjinKwon commented on code in PR #36793:
URL: https://github.com/apache/spark/pull/36793#discussion_r895251697
##
python/pyspark/sql/session.py:
##
@@ -952,12 +953,29 @@ def createDataFrame( # type: ignore[misc]
schema = [x.encode("utf-8") if not isinstance(x, s
HyukjinKwon commented on PR #36840:
URL: https://github.com/apache/spark/pull/36840#issuecomment-1153327788
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
HyukjinKwon closed pull request #36840: [SPARK-39443][PYTHON][DOC] Improve
docstring of pyspark.sql.functions.col/first
URL: https://github.com/apache/spark/pull/36840
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
U
gengliangwang commented on code in PR #36771:
URL: https://github.com/apache/spark/pull/36771#discussion_r895248971
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/V2SessionCatalog.scala:
##
@@ -43,6 +43,8 @@ class V2SessionCatalog(catalog: SessionCatalo
gengliangwang commented on code in PR #36771:
URL: https://github.com/apache/spark/pull/36771#discussion_r895248853
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/v2/V2SessionCatalog.scala:
##
@@ -43,6 +43,8 @@ class V2SessionCatalog(catalog: SessionCatalo
xinrong-databricks commented on PR #36793:
URL: https://github.com/apache/spark/pull/36793#issuecomment-1153244595
Rebased for conflicts in `python/docs/source/getting_started/install.rst`
only.
--
This is an automated message from the Apache Git Service.
To respond to the message, please
xinrong-databricks commented on code in PR #36793:
URL: https://github.com/apache/spark/pull/36793#discussion_r895210608
##
python/pyspark/sql/session.py:
##
@@ -952,12 +953,29 @@ def createDataFrame( # type: ignore[misc]
schema = [x.encode("utf-8") if not isinstan
xinrong-databricks commented on code in PR #36793:
URL: https://github.com/apache/spark/pull/36793#discussion_r895210329
##
python/pyspark/sql/session.py:
##
@@ -952,12 +953,29 @@ def createDataFrame( # type: ignore[misc]
schema = [x.encode("utf-8") if not isinstan
xinrong-databricks commented on code in PR #36793:
URL: https://github.com/apache/spark/pull/36793#discussion_r895208448
##
python/pyspark/sql/session.py:
##
@@ -952,12 +953,29 @@ def createDataFrame( # type: ignore[misc]
schema = [x.encode("utf-8") if not isinstan
xinrong-databricks commented on code in PR #36840:
URL: https://github.com/apache/spark/pull/36840#discussion_r895207134
##
python/pyspark/sql/functions.py:
##
@@ -1240,6 +1241,17 @@ def first(col: "ColumnOrName", ignorenulls: bool =
False) -> Column:
-
The functi
xinrong-databricks commented on code in PR #36793:
URL: https://github.com/apache/spark/pull/36793#discussion_r895206894
##
python/pyspark/sql/session.py:
##
@@ -952,12 +953,29 @@ def createDataFrame( # type: ignore[misc]
schema = [x.encode("utf-8") if not isinstan
AmplabJenkins commented on PR #36852:
URL: https://github.com/apache/spark/pull/36852#issuecomment-1153230519
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go
MaxGekk commented on PR #36811:
URL: https://github.com/apache/spark/pull/36811#issuecomment-1153206595
> What about exact numeric for fractional seconds?
I will implement this separately. This PR leverages existing functionality
of non-ANSI mode.
--
This is an automated message f
panbingkun commented on code in PR #36676:
URL: https://github.com/apache/spark/pull/36676#discussion_r895178269
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -61,13 +77,25 @@ private[spark] object SparkThrowableHelper {
queryContext: String = ""): String
panbingkun closed pull request #36676: [SPARK-38700][SQL][3.3] Use error
classes in the execution errors of save mode
URL: https://github.com/apache/spark/pull/36676
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
panbingkun commented on PR #36852:
URL: https://github.com/apache/spark/pull/36852#issuecomment-1153175452
FYI old pr: https://github.com/apache/spark/pull/36676, i have closed it!
master branch pr: https://github.com/apache/spark/pull/36350
ping @MaxGekk
--
This is an automated
panbingkun opened a new pull request, #36852:
URL: https://github.com/apache/spark/pull/36852
### What changes were proposed in this pull request?
Migrate the following errors in QueryExecutionErrors:
* unsupportedSaveModeError -> UNSUPPORTED_SAVE_MODE
### Why are the change
wangyum commented on PR #36851:
URL: https://github.com/apache/spark/pull/36851#issuecomment-1153170221
cc @Yaohua628 @viirya
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
wangyum opened a new pull request, #36851:
URL: https://github.com/apache/spark/pull/36851
### What changes were proposed in this pull request?
This PR makes `SchemaPruning` only pruning when `HadoopFsRelation`'s schema
contains nested column.
### Why are the changes needed?
srielau commented on PR #36811:
URL: https://github.com/apache/spark/pull/36811#issuecomment-1153161847
What about exact numeric for fractional seconds?
We keep coming back to that….
CAST(INTERVAL ‚12345.678‘ SECOND As DECIMAL(8,3))
--
This is an automated message from the Apache Git
MaxGekk commented on PR #36811:
URL: https://github.com/apache/spark/pull/36811#issuecomment-1153120819
cc @srielau
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsu
MaxGekk commented on code in PR #36676:
URL: https://github.com/apache/spark/pull/36676#discussion_r895141657
##
core/src/main/scala/org/apache/spark/ErrorInfo.scala:
##
@@ -61,13 +77,25 @@ private[spark] object SparkThrowableHelper {
queryContext: String = ""): String =
MaxGekk closed pull request #36818: [SPARK-39259][SQL][3.1] Evaluate timestamps
consistently in subqueries
URL: https://github.com/apache/spark/pull/36818
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to g
MaxGekk commented on PR #36818:
URL: https://github.com/apache/spark/pull/36818#issuecomment-1153109133
+1, LGTM. Merging to 3.1.
Thank you, @olaky.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to g
wangyum commented on PR #36709:
URL: https://github.com/apache/spark/pull/36709#issuecomment-1153101611
@dongjoon-hyun I don't think this is a regression since all these changes
are for push-based shuffles.
--
This is an automated message from the Apache Git Service.
To respond to the mes
wangyum opened a new pull request, #36850:
URL: https://github.com/apache/spark/pull/36850
### What changes were proposed in this pull request?
This PR enhances `PruneFilters` to push `EqualTo` with `Literal` to other
conditions. For example:
```sql
CREATE TABLE t1 (
id D
cxzl25 commented on code in PR #36769:
URL: https://github.com/apache/spark/pull/36769#discussion_r895128848
##
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/orc/OrcQuerySuite.scala:
##
@@ -844,6 +844,21 @@ abstract class OrcQuerySuite extends OrcQueryTest w
dongjoon-hyun commented on code in PR #36769:
URL: https://github.com/apache/spark/pull/36769#discussion_r895128188
##
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/orc/OrcQuerySuite.scala:
##
@@ -844,6 +844,21 @@ abstract class OrcQuerySuite extends OrcQuer
cxzl25 commented on code in PR #36769:
URL: https://github.com/apache/spark/pull/36769#discussion_r895125836
##
sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/orc/OrcQuerySuite.scala:
##
@@ -844,6 +844,21 @@ abstract class OrcQuerySuite extends OrcQueryTest w
79 matches
Mail list logo