MaxGekk closed pull request #38484: [SPARK-40998][SQL] Rename the error class
`_LEGACY_ERROR_TEMP_0040` to `INVALID_IDENTIFIER`
URL: https://github.com/apache/spark/pull/38484
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
MaxGekk commented on PR #38484:
URL: https://github.com/apache/spark/pull/38484#issuecomment-1301673423
Merging to master. Thank you, @LuciferYang @srielau and @cloud-fan for
review.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
ming95 commented on code in PR #38356:
URL: https://github.com/apache/spark/pull/38356#discussion_r1012507610
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/V1Writes.scala:
##
@@ -178,7 +180,15 @@ object V1WritesUtils {
} else {
// We
ming95 commented on PR #38356:
URL: https://github.com/apache/spark/pull/38356#issuecomment-1301649899
> @allisonwang-db can you elaborate on mapping `write.requiredOrdering` to
the projected columns via `attrMap` that you introduced in
cxzl25 commented on PR #38489:
URL: https://github.com/apache/spark/pull/38489#issuecomment-1301639710
## Current
### enable codegen
https://user-images.githubusercontent.com/3898450/199650431-d6443f45-03f9-489c-b1de-72c619acb37e.png;>
### disable codegen
HyukjinKwon commented on code in PR #38453:
URL: https://github.com/apache/spark/pull/38453#discussion_r1012493266
##
python/pyspark/sql/tests/connect/test_connect_plan_only.py:
##
@@ -128,6 +128,16 @@ def test_all_the_plans(self):
self.assertIsNotNone(plan.root, "Root
cxzl25 opened a new pull request, #38489:
URL: https://github.com/apache/spark/pull/38489
### What changes were proposed in this pull request?
BHJ LeftAnti does not update numOutputRows when codegen is disabled
### Why are the changes needed?
PR #29104 Only update
HyukjinKwon closed pull request #38453: [SPARK-40977][CONNECT][PYTHON] Complete
Support for Union in Python client
URL: https://github.com/apache/spark/pull/38453
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
HyukjinKwon commented on PR #38453:
URL: https://github.com/apache/spark/pull/38453#issuecomment-1301637894
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
HyukjinKwon commented on code in PR #38485:
URL: https://github.com/apache/spark/pull/38485#discussion_r1012492270
##
python/pyspark/sql/connect/client.py:
##
@@ -42,6 +43,125 @@
logging.basicConfig(level=logging.INFO)
+class ChannelBuilder:
+"""
+This is a helper
HyukjinKwon commented on code in PR #38485:
URL: https://github.com/apache/spark/pull/38485#discussion_r1012492460
##
python/pyspark/sql/connect/client.py:
##
@@ -42,6 +43,125 @@
logging.basicConfig(level=logging.INFO)
+class ChannelBuilder:
+"""
+This is a helper
HyukjinKwon commented on code in PR #38485:
URL: https://github.com/apache/spark/pull/38485#discussion_r1012492270
##
python/pyspark/sql/connect/client.py:
##
@@ -42,6 +43,125 @@
logging.basicConfig(level=logging.INFO)
+class ChannelBuilder:
+"""
+This is a helper
gaoyajun02 commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012466373
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
HyukjinKwon commented on code in PR #38462:
URL: https://github.com/apache/spark/pull/38462#discussion_r1012491071
##
python/pyspark/sql/connect/_typing.py:
##
@@ -15,5 +15,7 @@
# limitations under the License.
#
from typing import Union
+from datetime import date, time,
HyukjinKwon commented on code in PR #38462:
URL: https://github.com/apache/spark/pull/38462#discussion_r1012489260
##
python/pyspark/sql/connect/column.py:
##
@@ -99,11 +101,59 @@ def to_plan(self, session: Optional["RemoteSparkSession"])
-> "proto.Expression"
gaoyajun02 commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012466373
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
gaoyajun02 commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012466373
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
itholic commented on code in PR #38447:
URL: https://github.com/apache/spark/pull/38447#discussion_r1012485332
##
sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryParsingErrors.scala:
##
@@ -608,8 +608,12 @@ private[sql] object QueryParsingErrors extends
itholic commented on code in PR #38447:
URL: https://github.com/apache/spark/pull/38447#discussion_r1012485332
##
sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryParsingErrors.scala:
##
@@ -608,8 +608,12 @@ private[sql] object QueryParsingErrors extends
itholic commented on code in PR #38447:
URL: https://github.com/apache/spark/pull/38447#discussion_r1012485332
##
sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryParsingErrors.scala:
##
@@ -608,8 +608,12 @@ private[sql] object QueryParsingErrors extends
itholic commented on code in PR #38447:
URL: https://github.com/apache/spark/pull/38447#discussion_r1012485332
##
sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryParsingErrors.scala:
##
@@ -608,8 +608,12 @@ private[sql] object QueryParsingErrors extends
itholic commented on code in PR #38447:
URL: https://github.com/apache/spark/pull/38447#discussion_r1012485332
##
sql/catalyst/src/main/scala/org/apache/spark/sql/errors/QueryParsingErrors.scala:
##
@@ -608,8 +608,12 @@ private[sql] object QueryParsingErrors extends
amaliujia opened a new pull request, #38488:
URL: https://github.com/apache/spark/pull/38488
### What changes were proposed in this pull request?
1. Add `take(n)` API.
2. Change `head(n)` API to return `Union[Optional[Row], List[Row]]`.
### Why are the changes
beliefer commented on PR #38466:
URL: https://github.com/apache/spark/pull/38466#issuecomment-1301628080
This PR is only good for q93.
`q93 251.88 281.774 29.894 111.87%`
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
beliefer closed pull request #38466: [WIP][SPARK-40986][SQL] Add aggregate to
reduce the data size for bloom filter
URL: https://github.com/apache/spark/pull/38466
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
amaliujia commented on code in PR #38475:
URL: https://github.com/apache/spark/pull/38475#discussion_r1012477099
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala:
##
@@ -123,6 +125,24 @@ class SparkConnectPlanner(plan:
gaoyajun02 commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012468152
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
gaoyajun02 commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012466373
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
zhengruifeng commented on code in PR #38475:
URL: https://github.com/apache/spark/pull/38475#discussion_r1012465284
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala:
##
@@ -123,6 +125,24 @@ class SparkConnectPlanner(plan:
amaliujia commented on code in PR #38475:
URL: https://github.com/apache/spark/pull/38475#discussion_r1012464469
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala:
##
@@ -123,6 +125,24 @@ class SparkConnectPlanner(plan:
jzhuge commented on PR #37556:
URL: https://github.com/apache/spark/pull/37556#issuecomment-1301598157
Looking at error:
```
SparkThrowableSuite.Error classes are correctly formatted
```
--
This is an automated message from the Apache Git Service.
To respond to the message,
LuciferYang commented on code in PR #38481:
URL: https://github.com/apache/spark/pull/38481#discussion_r1012461316
##
project/build.properties:
##
@@ -15,4 +15,4 @@
# limitations under the License.
#
# Please update the version in appveyor-install-dependencies.ps1 together.
amaliujia commented on code in PR #38485:
URL: https://github.com/apache/spark/pull/38485#discussion_r1012460587
##
python/pyspark/sql/connect/client.py:
##
@@ -42,6 +43,125 @@
logging.basicConfig(level=logging.INFO)
+class ChannelBuilder:
+"""
+This is a helper
LuciferYang commented on PR #38465:
URL: https://github.com/apache/spark/pull/38465#issuecomment-1301588097
Thanks @srowen @HyukjinKwon @itholic
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
LuciferYang commented on PR #38469:
URL: https://github.com/apache/spark/pull/38469#issuecomment-1301587767
Thanks @srowen @dongjoon-hyun
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
pan3793 commented on PR #38483:
URL: https://github.com/apache/spark/pull/38483#issuecomment-1301579448
cc @dongjoon-hyun @Yikun, would you please take a look?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
amaliujia commented on code in PR #38475:
URL: https://github.com/apache/spark/pull/38475#discussion_r1012449031
##
connector/connect/src/main/protobuf/spark/connect/relations.proto:
##
@@ -250,3 +251,15 @@ message SubqueryAlias {
// Optional. Qualifier of the alias.
zhengruifeng commented on code in PR #38468:
URL: https://github.com/apache/spark/pull/38468#discussion_r1012441577
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/service/SparkConnectStreamHandler.scala:
##
@@ -117,7 +121,38 @@ class
HyukjinKwon closed pull request #38487: [SPARK-40995][CONNECT][DOC][FOLLOW-UP]
Fix the type in the doc name
URL: https://github.com/apache/spark/pull/38487
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
HyukjinKwon commented on PR #38487:
URL: https://github.com/apache/spark/pull/38487#issuecomment-1301565811
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
amaliujia commented on PR #38487:
URL: https://github.com/apache/spark/pull/38487#issuecomment-1301565529
R: @HyukjinKwon
cc @grundprinzip
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
HyukjinKwon closed pull request #38472: [SPARK-40989][CONNECT][PYTHON][TESTS]
Improve `session.sql` testing coverage in Python client
URL: https://github.com/apache/spark/pull/38472
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
amaliujia opened a new pull request, #38487:
URL: https://github.com/apache/spark/pull/38487
### What changes were proposed in this pull request?
Fix the type in the doc filename: `coient` -> `client`.
### Why are the changes needed?
Fix typo.
### Does
HyukjinKwon commented on PR #38472:
URL: https://github.com/apache/spark/pull/38472#issuecomment-1301565409
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
cloud-fan commented on PR #28450:
URL: https://github.com/apache/spark/pull/28450#issuecomment-1301565132
At that time, the ecosystem does not fully support standard parquet
timestamp yet. We can recheck now. If the latest version of popular data
systems (Hive, Presto, Flink, etc.) all
holdenk commented on PR #37556:
URL: https://github.com/apache/spark/pull/37556#issuecomment-1301564329
pending CI and any other concerns I plan to merge this on Friday.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
ming95 commented on PR #38358:
URL: https://github.com/apache/spark/pull/38358#issuecomment-1301560852
> @kristopherkane 3.1 is EOL unfortunately.
>
> @ming95 Does this PR fix your problem?
I have no problem with this issue. Another similar issue I found
hvanhovell commented on code in PR #38468:
URL: https://github.com/apache/spark/pull/38468#discussion_r1012428719
##
connector/connect/src/main/scala/org/apache/spark/sql/connect/service/SparkConnectStreamHandler.scala:
##
@@ -117,7 +121,38 @@ class
srowen commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1301541574
Probably, if it's internal, and being mutable/immutable doesn't matter in
the API
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
LuciferYang commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1301540958
@srowen Does this mean that for similar cases, if it is an internal api, we
can explicitly specify `Seq` as `collection.Seq` to avoid unnecessary memory
copying for Scala 2.13? There
ming95 commented on code in PR #38358:
URL: https://github.com/apache/spark/pull/38358#discussion_r1012409686
##
sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormatWriter.scala:
##
@@ -187,8 +188,17 @@ object FileFormatWriter extends Logging {
HyukjinKwon closed pull request #38477: [SPARK-40993][CONNECT][PYTHON][DOCS]
Migrate markdown style README to PySpark Development Documentation
URL: https://github.com/apache/spark/pull/38477
--
This is an automated message from the Apache Git Service.
To respond to the message, please log
HyukjinKwon commented on PR #38477:
URL: https://github.com/apache/spark/pull/38477#issuecomment-1301524099
Fixed in https://github.com/apache/spark/pull/38470 for now.
Per https://github.com/apache/spark/pull/38477#issuecomment-1299705536, let
me close this for now.
--
This is an
HyukjinKwon closed pull request #38470: [SPARK-40995] [CONNECT] [DOC] Defining
Spark Connect Client Connection String
URL: https://github.com/apache/spark/pull/38470
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
HyukjinKwon commented on PR #38470:
URL: https://github.com/apache/spark/pull/38470#issuecomment-1301523665
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
kelvinjian-db commented on PR #38486:
URL: https://github.com/apache/spark/pull/38486#issuecomment-1301517613
cc @sigmod @cloud-fan
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
github-actions[bot] commented on PR #37163:
URL: https://github.com/apache/spark/pull/37163#issuecomment-1301517583
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
github-actions[bot] commented on PR #37083:
URL: https://github.com/apache/spark/pull/37083#issuecomment-1301517598
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
github-actions[bot] closed pull request #37259: spark-submit: throw an error
when duplicate argument is provided
URL: https://github.com/apache/spark/pull/37259
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
github-actions[bot] commented on PR #37265:
URL: https://github.com/apache/spark/pull/37265#issuecomment-1301517562
We're closing this PR because it hasn't been updated in a while. This isn't
a judgement on the merit of the PR in any way. It's just a way of keeping the
PR queue manageable.
SandishKumarHN commented on code in PR #38344:
URL: https://github.com/apache/spark/pull/38344#discussion_r1012398593
##
core/src/main/resources/error/error-classes.json:
##
@@ -29,12 +44,22 @@
],
"sqlState" : "22007"
},
+ "CANNOT_LOAD_PROTOBUF_CLASS" : {
+
srielau commented on code in PR #38344:
URL: https://github.com/apache/spark/pull/38344#discussion_r1012385639
##
core/src/main/resources/error/error-classes.json:
##
@@ -742,6 +832,11 @@
],
"sqlState" : "22023"
},
+ "SQL_TYPE_TO_PROTOBUF_ENUM_TYPE_ERROR" : {
AmplabJenkins commented on PR #38475:
URL: https://github.com/apache/spark/pull/38475#issuecomment-1301480681
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
AmplabJenkins commented on PR #38472:
URL: https://github.com/apache/spark/pull/38472#issuecomment-1301480708
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
AmplabJenkins commented on PR #38477:
URL: https://github.com/apache/spark/pull/38477#issuecomment-1301480649
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
AmplabJenkins commented on PR #38470:
URL: https://github.com/apache/spark/pull/38470#issuecomment-1301480736
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
mridulm commented on PR #38333:
URL: https://github.com/apache/spark/pull/38333#issuecomment-1301464492
If there are hardware issues which are causing failures - it is better to
move the nodes to deny list and prevent them from getting used: we will keep
seeing more failures, including for
zhouyejoe commented on code in PR #37638:
URL: https://github.com/apache/spark/pull/37638#discussion_r1012349885
##
common/network-shuffle/src/test/java/org/apache/spark/network/shuffle/RemoteBlockPushResolverSuite.java:
##
@@ -413,6 +437,7 @@ public void
mridulm commented on code in PR #36165:
URL: https://github.com/apache/spark/pull/36165#discussion_r1012357823
##
core/src/main/scala/org/apache/spark/executor/ShuffleReadMetrics.scala:
##
@@ -146,6 +268,16 @@ private[spark] class TempShuffleReadMetrics extends
mridulm commented on PR #38377:
URL: https://github.com/apache/spark/pull/38377#issuecomment-1301454290
I am not sure I follow - can you give an example ?
Based on what the PR is doing, I would expect both to be equivalent.
--
This is an automated message from the Apache Git Service.
otterc commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012336511
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
otterc commented on code in PR #38333:
URL: https://github.com/apache/spark/pull/38333#discussion_r1012336511
##
core/src/main/scala/org/apache/spark/storage/ShuffleBlockFetcherIterator.scala:
##
@@ -794,7 +794,15 @@ final class ShuffleBlockFetcherIterator(
//
zhouyejoe commented on code in PR #36165:
URL: https://github.com/apache/spark/pull/36165#discussion_r1012326931
##
core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala:
##
@@ -227,6 +227,16 @@ class TaskMetrics private[spark] () extends Serializable {
zhouyejoe commented on code in PR #36165:
URL: https://github.com/apache/spark/pull/36165#discussion_r1012322498
##
core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala:
##
@@ -227,6 +227,16 @@ class TaskMetrics private[spark] () extends Serializable {
kelvinjian-db opened a new pull request, #38486:
URL: https://github.com/apache/spark/pull/38486
### What changes were proposed in this pull request?
We change `CommandResult` to extend `LeafCommand` (which extends `Command`)
instead of `LeafNode`.
### Why are the
dongjoon-hyun commented on PR #38433:
URL: https://github.com/apache/spark/pull/38433#issuecomment-1301158936
Thank you so much, @ben-zhang .
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
ben-zhang commented on PR #38433:
URL: https://github.com/apache/spark/pull/38433#issuecomment-1301158208
@dongjoon-hyun , sounds good. I will get back to you when the Databricks
docs are updated.
--
This is an automated message from the Apache Git Service.
To respond to the message,
grundprinzip opened a new pull request, #38485:
URL: https://github.com/apache/spark/pull/38485
### What changes were proposed in this pull request?
This PR implements the connection string for Spark Connect clients according
to the documentation.
### Why are the changes
HeartSaVioR commented on PR #38430:
URL: https://github.com/apache/spark/pull/38430#issuecomment-1301097792
@jerrypeng Could you please address post-review comments as followup PR? You
don't need a new JIRA ticket.
(Sigh, I forgot to check. Please make sure your PR title has a
clee704 commented on PR #28450:
URL: https://github.com/apache/spark/pull/28450#issuecomment-1301097763
Can someone explain why we reverted to INT96? I read
https://issues.apache.org/jira/browse/SPARK-31085 but want to know how the
discussion happened. To me the cost of breaking the API
MaxGekk commented on PR #38484:
URL: https://github.com/apache/spark/pull/38484#issuecomment-1301046405
@srielau @itholic @panbingkun @LuciferYang @cloud-fan Could you review this
PR, please.
--
This is an automated message from the Apache Git Service.
To respond to the message, please
swamirishi commented on PR #38377:
URL: https://github.com/apache/spark/pull/38377#issuecomment-1301028102
> Makes sense ... why not simply `val dfsLogFile = new Path(rootDir, appId +
DRIVER_LOG_FILE_SUFFIX)` instead btw ? If we do necessarily need the fully
qualified path, we can use
leewyang commented on code in PR #37734:
URL: https://github.com/apache/spark/pull/37734#discussion_r1012129510
##
python/pyspark/ml/functions.py:
##
@@ -106,6 +117,474 @@ def array_to_vector(col: Column) -> Column:
return
AmplabJenkins commented on PR #38482:
URL: https://github.com/apache/spark/pull/38482#issuecomment-1300983786
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
AmplabJenkins commented on PR #38483:
URL: https://github.com/apache/spark/pull/38483#issuecomment-1300983649
Can one of the admins verify this patch?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
srowen commented on PR #38465:
URL: https://github.com/apache/spark/pull/38465#issuecomment-1300728818
Merged to master
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
srowen closed pull request #38465: [SPARK-40985][BUILD] Upgrade RoaringBitmap
to 0.9.35
URL: https://github.com/apache/spark/pull/38465
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
srowen commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1300679989
The latter sounds better right? is there any downside?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
eejbyfeldt commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1300650411
> OK, I think we can't accept that much perf degradation. If there's a
simple way to refactor the code to make both faster, that seems OK. Ideally we
avoid separate code branches for
srowen commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1300620236
OK, I think we can't accept that much perf degradation. If there's a simple
way to refactor the code to make both faster, that seems OK. Ideally we avoid
separate code branches for 2.12
LuciferYang commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1300602808
> What's the current state here - this change is still much slower on 2.12?
@srowen
```
The overall trend is consistent with local tests. Using toIndexedSeq
improves the
dongjoon-hyun commented on code in PR #38430:
URL: https://github.com/apache/spark/pull/38430#discussion_r1011874757
##
sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala:
##
@@ -277,10 +295,34 @@ class HDFSMetadataLog[T <: AnyRef :
srowen commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1300578770
What's the current state here - this change is still much slower on 2.12?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
dongjoon-hyun commented on code in PR #38430:
URL: https://github.com/apache/spark/pull/38430#discussion_r1011874757
##
sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala:
##
@@ -277,10 +295,34 @@ class HDFSMetadataLog[T <: AnyRef :
MaxGekk opened a new pull request, #38484:
URL: https://github.com/apache/spark/pull/38484
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
### How
dongjoon-hyun commented on code in PR #38430:
URL: https://github.com/apache/spark/pull/38430#discussion_r1011869341
##
sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala:
##
@@ -168,7 +191,13 @@ class HDFSMetadataLog[T <: AnyRef :
dongjoon-hyun commented on code in PR #38430:
URL: https://github.com/apache/spark/pull/38430#discussion_r1011867576
##
sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala:
##
@@ -64,6 +67,17 @@ class HDFSMetadataLog[T <: AnyRef :
dongjoon-hyun commented on code in PR #38430:
URL: https://github.com/apache/spark/pull/38430#discussion_r1011867108
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -2007,6 +2007,14 @@ object SQLConf {
.booleanConf
LuciferYang commented on PR #38427:
URL: https://github.com/apache/spark/pull/38427#issuecomment-1300477025
Move forward?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
MaksGS09 commented on PR #37206:
URL: https://github.com/apache/spark/pull/37206#issuecomment-1300440849
Hi!
Any updates on this?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
1 - 100 of 165 matches
Mail list logo