wineternity commented on PR #38702:
URL: https://github.com/apache/spark/pull/38702#issuecomment-1340463701
> The change looks good to me. +CC @Ngone51
>
> Btw, do you also want to remove the `if (event.taskInfo == null) {` check
in beginning of `onTaskEnd` ?
>
> Make it a prec
wankunde commented on code in PR #38672:
URL: https://github.com/apache/spark/pull/38672#discussion_r1041811602
##
sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/LikeAnyBenchmark.scala:
##
@@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF
navinvishy commented on code in PR #38947:
URL: https://github.com/apache/spark/pull/38947#discussion_r1041817673
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/collectionOperations.scala:
##
@@ -119,21 +117,24 @@ case class Size(child: Expression, lega
beliefer commented on code in PR #38672:
URL: https://github.com/apache/spark/pull/38672#discussion_r1041818122
##
sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/LikeAnyBenchmark.scala:
##
@@ -0,0 +1,88 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF
infoankitp commented on code in PR #38865:
URL: https://github.com/apache/spark/pull/38865#discussion_r1041824037
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/collectionOperations.scala:
##
@@ -4600,3 +4600,133 @@ case class ArrayExcept(left: Expressi
LuciferYang commented on code in PR #38865:
URL: https://github.com/apache/spark/pull/38865#discussion_r1041827496
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/collectionOperations.scala:
##
@@ -4600,3 +4600,133 @@ case class ArrayExcept(left: Express
LuciferYang commented on code in PR #38865:
URL: https://github.com/apache/spark/pull/38865#discussion_r1041827973
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/collectionOperations.scala:
##
@@ -4600,3 +4600,133 @@ case class ArrayExcept(left: Express
zhengruifeng commented on PR #38914:
URL: https://github.com/apache/spark/pull/38914#issuecomment-1340493981
also cc @cloud-fan @grundprinzip
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng commented on PR #38958:
URL: https://github.com/apache/spark/pull/38958#issuecomment-1340494395
cc @grundprinzip @hvanhovell
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the spec
LuciferYang opened a new pull request, #38960:
URL: https://github.com/apache/spark/pull/38960
### What changes were proposed in this pull request?
`curdate()` throw
`QueryCompilationErrors.invalidFunctionArgumentNumberError` with `Seq.empty`
input when `expressions` is not empty, then
huaxingao commented on code in PR #38904:
URL: https://github.com/apache/spark/pull/38904#discussion_r1041840529
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/ColumnStatistics.java:
##
@@ -0,0 +1,60 @@
+/*
+ * Licensed to the Apache Software Foundati
zhengruifeng opened a new pull request, #38961:
URL: https://github.com/apache/spark/pull/38961
### What changes were proposed in this pull request?
Implement `collection` functions alphabetically, this PR contains `A` ~ `C`
except:
- aggregate, array_sort - need the support of L
huaxingao commented on code in PR #38904:
URL: https://github.com/apache/spark/pull/38904#discussion_r1041840770
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/ColumnStatistics.java:
##
@@ -0,0 +1,60 @@
+/*
+ * Licensed to the Apache Software Foundati
huaxingao commented on code in PR #38904:
URL: https://github.com/apache/spark/pull/38904#discussion_r1041840929
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/colstats/ColumnStatistics.java:
##
@@ -0,0 +1,60 @@
+/*
+ * Licensed to the Apache Software Foundati
huaxingao commented on code in PR #38904:
URL: https://github.com/apache/spark/pull/38904#discussion_r1041841092
##
sql/catalyst/src/main/java/org/apache/spark/sql/connector/read/Statistics.java:
##
@@ -31,4 +35,7 @@
public interface Statistics {
OptionalLong sizeInBytes();
zhengruifeng commented on code in PR #38961:
URL: https://github.com/apache/spark/pull/38961#discussion_r1041844991
##
python/pyspark/sql/tests/connect/test_connect_function.py:
##
@@ -63,6 +63,24 @@ class SparkConnectFunctionTests(SparkConnectFuncTestCase):
"""These test c
zhengruifeng commented on code in PR #38961:
URL: https://github.com/apache/spark/pull/38961#discussion_r1041844991
##
python/pyspark/sql/tests/connect/test_connect_function.py:
##
@@ -63,6 +63,24 @@ class SparkConnectFunctionTests(SparkConnectFuncTestCase):
"""These test c
LuciferYang commented on code in PR #38874:
URL: https://github.com/apache/spark/pull/38874#discussion_r1041845222
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/collectionOperations.scala:
##
@@ -4600,3 +4600,51 @@ case class ArrayExcept(left: Expressi
huaxingao commented on PR #38904:
URL: https://github.com/apache/spark/pull/38904#issuecomment-1340511363
> Also curious how this is to be used by Spark
The newly added `ColumnStatistics` is converted to logical `ColumnStat` in
this
[method](https://github.com/apache/spark/blob/0
huaxingao commented on code in PR #38904:
URL: https://github.com/apache/spark/pull/38904#discussion_r1041841165
##
sql/catalyst/src/test/scala/org/apache/spark/sql/connector/catalog/InMemoryBaseTable.scala:
##
@@ -294,7 +313,30 @@ abstract class InMemoryBaseTable(
val ob
zhengruifeng commented on code in PR #38961:
URL: https://github.com/apache/spark/pull/38961#discussion_r1041847513
##
python/pyspark/sql/tests/connect/test_connect_function.py:
##
@@ -413,6 +431,144 @@ def test_aggregation_functions(self):
sdf.groupBy("a").agg(SF.p
jerrypeng commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041848355
##
connector/kafka-0-10-sql/src/test/resources/error/kafka-error-classes.json:
##
@@ -0,0 +1,26 @@
+{
+ "TOPIC_PARTITIONS_IN_END_OFFSET_ARE_NOT_SAME_WITH_PREFETCHED"
jerrypeng commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041848809
##
connector/kafka-0-10-sql/src/test/resources/error/kafka-error-classes.json:
##
@@ -0,0 +1,26 @@
+{
+ "TOPIC_PARTITIONS_IN_END_OFFSET_ARE_NOT_SAME_WITH_PREFETCHED"
zhengruifeng commented on code in PR #38961:
URL: https://github.com/apache/spark/pull/38961#discussion_r1041844991
##
python/pyspark/sql/tests/connect/test_connect_function.py:
##
@@ -63,6 +63,24 @@ class SparkConnectFunctionTests(SparkConnectFuncTestCase):
"""These test c
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041856552
##
connector/kafka-0-10-sql/src/test/resources/error/kafka-error-classes.json:
##
@@ -0,0 +1,26 @@
+{
+ "TOPIC_PARTITIONS_IN_END_OFFSET_ARE_NOT_SAME_WITH_PREFETCHE
LuciferYang commented on PR #38874:
URL: https://github.com/apache/spark/pull/38874#issuecomment-1340524315
> Thanks for reviewing this. @LuciferYang let me know when you think it's
ready to go.
@HyukjinKwon @zhengruifeng The Scala part is good to me, please further
review, thanks ~
jerrypeng commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041856979
##
connector/kafka-0-10-sql/src/main/scala/org/apache/spark/sql/kafka010/KafkaMicroBatchStream.scala:
##
@@ -316,6 +320,50 @@ private[kafka010] class KafkaMicroBatchS
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041860498
##
connector/kafka-0-10-sql/src/main/scala/org/apache/spark/sql/kafka010/KafkaMicroBatchStream.scala:
##
@@ -316,6 +320,50 @@ private[kafka010] class KafkaMicroBatc
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041860498
##
connector/kafka-0-10-sql/src/main/scala/org/apache/spark/sql/kafka010/KafkaMicroBatchStream.scala:
##
@@ -316,6 +320,50 @@ private[kafka010] class KafkaMicroBatc
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041860498
##
connector/kafka-0-10-sql/src/main/scala/org/apache/spark/sql/kafka010/KafkaMicroBatchStream.scala:
##
@@ -316,6 +320,50 @@ private[kafka010] class KafkaMicroBatc
jerrypeng commented on PR #38898:
URL: https://github.com/apache/spark/pull/38898#issuecomment-1340530153
@wecharyu can you run one batch and then delete all the partitions?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
MaxGekk commented on code in PR #38937:
URL: https://github.com/apache/spark/pull/38937#discussion_r1041863502
##
sql/core/src/test/resources/sql-tests/results/except-all.sql.out:
##
@@ -230,10 +230,9 @@ org.apache.spark.sql.AnalysisException
{
"errorClass" : "NUM_COLUMNS_MI
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041864100
##
connector/kafka-0-10-sql/src/main/scala/org/apache/spark/sql/kafka010/KafkaMicroBatchStream.scala:
##
@@ -316,6 +320,50 @@ private[kafka010] class KafkaMicroBatc
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041864100
##
connector/kafka-0-10-sql/src/main/scala/org/apache/spark/sql/kafka010/KafkaMicroBatchStream.scala:
##
@@ -316,6 +320,50 @@ private[kafka010] class KafkaMicroBatc
zhengruifeng commented on code in PR #38961:
URL: https://github.com/apache/spark/pull/38961#discussion_r1041847513
##
python/pyspark/sql/tests/connect/test_connect_function.py:
##
@@ -413,6 +431,144 @@ def test_aggregation_functions(self):
sdf.groupBy("a").agg(SF.p
MaxGekk commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041866109
##
connector/kafka-0-10-sql/src/test/resources/error/kafka-error-classes.json:
##
@@ -0,0 +1,26 @@
+{
+ "TOPIC_PARTITIONS_IN_END_OFFSET_ARE_NOT_SAME_WITH_PREFETCHED" :
HeartSaVioR commented on code in PR #38911:
URL: https://github.com/apache/spark/pull/38911#discussion_r1041866576
##
connector/kafka-0-10-sql/src/test/resources/error/kafka-error-classes.json:
##
@@ -0,0 +1,26 @@
+{
+ "TOPIC_PARTITIONS_IN_END_OFFSET_ARE_NOT_SAME_WITH_PREFETCHE
jerrypeng commented on code in PR #38880:
URL: https://github.com/apache/spark/pull/38880#discussion_r1041868304
##
sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/state/RocksDB.scala:
##
@@ -310,6 +311,9 @@ class RocksDB(
"checkpoint" -> checkpointTime
zhengruifeng closed pull request #38914: [SPARK-41381][CONNECT][PYTHON]
Implement `count_distinct` and `sum_distinct` functions
URL: https://github.com/apache/spark/pull/38914
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and u
cloud-fan commented on code in PR #38942:
URL: https://github.com/apache/spark/pull/38942#discussion_r1041868772
##
sql/core/src/test/scala/org/apache/spark/sql/connector/V1WriteFallbackSuite.scala:
##
@@ -132,17 +132,21 @@ class V1WriteFallbackSuite extends QueryTest with
Shar
cloud-fan commented on PR #38942:
URL: https://github.com/apache/spark/pull/38942#issuecomment-1340537896
cc @viirya @gengliangwang
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific c
zhengruifeng commented on PR #38914:
URL: https://github.com/apache/spark/pull/38914#issuecomment-1340537857
merged into master
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific commen
grundprinzip commented on PR #38879:
URL: https://github.com/apache/spark/pull/38879#issuecomment-1340539415
@HyukjinKwon @zhengruifeng @amaliujia more opinions?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
HeartSaVioR commented on code in PR #38880:
URL: https://github.com/apache/spark/pull/38880#discussion_r1041871281
##
sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/state/RocksDB.scala:
##
@@ -310,6 +311,9 @@ class RocksDB(
"checkpoint" -> checkpointTi
beliefer opened a new pull request, #38962:
URL: https://github.com/apache/spark/pull/38962
### What changes were proposed in this pull request?
This PR adds document for `DataFrame.summary`.
### Why are the changes needed?
This PR adds document for `DataFrame.summary`.
wankunde commented on PR #38672:
URL: https://github.com/apache/spark/pull/38672#issuecomment-1340543632
After `LikeSimplification`, the combination of multiple like expressions
with `OR` can be pushdown to parquet reader, while `like any` can not.
So close this PR.
--
This is an auto
wankunde closed pull request #38672: [SPARK-41159][SQL] Optimize like any and
like all expressions
URL: https://github.com/apache/spark/pull/38672
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
301 - 347 of 347 matches
Mail list logo