yaooqinn commented on code in PR #42575:
URL: https://github.com/apache/spark/pull/42575#discussion_r1299641664
##
core/src/main/scala/org/apache/spark/ui/exec/ExecutorThreadDumpPage.scala:
##
@@ -67,18 +69,17 @@ private[ui] class ExecutorThreadDumpPage(
Updated at
ion-elgreco commented on PR #38624:
URL: https://github.com/apache/spark/pull/38624#issuecomment-1685688494
> I get that `cogroup` might not be possible tho. But we can just convert
pandas back to arrow batches easily. Is this really required for some scenario?
IIRC this is only useful for
yaooqinn commented on PR #42481:
URL: https://github.com/apache/spark/pull/42481#issuecomment-1685688005
thanks, merged to master
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
yaooqinn closed pull request #42481: [SPARK-44801][SQL][UI] Capture analyzing
failed queries in Listener and UI
URL: https://github.com/apache/spark/pull/42481
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
LuciferYang commented on PR #42580:
URL: https://github.com/apache/spark/pull/42580#issuecomment-1685685065
cc @dongjoon-hyun FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
cloud-fan commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299615448
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Expression.scala:
##
@@ -368,6 +368,15 @@ abstract class Expression extends
LuciferYang opened a new pull request, #42580:
URL: https://github.com/apache/spark/pull/42580
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
###
HyukjinKwon closed pull request #42579: [SPARK-44887][DOCS] Fix wildcard import
`from pyspark.sql.functions import *` in `Quick Start` Examples
URL: https://github.com/apache/spark/pull/42579
--
This is an automated message from the Apache Git Service.
To respond to the message, please log
HyukjinKwon commented on PR #42579:
URL: https://github.com/apache/spark/pull/42579#issuecomment-1685620282
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
HyukjinKwon commented on code in PR #42462:
URL: https://github.com/apache/spark/pull/42462#discussion_r1299583398
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/xml/StaxXmlGenerator.scala:
##
@@ -83,21 +86,21 @@ private[xml] object StaxXmlGenerator {
def
HyukjinKwon commented on code in PR #42462:
URL: https://github.com/apache/spark/pull/42462#discussion_r1299583398
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/xml/StaxXmlGenerator.scala:
##
@@ -83,21 +86,21 @@ private[xml] object StaxXmlGenerator {
def
HyukjinKwon commented on code in PR #42462:
URL: https://github.com/apache/spark/pull/42462#discussion_r1299581325
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/functions.scala:
##
@@ -7227,6 +7227,150 @@ object functions {
*/
def to_csv(e: Column):
HyukjinKwon commented on code in PR #42462:
URL: https://github.com/apache/spark/pull/42462#discussion_r1299580956
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/DataFrameReader.scala:
##
@@ -392,6 +392,46 @@ class DataFrameReader private[sql]
HyukjinKwon commented on code in PR #42462:
URL: https://github.com/apache/spark/pull/42462#discussion_r1299580676
##
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/DataFrameReader.scala:
##
@@ -392,6 +392,46 @@ class DataFrameReader private[sql]
HyukjinKwon commented on PR #38624:
URL: https://github.com/apache/spark/pull/38624#issuecomment-1685602103
adding @viirya @ueshin @BryanCutler in case you guys have some thought on
this PR.
--
This is an automated message from the Apache Git Service.
To respond to the message, please
cloud-fan commented on code in PR #41782:
URL: https://github.com/apache/spark/pull/41782#discussion_r1299573483
##
sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/WritableColumnVector.java:
##
@@ -955,4 +986,8 @@ protected WritableColumnVector(int capacity,
HyukjinKwon commented on PR #38624:
URL: https://github.com/apache/spark/pull/38624#issuecomment-1685601694
Yeah, I meant `df.repartition(grouping_cols).mapInArrow() `
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
itholic closed pull request #42528: [SPARK-44844][BUILD] Exclude
`python/build/*` path for local `lint-python` testing
URL: https://github.com/apache/spark/pull/42528
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
itholic commented on PR #42528:
URL: https://github.com/apache/spark/pull/42528#issuecomment-1685600378
IIRC they were generated when I upgrade the pip packages by running `pip
install -r dev/requirements.txt`, but seems not to be reproducible now for some
reason. Let me just close this
HyukjinKwon commented on PR #42377:
URL: https://github.com/apache/spark/pull/42377#issuecomment-1685594500
Would be great if we have the user-facing exception (and stacktrace) example
at the PR description.
--
This is an automated message from the Apache Git Service.
To respond to the
cloud-fan closed pull request #41335: [SPARK-43205][DOCS][SQL][FOLLOWUP]
IDENTIFIER clause docs
URL: https://github.com/apache/spark/pull/41335
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
cloud-fan commented on PR #41335:
URL: https://github.com/apache/spark/pull/41335#issuecomment-1685582467
the test failure is unrelated, thanks, merging to master!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
pan3793 commented on code in PR #42575:
URL: https://github.com/apache/spark/pull/42575#discussion_r1299557854
##
core/src/main/scala/org/apache/spark/ui/exec/ExecutorThreadDumpPage.scala:
##
@@ -67,18 +69,17 @@ private[ui] class ExecutorThreadDumpPage(
Updated at
cloud-fan closed pull request #41100: [SPARK-43420][SQL] Make
DisableUnnecessaryBucketedScan smart with table cache
URL: https://github.com/apache/spark/pull/41100
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
cloud-fan commented on code in PR #40390:
URL: https://github.com/apache/spark/pull/40390#discussion_r1299552593
##
sql/core/src/test/scala/org/apache/spark/sql/sources/DisableUnnecessaryBucketedScanSuite.scala:
##
@@ -244,7 +244,8 @@ abstract class
cloud-fan commented on code in PR #40390:
URL: https://github.com/apache/spark/pull/40390#discussion_r1299551915
##
sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala:
##
@@ -512,6 +512,9 @@ class CachedTableSuite extends QueryTest with SQLTestUtils
*
wankunde commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299551600
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodegenFallback.scala:
##
@@ -46,21 +46,54 @@ trait CodegenFallback extends
gengliangwang commented on PR #42553:
URL: https://github.com/apache/spark/pull/42553#issuecomment-1685566875
TBH `%s/%s/statistics?id=%s` is more "restful".
(And, of course it would be totally restful if it is `%s/%s/statistics/%s`,
but we can't make such changes.)
--
This is an
cloud-fan commented on code in PR #41782:
URL: https://github.com/apache/spark/pull/41782#discussion_r1299550767
##
sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/WritableColumnVector.java:
##
@@ -846,7 +849,14 @@ public final void addElementsAppended(int num)
cloud-fan commented on code in PR #41782:
URL: https://github.com/apache/spark/pull/41782#discussion_r1299550466
##
sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OffHeapColumnVector.java:
##
@@ -84,9 +84,7 @@ public long valuesNativeAddress() {
return
HyukjinKwon commented on PR #42455:
URL: https://github.com/apache/spark/pull/42455#issuecomment-1685563329
Fixed in https://github.com/apache/spark/pull/42464
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
HyukjinKwon closed pull request #42455: [DRAFT] Fix Spark Connect Behavior for
Default Session
URL: https://github.com/apache/spark/pull/42455
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
HyukjinKwon commented on PR #42467:
URL: https://github.com/apache/spark/pull/42467#issuecomment-1685562669
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
cloud-fan commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299549288
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodegenFallback.scala:
##
@@ -46,21 +46,54 @@ trait CodegenFallback extends
HyukjinKwon closed pull request #42471: [SPARK-44785][SQL][CONNECT] Convert
common alreadyExistsExceptions and noSuchExceptions
URL: https://github.com/apache/spark/pull/42471
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
HyukjinKwon commented on PR #42471:
URL: https://github.com/apache/spark/pull/42471#issuecomment-1685562030
Merged to master and branch-3.5.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
cloud-fan commented on PR #42534:
URL: https://github.com/apache/spark/pull/42534#issuecomment-1685560046
late LGTM
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To
HyukjinKwon commented on code in PR #42475:
URL: https://github.com/apache/spark/pull/42475#discussion_r1299546504
##
sql/core/src/main/scala/org/apache/spark/sql/execution/WholeStageCodegenEvaluatorFactory.scala:
##
@@ -41,7 +41,7 @@ class WholeStageCodegenEvaluatorFactory(
HyukjinKwon commented on PR #42498:
URL: https://github.com/apache/spark/pull/42498#issuecomment-1685557081
Seems like it dose trigger sth :-).
https://github.com/grundprinzip/spark/actions/runs/5870189292/job/15916811394#step:12:1425
--
This is an automated message from the Apache Git
goodwanghan commented on PR #38624:
URL: https://github.com/apache/spark/pull/38624#issuecomment-1685556987
> qq, can't we workaround by `df.repartitionByExpression().mapInArrow()` for
`groupby` case?
Hi @HyukjinKwon i understand what you mean, I am curious if df.repartition
will
hdaikoku commented on PR #42572:
URL: https://github.com/apache/spark/pull/42572#issuecomment-1685554405
This seems to be the same issue as https://github.com/apache/spark/pull/42426
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
HyukjinKwon commented on PR #42528:
URL: https://github.com/apache/spark/pull/42528#issuecomment-1685554336
how were `python/build/*` generated?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to
wankunde commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299537222
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodegenFallback.scala:
##
@@ -46,21 +46,54 @@ trait CodegenFallback extends
itholic commented on code in PR #42551:
URL: https://github.com/apache/spark/pull/42551#discussion_r1299537975
##
python/pyspark/pandas/namespace.py:
##
@@ -985,11 +975,6 @@ def read_excel(
* If list of string, then indicates list of column names to be parsed.
zhengruifeng commented on PR #42579:
URL: https://github.com/apache/spark/pull/42579#issuecomment-1685552671
cc @HyukjinKwon @allisonwang-db
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
HyukjinKwon closed pull request #42534: [SPARK-44868][SQL] Convert datetime to
string by `to_char`/`to_varchar`
URL: https://github.com/apache/spark/pull/42534
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
HyukjinKwon commented on PR #42534:
URL: https://github.com/apache/spark/pull/42534#issuecomment-1685552067
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
wankunde commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299537222
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodegenFallback.scala:
##
@@ -46,21 +46,54 @@ trait CodegenFallback extends
zhengruifeng commented on PR #42579:
URL: https://github.com/apache/spark/pull/42579#issuecomment-1685551050
there are two wildcard import under `docs`:
```
(spark_dev_310) ➜ spark git:(master) ag -i 'import \*' docs
docs/sql-ref-datatypes.md
117:from pyspark.sql.types import *
HyukjinKwon commented on code in PR #42541:
URL: https://github.com/apache/spark/pull/42541#discussion_r1299534962
##
python/pyspark/sql/types.py:
##
@@ -442,7 +442,7 @@ def needConversion(self) -> bool:
def toInternal(self, dt: datetime.timedelta) -> Optional[int]:
HyukjinKwon commented on code in PR #42541:
URL: https://github.com/apache/spark/pull/42541#discussion_r1299533479
##
python/pyspark/sql/types.py:
##
@@ -442,7 +442,7 @@ def needConversion(self) -> bool:
def toInternal(self, dt: datetime.timedelta) -> Optional[int]:
zhengruifeng opened a new pull request, #42579:
URL: https://github.com/apache/spark/pull/42579
### What changes were proposed in this pull request?
Fix wildcard import `from pyspark.sql.functions import *`
### Why are the changes needed?
to follow the [PEP 8 - Style Guide
hdaikoku commented on PR #42426:
URL: https://github.com/apache/spark/pull/42426#issuecomment-1685547323
> To make sure I understand correctly - there is an OOM which is thrown,
which happens to be within `initiateRetry` and so shuffle fetch stalled
indefinitely, and so task appeared to be
HyukjinKwon commented on PR #42541:
URL: https://github.com/apache/spark/pull/42541#issuecomment-1685546533
Seems pretty good - mind retriggering
https://github.com/hdaly0/spark/runs/15986658618 please?
Also please fix the PR title (see also
itholic opened a new pull request, #42578:
URL: https://github.com/apache/spark/pull/42578
### What changes were proposed in this pull request?
This PR followups for https://github.com/apache/spark/pull/42525.
### Why are the changes needed?
To fill
cloud-fan commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299530738
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/CodegenFallback.scala:
##
@@ -46,21 +46,54 @@ trait CodegenFallback extends
HyukjinKwon commented on code in PR #42550:
URL: https://github.com/apache/spark/pull/42550#discussion_r1299527368
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/service/ExecuteEventsManager.scala:
##
@@ -278,6 +278,7 @@ case class
wankunde commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299526542
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/EquivalentExpressions.scala:
##
@@ -150,15 +150,15 @@ class EquivalentExpressions(
// 1.
HyukjinKwon commented on code in PR #42550:
URL: https://github.com/apache/spark/pull/42550#discussion_r1299526592
##
connector/connect/server/src/main/scala/org/apache/spark/sql/connect/service/ExecuteEventsManager.scala:
##
@@ -278,6 +278,7 @@ case class
wankunde commented on code in PR #42450:
URL: https://github.com/apache/spark/pull/42450#discussion_r1299526542
##
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/EquivalentExpressions.scala:
##
@@ -150,15 +150,15 @@ class EquivalentExpressions(
// 1.
HyukjinKwon commented on code in PR #42551:
URL: https://github.com/apache/spark/pull/42551#discussion_r1299518780
##
python/pyspark/pandas/namespace.py:
##
@@ -985,11 +975,6 @@ def read_excel(
* If list of string, then indicates list of column names to be parsed.
imback82 opened a new pull request, #42577:
URL: https://github.com/apache/spark/pull/42577
### What changes were proposed in this pull request?
### Why are the changes needed?
### Does this PR introduce _any_ user-facing change?
### How
HyukjinKwon commented on PR #42553:
URL: https://github.com/apache/spark/pull/42553#issuecomment-1685526203
cc @gengliangwang and @sarutak FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
HyukjinKwon commented on PR #42554:
URL: https://github.com/apache/spark/pull/42554#issuecomment-1685525915
Mind filing a JIRA please? See also
https://spark.apache.org/contributing.html
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on
HyukjinKwon commented on PR #42556:
URL: https://github.com/apache/spark/pull/42556#issuecomment-1685525416
@allanf-db FYI
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
zekai-li commented on code in PR #42529:
URL: https://github.com/apache/spark/pull/42529#discussion_r1299512887
##
resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ClientSuite.scala:
##
@@ -670,7 +670,7 @@ class ClientSuite extends SparkFunSuite with Matchers
dzypersonal commented on PR #36162:
URL: https://github.com/apache/spark/pull/36162#issuecomment-1685522302
> It helps in two cases @weixiuli - the example you gave (generated input
(like range()), etc where there is no input metrics). It also helps when
reading shuffle input where there
HyukjinKwon commented on PR #42566:
URL: https://github.com/apache/spark/pull/42566#issuecomment-1685515524
Yeah, let's probably don't backport to 3.4 although it's sort of safe.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to
zekai-li commented on code in PR #42529:
URL: https://github.com/apache/spark/pull/42529#discussion_r1299504495
##
resource-managers/yarn/src/test/scala/org/apache/spark/deploy/yarn/ClientSuite.scala:
##
@@ -670,7 +670,7 @@ class ClientSuite extends SparkFunSuite with Matchers
HyukjinKwon commented on PR #42575:
URL: https://github.com/apache/spark/pull/42575#issuecomment-1685501692
also @sarutak and @jasonli-db
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
wankunde commented on code in PR #41782:
URL: https://github.com/apache/spark/pull/41782#discussion_r1299497876
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -487,6 +487,25 @@ object SQLConf {
.intConf
.createWithDefault(1)
+
srowen commented on code in PR #42428:
URL: https://github.com/apache/spark/pull/42428#discussion_r1299494936
##
python/docs/source/_static/versions.json:
##
@@ -0,0 +1,278 @@
+[
Review Comment:
Yes, let's just start with latest versions even, as a convenience to switch.
HyukjinKwon commented on code in PR #42548:
URL: https://github.com/apache/spark/pull/42548#discussion_r1299494819
##
python/pyspark/sql/tests/connect/test_connect_basic.py:
##
@@ -3347,6 +3347,22 @@ def
test_can_create_multiple_sessions_to_different_remotes(self):
HyukjinKwon commented on PR #42548:
URL: https://github.com/apache/spark/pull/42548#issuecomment-1685489213
Three of them are actually runtime configurations :-). Some of
`spark.connect.*` are runtime and others are static so we might need to clarify
them tho.
--
This is an automated
HyukjinKwon commented on code in PR #42521:
URL: https://github.com/apache/spark/pull/42521#discussion_r1299493549
##
python/pyspark/sql/tests/connect/streaming/test_parity_listener.py:
##
@@ -19,38 +19,153 @@
import time
from
HyukjinKwon commented on PR #38624:
URL: https://github.com/apache/spark/pull/38624#issuecomment-1685484092
I get that `cogroup` might not be possible tho. But we can just convert
pandas back to arrow batches easily. Is this really required for some scenario?
IIRC this is only useful for
HyukjinKwon commented on PR #38624:
URL: https://github.com/apache/spark/pull/38624#issuecomment-1685483237
qq, can't we workaround by `df.repartitionByExpression().mapInArrow()` for
`groupby` case?
--
This is an automated message from the Apache Git Service.
To respond to the message,
HyukjinKwon commented on code in PR #41782:
URL: https://github.com/apache/spark/pull/41782#discussion_r1299491118
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -487,6 +487,25 @@ object SQLConf {
.intConf
.createWithDefault(1)
HyukjinKwon commented on code in PR #41782:
URL: https://github.com/apache/spark/pull/41782#discussion_r1299490804
##
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##
@@ -487,6 +487,25 @@ object SQLConf {
.intConf
.createWithDefault(1)
zhengruifeng commented on PR #42556:
URL: https://github.com/apache/spark/pull/42556#issuecomment-1685477162
cc @grundprinzip @hvanhovell
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng commented on PR #42563:
URL: https://github.com/apache/spark/pull/42563#issuecomment-1685475524
merged to master and branch-3.5
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng closed pull request #42563: [SPARK-44877][CONNECT][PYTHON] Support
python protobuf functions for Spark Connect
URL: https://github.com/apache/spark/pull/42563
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use
HyukjinKwon commented on code in PR #40085:
URL: https://github.com/apache/spark/pull/40085#discussion_r1299485759
##
python/pyspark/sql/functions.py:
##
@@ -13068,6 +13068,46 @@ def _invoke_higher_order_function(
return Column(cast(JVMView, sc._jvm).Column(expr(*jcols +
HyukjinKwon commented on code in PR #42428:
URL: https://github.com/apache/spark/pull/42428#discussion_r1299471107
##
python/docs/source/_templates/version-switcher.html:
##
@@ -0,0 +1,60 @@
+
Review Comment:
Let's put the license header:
```
```
--
This
HyukjinKwon commented on code in PR #42428:
URL: https://github.com/apache/spark/pull/42428#discussion_r1299470607
##
python/docs/source/_static/versions.json:
##
@@ -0,0 +1,278 @@
+[
Review Comment:
I wonder if we better remove EOL releases ... but no strong opinion WDYT
HyukjinKwon commented on code in PR #42392:
URL: https://github.com/apache/spark/pull/42392#discussion_r1299469771
##
python/pyspark/pandas/tests/test_resample.py:
##
@@ -252,14 +254,32 @@ def test_dataframe_resample(self):
self._test_resample(self.pdf5, self.psdf5,
zhengruifeng commented on PR #42526:
URL: https://github.com/apache/spark/pull/42526#issuecomment-1685452613
merged to master
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
zhengruifeng closed pull request #42526: [SPARK-44842][SPARK-43812][PS] Support
stat functions for pandas 2.0.0 and enabling tests.
URL: https://github.com/apache/spark/pull/42526
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub
zhengruifeng commented on PR #42547:
URL: https://github.com/apache/spark/pull/42547#issuecomment-1685449633
merged to master and branch-3.5
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
zhengruifeng closed pull request #42547: [SPARK-44858][PYTHON][DOCS] Refine
dostring of DataFrame.isEmpty
URL: https://github.com/apache/spark/pull/42547
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
HyukjinKwon closed pull request #42255: [SPARK-40178][SQL][COONECT] Support
coalesce hints with ease for PySpark and R
URL: https://github.com/apache/spark/pull/42255
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
HyukjinKwon commented on PR #42255:
URL: https://github.com/apache/spark/pull/42255#issuecomment-1685447669
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
zhengruifeng commented on PR #42575:
URL: https://github.com/apache/spark/pull/42575#issuecomment-1685447587
cc @gengliangwang @gatorsmile
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
HyukjinKwon commented on code in PR #42513:
URL: https://github.com/apache/spark/pull/42513#discussion_r1299463703
##
python/pyspark/sql/dataframe.py:
##
@@ -3793,6 +3793,8 @@ def union(self, other: "DataFrame") -> "DataFrame":
Example 2: Combining two DataFrames with
HyukjinKwon closed pull request #42569: [SPARK-44879][PYTHON][DOCS] Refine the
docstring of spark.createDataFrame
URL: https://github.com/apache/spark/pull/42569
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL
HyukjinKwon commented on PR #42569:
URL: https://github.com/apache/spark/pull/42569#issuecomment-1685443463
Merged to master.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific
HyukjinKwon closed pull request #42568: [SPARK-44876][PYTHON] Fix
Arrow-optimized Python UDF on Spark Connect
URL: https://github.com/apache/spark/pull/42568
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above
HyukjinKwon commented on PR #42568:
URL: https://github.com/apache/spark/pull/42568#issuecomment-1685440028
Merged to master and branch-3.5
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
github-actions[bot] closed pull request #41113: [SPARK-43400][SQL] Add Primary
Key syntax support
URL: https://github.com/apache/spark/pull/41113
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the
github-actions[bot] closed pull request #40467: [SPARK-42584][CONNECT] Improve
output of `Column.explain`
URL: https://github.com/apache/spark/pull/40467
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to
1 - 100 of 114 matches
Mail list logo