GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/14010
[GRAPHX][EXAMPLES] move graphx test data directory and update graphx
document
## What changes were proposed in this pull request?
There are two test data for graphx examples which
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13946
[MINOR][SparkR] update sparkR DataFrame.R comment
## What changes were proposed in this pull request?
update sparkR DataFrame.R comment
SQLContext ==> SparkSess
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/13544
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
cc @liancheng Thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
@andrewor14 The PR is OK now.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13558#discussion_r67496513
--- Diff: python/pyspark/sql/catalog.py ---
@@ -232,6 +232,11 @@ def clearCache(self):
"""Removes all cached tables fro
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
@andrewor14 It looks strange, I test on my own machine and it is all OK. If
it is the test server's problem?
---
If your project is set up for it, you can reply to this email and have
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13381#discussion_r66961848
--- Diff: examples/src/main/python/mllib/isotonic_regression_example.py ---
@@ -23,18 +23,22 @@
from pyspark import SparkContext
# $example on
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13381#discussion_r66715565
--- Diff: docs/ml-classification-regression.md ---
@@ -685,6 +685,76 @@ The implementation matches the result from R's
survival fun
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13381#discussion_r66700697
--- Diff:
examples/src/main/scala/org/apache/spark/examples/ml/IsotonicRegressionExample.scala
---
@@ -0,0 +1,62 @@
+/*
+ * Licensed to the
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13592#discussion_r66700281
--- Diff: docs/sql-programming-guide.md ---
@@ -517,24 +517,26 @@ types such as Sequences or Arrays. This RDD can be
implicitly converted to a Dat
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13592#discussion_r66699400
--- Diff: docs/sql-programming-guide.md ---
@@ -1607,13 +1600,13 @@ a regular multi-line JSON file will most often fail.
{% highlight r
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13544
@liancheng OK, no problem !
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13578
@srowen Hi srowen, I have another similar PR #13558 which past test on my
machine, but the official test fail. It seems to be the test server's problem,
can you help to check it ?
-
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13578
[SPARK-15837][ML][PySpark]Word2vec python add maxsentence parameter
## What changes were proposed in this pull request?
Word2vec python add maxsentence parameter.
## How was
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13558
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13558
[SPARK-15820][pyspark][SQL] update python sql interface refreshTable
## What changes were proposed in this pull request?
Add Catalog.refreshTable API into python interface for Spark
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13544
@rxin
a small problem:
in `HiveContext` there is a method `refreshTable` for refreshing metadata
of Hive table.
now using new SparkSession API with hive support, the method is
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13544
@rxin
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13544
[SPARK-15805][SQL][Documents] update sql programming guide
## What changes were proposed in this pull request?
Update the whole sql programming guide doc file , including:
update
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13538
@srowen Yes, I check each md files, and I think it is done.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13538#discussion_r66059465
--- Diff: docs/streaming-programming-guide.md ---
@@ -2037,7 +2037,7 @@ and configuring them to receive different partitions
of the data stream from
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13538
[MINOR] fix typo in documents
## What changes were proposed in this pull request?
I use spell check tools checks typo in spark documents and fix them.
## How was this patch
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/13525
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13525
@srowen
ok, so the comand don't work correctly
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project doe
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13525
[MINOR]fix typo a -> an
## What changes were proposed in this pull request?
a->an
similar to #13515
Use cmds like `find . -name '*.R' | xargs -i sh -c &quo
Github user WeichenXu123 commented on the issue:
https://github.com/apache/spark/pull/13381
@yanboliang Done.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/12987
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13441
[SPARK-15702][Documentation]Update document programming-guide accumulator
section
## What changes were proposed in this pull request?
Update document programming-guide accumulator
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13412
[SPARK-15670][Java API][Spark
Core]label_accumulator_deprecate_in_java_spark_context
## What changes were proposed in this pull request?
Add deprecate annotation for acumulator V1
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13381#issuecomment-222353670
@holdenk Java & python example added. Thanks!
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13381
[SPARK-15608][ml][doc] add_isotonic_regression_doc
## What changes were proposed in this pull request?
add ml doc for ml isotonic regression
add scala example for ml isotonic
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13275#issuecomment-222069163
@jkbradley
--modules='pyspark-ml' will run a bunch of test in pyspark sub directory
parallel, not single python file.
and my purpose is add a wa
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/13313
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13313
[SPARK-15533][SQL]Deprecate Dataset.explode
## What changes were proposed in this pull request?
Deprecate Dataset.explode
## How was this patch tested?
Existing
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13275#issuecomment-221210606
@davies How do you think about it ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13275#issuecomment-221203315
Jenkins, test this please.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13275
[SPARK-15499][PySpark][Tests] Add python testsuite with remote debug and
single test parameter to help developer debug code easier
## What changes were proposed in this pull request
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/13007
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/13240
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/13242#discussion_r64147805
--- Diff: python/pyspark/ml/clustering.py ---
@@ -933,21 +933,20 @@ def getKeepLastCheckpoint(self):
if __name__ == "__main__":
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13242
[SPARK-15464][ML][MLlib][SQL][Tests] Replace SQLContext and SparkContext
with SparkSession using builder pattern in python test code
## What changes were proposed in this pull request
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13240
[SPARK-15461][tests][pyspark]modify python test script using version
default 2.7
## What changes were proposed in this pull request?
update the default python version used in pytion
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/13224
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13224
[SPARK-15446][build][sql] modify catalyst using longValueExact not
supporting java 7
## What changes were proposed in this pull request?
(Please fill in changes proposed in this fix
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13172#issuecomment-220260517
@srowen
Modified as you expected.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user WeichenXu123 closed the pull request at:
https://github.com/apache/spark/pull/12978
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/12978#discussion_r63708450
--- Diff: sbin/spark-daemon.sh ---
@@ -162,6 +162,15 @@ run_command() {
esac
echo "$newpid" > "$pid"
+
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13172
[SPARK-15203][Deploy]fix bug SPARK-15203
## What changes were proposed in this pull request?
fix bug SPARK-15203
## How was this patch tested?
existing test
Github user WeichenXu123 commented on a diff in the pull request:
https://github.com/apache/spark/pull/12978#discussion_r63700865
--- Diff: sbin/spark-daemon.sh ---
@@ -162,6 +162,15 @@ run_command() {
esac
echo "$newpid" > "$pid"
+
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/12978#issuecomment-220021066
In order to check this potential problem more carefully, We can add the
following test code like this:
`
echo "$newpid" > "$pid&qu
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/12978#issuecomment-220016724
@srowen According to your suggestion, I add a loop to check whether it pass
STAGE-1 and launch java daemon. And I recover the check statement `! $(ps -p
"$n
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13112#issuecomment-219980697
@srowen updated. Seems no problem.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13136
[SPARK-15350][mllib]add unit test function for LogisticRegressionWithLBFGS
in JavaLogisticRegressionSuite
## What changes were proposed in this pull request?
add unit test function
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13112#issuecomment-219291622
@srowen I use Intellj-IDEA to search usage of deprecate
SparkContext.accumulator in the whole spark project, and update the
code.(except those test code for
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13112
[SPARK-15322][mllib]update deprecate accumulator usage into accumulatorV2
in mllib
## What changes were proposed in this pull request?
MLlib code has two position use sc.accumulator
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/12978#issuecomment-218097327
@srowen Er...I am also a little strange while stage 1 may took a long time
but it really happen several times... If there is time I will do a more
detailed test
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13007#issuecomment-218094762
@HyukjinKwon
En..current cvs load code use Hadoop `LineRecordReader`, so not allow a row
split into mulit-lines, so I think the code should disable csv multi
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/12978#issuecomment-218089784
@srowen At my virtual machine, After OS started and start spark daemon, the
stage 1 describe above will took a long time,often exceeding 2s, I think the
java
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/12978#issuecomment-218040874
@srowen
because spark-daemon.sh using exec command to start the java daemon process.
when run script, the spark-daemon.sh process will exists for a little
Github user WeichenXu123 commented on the pull request:
https://github.com/apache/spark/pull/13007#issuecomment-218037385
@HyukjinKwon I run existing test against this patch and all pass. If need I
will add a new test in CSVSuit.
And I think the only reason cause the bug is
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/13007
fix CSV file data-line with newline at first line load error
## What changes were proposed in this pull request?
fix CSV file data-line with newline at first line load error
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/12987
[spark-15212][SQL]CSV file reader when read file with first line schema do
not filter blank in schema column name
## What changes were proposed in this pull request?
When load csv
GitHub user WeichenXu123 opened a pull request:
https://github.com/apache/spark/pull/12978
[SPARK-15203][Deploy]The spark daemon shell script error, daemon process
start successfully but script output fail message.
This bug is because, sbin/spark-daemon.sh script use bin/spark
1101 - 1170 of 1170 matches
Mail list logo