Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/14262
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the issue:
https://github.com/apache/spark/pull/14262
close this and open the same one base on new master branch.
https://github.com/apache/spark/pull/15914
---
If your project is set up for it, you can reply to this email and have your
reply
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/15914
delete temporary folder after insert hive table
## What changes were proposed in this pull request?
Modify the code of InsertIntoHiveTable.scala. To fix
https://issues.apache.org/jira
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/14262
[SPARK-14974][SQL]delete temporary folder after insert hive table
## What changes were proposed in this pull request?
Modify the code of InsertIntoHiveTable.scala. To fix
https
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/8664#issuecomment-142157532
I am so sorry that I worked on trip and diffcult to connect github.
@andrewor14 @srowen
close this PR
I will reopen it when I go back to beijing
---
If your
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/8664
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/4980
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4980#issuecomment-141668675
no problem. close it :) @andrewor14 @liancheng
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/8664#issuecomment-139138023
hi @srowen @andrewor14 after debug, I found before we call
UTF8String.fromBytes(output.toByteArray) , we must call generator.close(),
otherwise UTF8String.fromBytes
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/8664#issuecomment-138793183
test failed by jenkins ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/8664
[SPARK-10506][SQL]move .close() to finally code block
move close to finally to avoid potential resource leak
You can merge this pull request into a Git repository by running:
$ git pull https
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4980#issuecomment-138765865
hi @andrewor14 @liancheng , I think this is a useful patch :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6745#issuecomment-138462580
hi @andrewor14 @liancheng , This patch didnt pass some test. Can close it :)
---
If your project is set up for it, you can reply to this email and have your
reply
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/6951
[SPARK-8561][SQL]drop table under specific database
we run the following code
hivecontext.sql("use dbname")
hivecontext.sql("drop table tablename")
it will
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6868#issuecomment-114422878
Hi @marmbrus @yhuai ,would you please help me check whether this patch can
be merged? :)
---
If your project is set up for it, you can reply to this email and have
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6494#issuecomment-113034110
get error after rebase, close this and create a same one as this :
https://github.com/apache/spark/pull/6868
---
If your project is set up for it, you can reply to
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/6494
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/6868
[SPARK-7943][SQL]support DataFrame created by hiveContext can create table
to specific database by saveAstable
You can merge this pull request into a Git repository by running:
$ git pull
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6695#issuecomment-112373658
hi @marmbrus ,thanks for review this PR. Had modified it according your
comment. Would you please help to check whether it can be merged :)
---
If your project is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6745#issuecomment-110994977
@liancheng thank you for review this PR, had modify the style :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/6745
[Spark-7179][SQL]add tablenamepattern to hive sql of "show tables"
the whole hive sql grammar of "show tables " is like "show tables [in
databasename] [tablenamepat
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6695#issuecomment-110202776
hi @yhuai would you please help me review this pr when you have time? i
think may it was the base of https://github.com/apache/spark/pull/6494 .
thanks:)
---
If
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/6695
[SPARK-8516][SQL]create table to specific database by 'use dbname'
when i test the following code:
hiveContext.sql("""use testdb""")
val df = (1 to 3)
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/6494
[SPARK-7943][SQL][WIP]save table to specified database
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/baishuo/spark SPARK-7943-20150529
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/6486
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6486#issuecomment-106721438
@rxin no problem,close this PR :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/6486#issuecomment-106683828
Hi @rxin , sparkContext should be transient, sorry for forget add it. I
think if define as function it will be computed when it was be referenced but
val need not be
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/6486
change sparkContext in sparkPlan as val
I thinke sparkContext is better defined as val than a function
You can merge this pull request into a Git repository by running:
$ git pull https
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/5781
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/5781#issuecomment-100746741
@marmbrus no problem, close it
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/5660#issuecomment-99758978
hi @srowen , my jira name is also baishuo
https://issues.apache.org/jira/secure/ViewProfile.jspa?name=baishuo
thanks :)
---
If your project is set up for it
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4980#issuecomment-97636824
can this patch be mergedï¼:)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/5781
[SPARK-5084][SQL][WIP]Replaces TestHiveContext.configure() with
HiveContext.overrideHiveConf()
You can merge this pull request into a Git repository by running:
$ git pull https
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/5660#issuecomment-96318903
can this patch be merged? :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/5660
[SPARK-6505][SQL]Remove the reflection call in HiveFunctionWrapper
according @lianchengâs comment in
https://issues.apache.org/jira/browse/SPARK-6505, this patch remove the
reflection call in
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4980#issuecomment-95413614
thank you @srowen
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-82054358
@marmbrus no problem, let me resolve the conflicts :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-82051247
had modify the Title of this PR @marmbrus @liancheng
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-82048390
thank you @liancheng , I had study baishuo/spark#2 ï¼ and I think that
is good :) @marmbrus
---
If your project is set up for it, you can reply to this email
Github user baishuo commented on a diff in the pull request:
https://github.com/apache/spark/pull/3895#discussion_r26364448
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala ---
@@ -57,8 +57,10 @@ class LocalHiveContext(sc: SparkContext) extends
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-78794496
Hi guys, sorry for response this so late. I just work on trip before
china new year. And I rebase the code.
---
If your project is set up for it, you can reply to
Github user baishuo commented on a diff in the pull request:
https://github.com/apache/spark/pull/3895#discussion_r26363950
--- Diff:
sql/hive/v0.13.1/src/main/scala/org/apache/spark/sql/hive/Shim13.scala ---
@@ -297,7 +297,7 @@ private[hive] object HiveShim {
def
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4980#issuecomment-78246998
I think this patch should test by the Reporter of SPARK-6067
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/4980
[SPARK-6067][SQL][WIP]delete the file with the same name before create the
path for HiveFileFormatUtils.getHiveRecordWriter
this is commit for https://issues.apache.org/jira/browse/SPARK-6067
You
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/4001
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4001#issuecomment-70778989
no problem,clolse it :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-70046564
Hi @marmbrus ,can this PR be merged? :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4001#issuecomment-70046499
Hi @marmbrus ,can this PR be merged? :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4001#issuecomment-69884566
Indeed, the code passed all the test when I do test locally, I had
[hotfix] to title just because i want illustrate that this is not the final
solution of [SPARK-4908
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/4001#issuecomment-69883507
Hi @liancheng and @marmbrus I had remove [hotfix] from the title.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/4001
[SPARK-4908][SQL][hotfix]narrow the scope of synchronized for PR 3834
compared with https://github.com/apache/spark/pull/3834, this PR narrow
the scope of synchronized
You can merge this pull
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-68840957
I had modify some code and do test locally
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3895#issuecomment-68675593
some explain:
if we want to use mysql instead of derby to store the metadata for
spark-sql, we add the param such as
"javax.jdo.option.ConnectionURL"
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/3895
[SPARK-5084][SQL]add if not exists after create database-in Shim13.scala
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/baishuo/spark SPARK
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3740#issuecomment-68230642
I learn a lot when review this PRï¼thanks
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3526#issuecomment-65176378
had modified according @marmbrus commentï¼)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/3526#issuecomment-65053699
@rxin no problem. Had modify it :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/3526
[SPARK-4663][sql]add finally to avoid resource leak
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/baishuo/spark master-trycatch
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2876#issuecomment-60341700
hi @vanzin , I had modify 4 pom.xml, change the scope of guava to
"provided" at root pom.xml. And all test of sql project can passed. can this
change
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2876#issuecomment-59910134
Hi @srowen and @vanzin, If we do not do Reimport, there is no problem.
But if we do (Reimport can help idea refresh the jars) and run
SparkSQLCLIDriver. The exception
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2876#issuecomment-59907899
i think the root cause is: the scope of guava in root pom.xml is
"provided", every time when we do reimport (right click the whole project,
click maven->R
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/2876
[SPARK-4034]change the scope of guava to compile
After click maven->reimport for spark project in idea, and begin to start
"sparksqlclidriver" in idea, we will get a exception:
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/2842
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2842#issuecomment-59688527
@JoshRosen @pwendell I know the reason of this problemã In idea, I
should right click the project and click maven->reimport
---
If your project is set up for
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/2842
[SPARK-3999][deploy] resolve the wrong number of arguments for pattern error
AssociationErrorEvent which is provided by
akka-remote_2.10-2.2.3-shaded-protobuf.jar only have 4 arguments
You can
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-56770301
hi @marmbrus ,would you please run the merge script again? :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-56621834
I think I should say thank you to @liancheng and @yhuai. During the
communication with you, I had learned a lot :)
---
If your project is set up for it, you can reply
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-56616245
had remove "s from title @marmbrus
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-56473456
thanks a lot to @liancheng :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not
Github user baishuo commented on a diff in the pull request:
https://github.com/apache/spark/pull/2226#discussion_r17731044
--- Diff:
sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala
---
@@ -522,6 +523,52 @@ class HiveQuerySuite extends
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-55278210
can this PR merged?:)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-55225133
steps to verify this PR by SparkSQLCliDriver:
firstï¼create two table:
run the following sql:
---
If your project is set up for it, you can reply to this email
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-55067304
had update the file according liancheng's comment. and test it locally
---
If your project is set up for it, you can reply to this email and have your
reply appe
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-54966993
after check the consoleFull
there is a error occurs when run the test "full outer join"
[info] - full outer join
05:02:22
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-54949823
I try to explain my design idea(the code is mostly in
InsertIntoHiveTable.scala) :
lets assume there is a table called table1,which has 2 columns:col1,col2,
and two
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-54949644
Hi @marmbrus thanks a lot for your advice. I had modify the code according
to your advice.
I try to seperate dynamic partition support by use the condition &qu
Github user baishuo commented on a diff in the pull request:
https://github.com/apache/spark/pull/2226#discussion_r17290567
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/InsertIntoHiveTable.scala
---
@@ -178,6 +253,40 @@ case class InsertIntoHiveTable
Github user baishuo commented on a diff in the pull request:
https://github.com/apache/spark/pull/2226#discussion_r17287305
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/InsertIntoHiveTable.scala
---
@@ -101,62 +103,135 @@ case class InsertIntoHiveTable
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-54575671
can this PR be tested? :)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-54574495
can this PR be tested? The golden file related HiveCompatibilitySuite with
had already exists in master branch of spark. So do not need to add them.
---
If your
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2226#issuecomment-54259701
Hi @marmbrus and @liancheng, the latest code had pass "dev/lint-scala" and
"sbt/sbt catalyst/test sql/test hive/test" locally.
---
If your proj
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-54244065
no problemï¼ close this PR
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/1919
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-54032088
Hi @
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/2226
[SPARK-3007][SQL]Add "Dynamic Partition" support to Spark Sql hive
a new PR base on new master. changes are the same as
https://github.com/apache/spark/pull/1919
You can merge this pu
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2157#issuecomment-53669323
yes,no problem :) close this issue
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does
Github user baishuo closed the pull request at:
https://github.com/apache/spark/pull/2157
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/2157#issuecomment-53544332
thank you @chenghao-intel . I think I didnt express what I think clearly.
why there is a threadlocal is to ensure there is one and only one NumberFormat
instance in
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/2157
[SPARK-3241][SQL] create NumberFormat instance by threadsafe way
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/baishuo/spark patch
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-53390280
Hi @marmbrus i had update the file relating with test. all test passed on
my machine. Would you please help to verify this patch when you have time:) I
had write out
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52758525
Here I try to explain my design idea(the code is mostly in
InsertIntoHiveTable.scala) :
lets assume there is a table called table1,which has 2 columns:col1,col2,
and
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52734543
I also curious about that.
I down the master branch,and check the folder
sql/hive/src/test/resources/golden
I find that files begin with
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52728496
here I try to express my design idea clearly:
lets assume there is a table called table1,which has 2 columns:col1,col2,
and two partitions: part1, part2.
---
If
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52670830
Hi @marmbrus and @liancheng I had made some modification and do the test
with "sbt/sbt catalyst/test sql/test hive/test" . Please help me to check if
it
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52583642
thanks a lot @yhuai and @liancheng:)
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52141005
hi @marmbrus , when I study the HiveQuerySuite.scala, I found there is a
important table : src, but I didnt find where and how the table created, would
please give more
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1919#issuecomment-52026271
I didnt have add the related test since I dont know how to write it. but
I had test the function by SparkSQLCLIDriver
---
If your project is set up for it, you can
GitHub user baishuo opened a pull request:
https://github.com/apache/spark/pull/1919
[SPARK-3007][SQL]Add "Dynamic Partition" support to Spark Sql hive
the detail please refer the comment of
https://issues.apache.org/jira/browse/SPARK-3007
You can merge this pull requ
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1569#issuecomment-50180682
thank you @marmbrus , I had modify it to "@transient lazy val" and the run
"sbt/sbt catalyst/test sql/test hive/test" at master branch , all test pa
Github user baishuo commented on the pull request:
https://github.com/apache/spark/pull/1569#issuecomment-50101782
modify the title, add [SQL]
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have
1 - 100 of 160 matches
Mail list logo