Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r187907410
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/joins/HashedRelation.scala
---
@@ -568,13 +568,16 @@ private[execution] final class
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r187907559
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/joins/HashedRelation.scala
---
@@ -568,13 +568,16 @@ private[execution] final class
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r187907473
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/joins/HashedRelation.scala
---
@@ -568,13 +568,16 @@ private[execution] final class
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21311
Thanks for your review. @maropu @kiszk @cloud-fan
I submitted a modification including the following:
1. spliting append func into two parts:grow/appendG
2. doubling the size when
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/21311
[SPARK-24257][SQL]LongToUnsafeRowMap calculate the new size may be wrong
## What changes were proposed in this pull request?
LongToUnsafeRowMap
Calculate the new size simply
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21311
@JoshRosen @cloud-fan @gatorsmile
When introducing
[SPARK-10399](https://issues.apache.org/jira/browse/SPARK-10399),UnsafeRow#getUTF8String
check the size at this time.
[UnsafeRow
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21311
@cloud-fan
LongToUnsafeRowMap#append(key: Long, row: UnsafeRow)
when row.getSizeInBytes > newPageSize( oldPage.length * 8L * 2),still use
newPageSize value.
When the new page s
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r189905873
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/execution/joins/HashedRelation.scala
---
@@ -626,6 +618,32 @@ private[execution] final class
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r190146533
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/joins/HashedRelationSuite.scala
---
@@ -254,6 +254,30 @@ class HashedRelationSuite extends
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21311
@cloud-fan Thank you very much for your help.
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional
GitHub user cxzl25 reopened a pull request:
https://github.com/apache/spark/pull/21364
[SPARK-24317][SQL]Float-point numbers are displayed with different
precision in ThriftServer2
## What changes were proposed in this pull request?
When querying float-point numbers
Github user cxzl25 closed the pull request at:
https://github.com/apache/spark/pull/21364
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/21364
[SPARK-20173][SQL]Float-point numbers are displayed with different
precision in ThriftServer2
## What changes were proposed in this pull request?
When querying float-point numbers , the values
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r189997697
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/joins/HashedRelationSuite.scala
---
@@ -254,6 +254,30 @@ class HashedRelationSuite extends
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21311#discussion_r190345942
--- Diff:
sql/core/src/test/scala/org/apache/spark/sql/execution/joins/HashedRelationSuite.scala
---
@@ -254,6 +254,30 @@ class HashedRelationSuite extends
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/18900
**Modify the partition will lose createTime.**
Reading the hive partitions ignores createTime when converting the
CatalogTablePartition, it will also be lost when modifying partitions
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21164
@liutang123 cc @cloud-fan @gatorsmile
I also encountered this problem.
![image](https://user-images.githubusercontent.com/3898450/40981493-9b0b0d46-690d-11e8-8607-c14756610d59.png
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/18900#discussion_r193685282
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/client/HiveClientImpl.scala
---
@@ -1019,6 +1021,8 @@ private[hive] object HiveClientImpl
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21364
ping @gatorsmile @liufengdb
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21656
@maropu
I have added a unit test.
Can you trigger a test for this?
---
-
To unsubscribe, e-mail: reviews-unsubscr
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/20406
[SPARK-23230][SQL]Error by creating a data table when using
hive.default.fileformat=orc
When hive.default.fileformat is other kinds of file types, create textfile
table cause a serda error.
We
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/20593
[SPARK-23230][SQL][BRANCH-2.2]When hive.default.fileformat is other kinds
of file types, create textfile table cause a serde error
When hive.default.fileformat is other kinds of file types, create
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/20406
Thanks for your help , @dongjoon-hyun @gasparms .
I submit a separate PR to 2.2
https://github.com/apache/spark/pull/20593
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/20406
ping @gatorsmile @dongjoon-hyun
```
set hive.default.fileformat=orc;
create table tbl stored as textfile
as
select 1
```
It failed because it used the wrong SERDE
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/20406#discussion_r167482913
--- Diff:
sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveSerDeSuite.scala
---
@@ -100,6 +100,25 @@ class HiveSerDeSuite extends
Github user cxzl25 closed the pull request at:
https://github.com/apache/spark/pull/20593
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21596
https://github.com/apache/spark/pull/20738
Bump jackson from 2.6.7&2.6.7.1 to 2.7.7
Jackson(>=2.7.7) fixes the possibility of missing tail data when the length
of the value is in
Github user cxzl25 commented on a diff in the pull request:
https://github.com/apache/spark/pull/21656#discussion_r200236204
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala ---
@@ -772,6 +772,12 @@ private[spark] class TaskSetManager(
private
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21656
@tgravescs
This is really not difficult.
I'm just not sure if we want to ignore or send down the real time.
Now I have submitted a change, use actual time of successful task
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/21656
@maropu @cloud-fan @squito
Can you trigger a test for this?
This is the exception stack in the log:
```
ERROR Utils: uncaught error in thread task-scheduler-speculation, stopping
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/20739
Another solution:
#20738
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/20738
Another solution:
https://github.com/apache/spark/pull/20739
---
-
To unsubscribe, e-mail: reviews-unsubscr
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/20738
[SPARK-23603][SQL]When the length of the json is in a range,get_json_object
will result in missing tail data
## What changes were proposed in this pull request?
Bump jackson from 2.6.7
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/20739
[SPARK-23603][SQL]When the length of the json is in a range,get_json_object
will result in missing tail data
## What changes were proposed in this pull request?
Replace writeRaw(char[] text
GitHub user cxzl25 opened a pull request:
https://github.com/apache/spark/pull/21656
[SPARK-24677][Core]MedianHeap is empty when speculation is enabled, causing
the SparkContext to stop
## What changes were proposed in this pull request?
When speculation is enabled
Github user cxzl25 closed the pull request at:
https://github.com/apache/spark/pull/20739
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org
Github user cxzl25 closed the pull request at:
https://github.com/apache/spark/pull/20738
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org
Github user cxzl25 commented on the issue:
https://github.com/apache/spark/pull/22887
user set hadoop conf can't overwrite spark-defaults.conf
**SparkHadoopUtil.get.appendS3AndSparkHadoopConfigurations** overwrite the
user-set spark.hadoop with the default configuration
Github user cxzl25 closed the pull request at:
https://github.com/apache/spark/pull/21364
---
-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org
39 matches
Mail list logo