[GitHub] spark pull request #23104: [SPARK-26138][SQL] Cross join requires push Local...

2018-11-27 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/23104#discussion_r236929433
  
--- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala
 ---
@@ -459,6 +459,7 @@ object LimitPushDown extends Rule[LogicalPlan] {
   val newJoin = joinType match {
 case RightOuter => join.copy(right = maybePushLocalLimit(exp, 
right))
 case LeftOuter => join.copy(left = maybePushLocalLimit(exp, left))
+case Cross => join.copy(left = maybePushLocalLimit(exp, left), 
right = maybePushLocalLimit(exp, right))
--- End diff --

There are two tables as follows:
CREATE TABLE `**test1**`(`id` int, `name` int);
CREATE TABLE `**test2**`(`id` int, `name` int);

test1 table data:
2,2
1,1

test2 table data:
2,2
3,3
4,4

Execute sql select * from test1 t1 **left anti join** test2 t2 on 
t1.id=t2.id limit 1; The result:
1,1

But 
   we push the limit 1 on left side, the result is not correct. Result is 
empty.
   we push the limit 1 on right side, the result is not correct. Result is 
empty.

So
left anti join no need to push down limit. Similarly, left semi join is 
the same logic.



---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #23104: [SPARK-26138][SQL] Cross join requires push Local...

2018-11-27 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/23104#discussion_r236589331
  
--- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala
 ---
@@ -459,6 +459,7 @@ object LimitPushDown extends Rule[LogicalPlan] {
   val newJoin = joinType match {
 case RightOuter => join.copy(right = maybePushLocalLimit(exp, 
right))
 case LeftOuter => join.copy(left = maybePushLocalLimit(exp, left))
+case Cross => join.copy(left = maybePushLocalLimit(exp, left), 
right = maybePushLocalLimit(exp, right))
--- End diff --

@cloud-fan 
Please give me some advice. Thank you.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #23104: [SPARK-26138][SQL] Cross join requires push Local...

2018-11-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/23104#discussion_r236143436
  
--- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala
 ---
@@ -459,6 +459,7 @@ object LimitPushDown extends Rule[LogicalPlan] {
   val newJoin = joinType match {
 case RightOuter => join.copy(right = maybePushLocalLimit(exp, 
right))
 case LeftOuter => join.copy(left = maybePushLocalLimit(exp, left))
+case Cross => join.copy(left = maybePushLocalLimit(exp, left), 
right = maybePushLocalLimit(exp, right))
--- End diff --

I think, if when set spark.sql.crossJoin.enabled=true, if Inner join 
without condition, LeftOuter join without condition, RightOuter join without 
condition, FullOuter join without condition , limit should be pushed down on 
both sides, just like cross join limit in this PR.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #23104: [SPARK-26138][SQL] Cross join requires push LocalLimit i...

2018-11-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/23104
  
> The title has a typo.

Sorry, it has been fixed.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #23104: [SPARK-26138][SQL] LimitPushDown cross join requi...

2018-11-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/23104#discussion_r236137253
  
--- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala
 ---
@@ -459,6 +459,7 @@ object LimitPushDown extends Rule[LogicalPlan] {
   val newJoin = joinType match {
 case RightOuter => join.copy(right = maybePushLocalLimit(exp, 
right))
 case LeftOuter => join.copy(left = maybePushLocalLimit(exp, left))
+case Cross => join.copy(left = maybePushLocalLimit(exp, left), 
right = maybePushLocalLimit(exp, right))
--- End diff --

When set  spark.sql.crossJoin.enabled=true, 
inner join without condition, LeftOuter without condition, RightOuter 
without condition, FullOuter without condition, all these are  iterally cross 
join?




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #23104: [SPARK-26138][SQL] LimitPushDown cross join requi...

2018-11-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/23104#discussion_r236115582
  
--- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/optimizer/Optimizer.scala
 ---
@@ -459,6 +459,7 @@ object LimitPushDown extends Rule[LogicalPlan] {
   val newJoin = joinType match {
 case RightOuter => join.copy(right = maybePushLocalLimit(exp, 
right))
 case LeftOuter => join.copy(left = maybePushLocalLimit(exp, left))
+case Cross => join.copy(left = maybePushLocalLimit(exp, left), 
right = maybePushLocalLimit(exp, right))
--- End diff --

A = {(a, 0), (b, 1), (c, 2), (d, 0), **(e, 1), (f, 2)**}
B = {**(e, 1), (f, 2)**}

A inner join B limit 2
If there is limit 2, (a, 0), (b, 1) inner join {(e, 1), (f, 2)}, the result 
is empty. But the real result is not empty.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #23104: [SPARK-26138][SQL] LimitPushDown cross join requires may...

2018-11-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/23104
  
@cloud-fan @dongjoon-hyun @gatorsmile
Help review the code.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #23104: [SPARK-26138][SQL] LimitPushDown cross join requires may...

2018-11-22 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/23104
  
Yes I tested and understood, you are right. @mgaido91


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #23104: [SPARK-26138][SQL] LimitPushDown cross join requires may...

2018-11-22 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/23104
  
Cartesian product refers to the Cartesian product of two sets X and Y in 
mathematics , also known as direct product , expressed as X × Y , the first 
object is a member of X and the second object is One of all possible ordered 
pairs of Y.So cross join mustpush it on the left side.
   For example, A={a,b}, B={0,1,2}, then
   A × B = {(a, 0), (a, 1), (a, 2), (b, 0), (b, 1), (b, 2)}
   B×A={(0, a), (0, b), (1, a), (1, b), (2, a), (2, b)}


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #23104: [SPARK-26138][SQL] LimitPushDown cross join requires may...

2018-11-21 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/23104
  
OK,  I will add some UTs.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #23104: [SPARK-26138][SQL] LimitPushDown cross join requi...

2018-11-21 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/23104

[SPARK-26138][SQL] LimitPushDown cross join requires maybeBushLocalLimit

## What changes were proposed in this pull request?

In LimitPushDown batch, cross join can push down the limit.

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-26138

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/23104.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #23104


commit deb18093b1a1b1c6b7e6ad1fd148448b761297ea
Author: guoxiaolong 
Date:   2018-11-21T13:36:24Z

[SPARK-26138][SQL] LimitPushDown cross join requires maybeBushLocalLimit




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21784: [SPARK-24873][YARN] Turn off spark-shell noisy log outpu...

2018-07-22 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21784
  
We need to listen to @vanzin opinion.
Because the relevant code is what he wrote.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21784: [SPARK-24873][YARN] Turn off spark-shell noisy log outpu...

2018-07-22 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21784
  
But for some spark-submit applications, I want these Application report for 
information. 
What should I do?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21827: [SPARK-24873]Increase switch to shielding frequent inter...

2018-07-22 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21827
  
Please add a switch. And represented by a constant. This configuration is 
added to the running-on-yarn.md document. @hejiefang


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21784: [SPARK-24873][YARN] Turn off spark-shell noisy log outpu...

2018-07-22 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21784
  
what? I think we need to add a switch.
https://github.com/apache/spark/pull/21827


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21036: [SPARK-23958][CORE] HadoopRdd filters empty files to avo...

2018-04-17 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21036
  
Thank you for your comments, I will close this PR, thanks.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #21036: [SPARK-23958][CORE] HadoopRdd filters empty files...

2018-04-17 Thread guoxiaolongzte
Github user guoxiaolongzte closed the pull request at:

https://github.com/apache/spark/pull/21036


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21036: [SPARK-23958][CORE] HadoopRdd filters empty files to avo...

2018-04-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21036
  
1.No need to loop twice to filter to determine if the length is greater 
than 0 
2.This feature is to improve performance, the default switch needs to open


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #21036: [SPARK-23958][CORE] HadoopRdd filters empty files to avo...

2018-04-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/21036
  
Thanks, I will try to add test cases. @felixcheung 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #21036: [SPARK-23958][CORE] HadoopRdd filters empty files...

2018-04-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/21036#discussion_r180655799
  
--- Diff: core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala ---
@@ -55,7 +56,8 @@ private[spark] class HadoopPartition(rddId: Int, override 
val index: Int, s: Inp
 
   /**
* Get any environment variables that should be added to the users 
environment when running pipes
-   * @return a Map with the environment variables and corresponding 
values, it could be empty
+*
+* @return a Map with the environment variables and corresponding 
values, it could be empty
--- End diff --

Thanks.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #21036: [SPARK-23958][CORE] HadoopRdd filters empty files...

2018-04-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/21036#discussion_r180652894
  
--- Diff: core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala ---
@@ -55,7 +56,8 @@ private[spark] class HadoopPartition(rddId: Int, override 
val index: Int, s: Inp
 
   /**
* Get any environment variables that should be added to the users 
environment when running pipes
-   * @return a Map with the environment variables and corresponding 
values, it could be empty
+*
+* @return a Map with the environment variables and corresponding 
values, it could be empty
--- End diff --

what mean?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #21036: [SPARK-23958][CORE] HadoopRdd filters empty files...

2018-04-10 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/21036

[SPARK-23958][CORE] HadoopRdd filters empty files to avoid generating empty 
tasks that affect the performance of the Spark computing performance.

## What changes were proposed in this pull request?

HadoopRdd filter empty files to avoid generating empty tasks that affect 
the performance of the Spark computing performance.

Empty file's length is zero.

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23958

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/21036.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #21036


commit e4ccdf913157b45f11efe8b8900d1f805d853278
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-04-11T02:48:51Z

[SPARK-23958][CORE] HadoopRdd filters empty files to avoid generating empty 
tasks that affect the performance of the Spark computing performance.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20818: [SPARK-23675][WEB-UI]Title add spark logo, use spark log...

2018-03-19 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20818
  
@ajbozarth @srowen 
Help to review the code.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20818: [SPARK-23675][WEB-UI]Title add spark logo, use sp...

2018-03-14 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20818

[SPARK-23675][WEB-UI]Title add spark logo, use spark logo image

## What changes were proposed in this pull request?

Title add spark logo, use spark logo image. reference other big data system 
ui, so i think spark should add it.

spark fix before: 

![spark_fix_before](https://user-images.githubusercontent.com/26266482/37387866-2d5add0e-2799-11e8-9165-250f2b59df3f.png)


spark fix after: 

![spark_fix_after](https://user-images.githubusercontent.com/26266482/37387874-329e1876-2799-11e8-8bc5-c619fc1e680e.png)

 

reference kafka ui: 

![kafka](https://user-images.githubusercontent.com/26266482/37387878-35ca89d0-2799-11e8-834e-1598ae7158e1.png)

 

reference storm ui: 

![storm](https://user-images.githubusercontent.com/26266482/37387880-3854f12c-2799-11e8-8968-b428ba361995.png)

 

reference yarn ui: 

![yarn](https://user-images.githubusercontent.com/26266482/37387881-3a72e130-2799-11e8-97bb-dea85f573e95.png)

 

reference nifi ui: 

![nifi](https://user-images.githubusercontent.com/26266482/37387887-3cecfea0-2799-11e8-9a71-6c454d25840b.png)

 

reference flink ui: 

![flink](https://user-images.githubusercontent.com/26266482/37387888-3f16b1ee-2799-11e8-9d37-8355f0100548.png)

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23675

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20818.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20818


commit 964439be7a592b2a94f93008dabc45a18f97c3c6
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-03-14T07:04:00Z

[SPARK-23675][WEB-UI]Title add spark logo, use spark logo image




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20557: [SPARK-23364][SQL]'desc table' command in spark-sql add ...

2018-02-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20557
  
Well, for now, I don't have a better solution.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20543: [SPARK-23357][CORE] 'SHOW TABLE EXTENDED LIKE pattern=ST...

2018-02-25 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20543
  
Oh, I just think it adds to make it clearer.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20543: [SPARK-23357][CORE] 'SHOW TABLE EXTENDED LIKE pattern=ST...

2018-02-12 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20543
  
@gatorsmile 
Help to review the code.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20570: [spark-23382][WEB-UI]Spark Streaming ui about the conten...

2018-02-12 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20570
  
Okay, I check the other pages again today.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20557: [SPARK-23364][SQL]'desc table' command in spark-sql add ...

2018-02-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20557
  
@srowen @gatorsmile 

![4](https://user-images.githubusercontent.com/26266482/36081707-86d3a7cc-0fdd-11e8-9ee8-1c17efd5d690.png)

Can I overload hive's org.apache.hive.beeline.Commands.java in spark sql 
package, modify the hive's hard-coded statistics rows?
0: jdbc:hive2://localhost:18001> desc cyj.partition_table;
+--++--+--+
| col_name | data_type  | comment  |
+--++--+--+
| # col_name   | data_type  | comment  |
| id   | int| NULL |
| name | string | NULL |
| age  | int| NULL |
| dt   | string | NULL |
| day  | string | NULL |
| hour | int| NULL |
| # Partition Information  ||  |
| # col_name   | data_type  | comment  |
| dt   | string | NULL |
| day  | string | NULL |
| hour | int| NULL |
+--++--+--+
**12 rows** selected (0.092 seconds)

**| # Partition Information  ||  |
| # col_name  | data_type  | comment  |
Also counted as two lines. Still in question.**


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20573: [SPARK-23384][WEB-UI]When it has no incomplete(co...

2018-02-10 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20573

[SPARK-23384][WEB-UI]When it has no incomplete(completed) applications 
found, the last updated time is not formatted and client local time zone is not 
show in history server web ui.

## What changes were proposed in this pull request?

When it has no incomplete(completed) applications found, the last updated 
time is not formatted and client local time zone is not show in history server 
web ui. It is a bug.

fix before:

![1](https://user-images.githubusercontent.com/26266482/36070635-264d7cf0-0f3a-11e8-8426-14135ffedb16.png)

fix after:

![2](https://user-images.githubusercontent.com/26266482/36070651-8ec3800e-0f3a-11e8-991c-6122cc9539fe.png)


## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration 
tests, manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, 
remove this)

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23384

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20573.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20573


commit 0575d5eb402edcca0c67a5fa9001fd5e5183e34e
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-02-11T06:43:20Z

[SPARK-23384][WEB-UI]When it has no incomplete(completed) applications 
found, the last updated time is not formatted and client local time zone is not 
show in history server web ui.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20557: [SPARK-23364][SQL]'desc table' command in spark-s...

2018-02-10 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20557#discussion_r167419765
  
--- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/execution/command/tables.scala ---
@@ -539,15 +539,15 @@ case class DescribeTableCommand(
 throw new AnalysisException(
   s"DESC PARTITION is not allowed on a temporary view: 
${table.identifier}")
   }
-  describeSchema(catalog.lookupRelation(table).schema, result, header 
= false)
+  describeSchema(catalog.lookupRelation(table).schema, result, header 
= true)
--- End diff --

The snapshot is correct fix code effect, the statistics rows does not 
contain the head

![2](https://user-images.githubusercontent.com/26266482/36069344-ba833c56-0f22-11e8-9ab6-26f0ae6285b7.png)


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20570: [spark-23382][WEB-UI]Spark Streaming ui about the...

2018-02-10 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20570

[spark-23382][WEB-UI]Spark Streaming ui about the contents of the for need 
to have hidden and show features, when the table records very much.


## What changes were proposed in this pull request?
Spark Streaming ui about the contents of the for need to have hidden and 
show features, when the table records very much.
please refer to https://github.com/apache/spark/pull/20216

fix after:

![1](https://user-images.githubusercontent.com/26266482/36068644-df029328-0f14-11e8-8350-cfdde9733ffc.png)




## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration 
tests, manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, 
remove this)

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23382

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20570.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20570


commit c6ffe3025af5129a807885f9d757d2ddad641b62
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-02-11T02:13:05Z

[spark-23382][WEB-UI]Spark Streaming ui about the contents of the form need 
to have hidden and show features, when the table records very much.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20557: [SPARK-23364][SQL]'desc table' command in spark-s...

2018-02-10 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20557#discussion_r167416457
  
--- Diff: 
sql/core/src/main/scala/org/apache/spark/sql/execution/command/tables.scala ---
@@ -539,15 +539,15 @@ case class DescribeTableCommand(
 throw new AnalysisException(
   s"DESC PARTITION is not allowed on a temporary view: 
${table.identifier}")
   }
-  describeSchema(catalog.lookupRelation(table).schema, result, header 
= false)
+  describeSchema(catalog.lookupRelation(table).schema, result, header 
= true)
--- End diff --

# Partition Information
# col_name  data_type   comment

Partition information also takes up two rows.
I try to keep the head of the case, let rows number is displayed correctly.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20557: [SPARK-23364][SQL]'desc table' command in spark-s...

2018-02-08 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20557

[SPARK-23364][SQL]'desc table' command in spark-sql add column head display

## What changes were proposed in this pull request?

Use 'desc partition_table'  command in spark-sql client, i think it should 
add column head display.

Add 'col_name' ‘data_type’ 'comment'  column head display.

fix before:

![2](https://user-images.githubusercontent.com/26266482/36013945-283fea8c-0da2-11e8-8265-63d816dabd9b.png)

fix after:

![1](https://user-images.githubusercontent.com/26266482/36013954-3252fd7a-0da2-11e8-8e63-3b586f238072.png)

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23364

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20557.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20557


commit 5699c0dc2810a4500f0ee34414b77b80afd0e9c1
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-02-09T06:00:40Z

[SPARK-23364][SQL]'desc table' command in spark-sql add column head display




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20543: [SPARK-23357][CORE] 'SHOW TABLE EXTENDED LIKE pat...

2018-02-08 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20543

[SPARK-23357][CORE] 'SHOW TABLE EXTENDED LIKE pattern=STRING' add 
‘Partitioned’ display similar to hive, and partition is empty, also need to 
show empty partition field []

## What changes were proposed in this pull request?
'SHOW TABLE EXTENDED LIKE pattern=STRING' add ‘Partitioned’ display 
similar to hive, and partition is empty, also need to show empty partition 
field []

hive:

![3](https://user-images.githubusercontent.com/26266482/35967523-12a15c70-0cfc-11e8-88ce-36b2595c1512.png)


sparkSQL Non-partitioned table fix before:

![1](https://user-images.githubusercontent.com/26266482/35967561-32098ede-0cfc-11e8-8382-57ae4857556b.png)


sparkSQL partitioned table fix before:

![2](https://user-images.githubusercontent.com/26266482/35967572-3e4f1150-0cfc-11e8-9956-5007ccb50761.png)


sparkSQL Non-partitioned table fix after:

![4](https://user-images.githubusercontent.com/26266482/35967586-493376b0-0cfc-11e8-8652-0618912bd63f.png)


sparkSQL partitioned table fix after:

![5](https://user-images.githubusercontent.com/26266482/35967602-52474588-0cfc-11e8-9b34-7532c6abae00.png)


## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23357

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20543.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20543


commit 073542d7199acddfbb122d28ab5110f638c2ec82
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-02-08T10:12:22Z

[SPARK-23357][CORE] 'SHOW TABLE EXTENDED LIKE pattern=STRING' add 
‘Partitioned’ display similar to hive, and partition is empty, also need to 
show empty partition field []




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r165253828
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

I am very sad. I'm working on whether there's a better way.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r165251810
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

If you can add a switch parameter, the default value is false.

If it is true, then it needs to be count (read the file again) so that the 
records can be correctly counted. Of course, it shows that when the parameter 
is opened to true, the streaming performance problem will be affected.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r165247567
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

Asynchronous processing, does not affect the backbone of the Streaming job, 
also can get the number of records.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r165239860
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

The number of rows in a file. Is this solution possible?



---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r164996836
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

I would like another from a thread, try hdfs api to count the number of 
documents, does not affect the main thread spark jobs, i will try. ok? 
@jerryshao


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r164975752
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

I see what you mean. I'll try to make it read once. Can you give me some 
idea?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-30 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20437#discussion_r164973156
  
--- Diff: 
streaming/src/main/scala/org/apache/spark/streaming/dstream/FileInputDStream.scala
 ---
@@ -157,7 +157,7 @@ class FileInputDStream[K, V, F <: NewInputFormat[K, V]](
 val metadata = Map(
   "files" -> newFiles.toList,
   StreamInputInfo.METADATA_KEY_DESCRIPTION -> newFiles.mkString("\n"))
-val inputInfo = StreamInputInfo(id, 0, metadata)
+val inputInfo = StreamInputInfo(id, rdds.map(_.count).sum, metadata)
--- End diff --

Because of this little overhead, that 'Records' is not recorded? This is a 
obvious bug.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream Streami...

2018-01-30 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20437
  
thanks, Thank you for your review.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20437: [SPARK-23270][Streaming][WEB-UI]FileInputDStream ...

2018-01-30 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20437

[SPARK-23270][Streaming][WEB-UI]FileInputDStream Streaming UI 's records 
should not be set to the default value of 0, it should be the total number of 
rows of new files.

## What changes were proposed in this pull request?

FileInputDStream Streaming UI 's records should not be set to the default 
value of 0, it should be the total number of rows of new files.
--in FileInputDStream.scala 
start
val inputInfo = StreamInputInfo(id, 0, metadata) // set to the default 
value of 0
ssc.scheduler.inputInfoTracker.reportInfo(validTime, inputInfo)
case class StreamInputInfo(
inputStreamId: Int, numRecords: Long, metadata: Map[String, Any] = 
Map.empty)
in FileInputDStream.scala 
end--
 
--in DirectKafkaInputDStream.scala 
start
val inputInfo = StreamInputInfo(id, rdd.count, metadata) //set to rdd count 
as numRecords
ssc.scheduler.inputInfoTracker.reportInfo(validTime, inputInfo)
case class StreamInputInfo(
inputStreamId: Int, numRecords: Long, metadata: Map[String, Any] = 
Map.empty)
in DirectKafkaInputDStream.scala 
end--
 
test method:
./bin/spark-submit --class 
org.apache.spark.examples.streaming.HdfsWordCount 
examples/jars/spark-examples_2.11-2.4.0-SNAPSHOT.jar /spark/tmp/

fix after:

![1](https://user-images.githubusercontent.com/26266482/35564207-5c19c946-05f4-11e8-9367-a36dd321a111.png)


## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23270

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20437.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20437


commit 41148c605ddf48c155fc03611bca03af9d4e25a3
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-01-30T11:30:49Z

[SPARK-23270][Streaming][WEB-UI]FileInputDStream Streaming UI 's records 
should not be set to the default value of 0, it should be the total number of 
rows of new files.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20259: [SPARK-23066][WEB-UI] Master Page increase master start-...

2018-01-21 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20259
  
Thank you for review, I will close this list. I'm going to use a script to 
monitor the health of the Master process.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20259: [SPARK-23066][WEB-UI] Master Page increase master...

2018-01-21 Thread guoxiaolongzte
Github user guoxiaolongzte closed the pull request at:

https://github.com/apache/spark/pull/20259


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20287: [SPARK-23121][WEB-UI] When the Spark Streaming ap...

2018-01-21 Thread guoxiaolongzte
Github user guoxiaolongzte closed the pull request at:

https://github.com/apache/spark/pull/20287


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20287: [SPARK-23121][WEB-UI] When the Spark Streaming app is ru...

2018-01-19 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20287
  
@smurakozi @vanzin @srowen 
Thanks, i will close the PR.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-17 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
Please help merge code, thank you.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20287: [SPARK-23121][WEB-UI] When the Spark Streaming app is ru...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20287
  
@smurakozi 
Help review the code, this bug results from your added functionality. 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20287: [SPARK-23121][WEB-UI] When the Spark Streaming app is ru...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20287
  
Well, then you can tell me how specific changes? I do not have a good idea 
right now. The problem is that the page crashes, it should be a fatal bug.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
test this please


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20259: [SPARK-23066][WEB-UI] Master Page increase master...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20259#discussion_r161959792
  
--- Diff: core/src/main/scala/org/apache/spark/deploy/master/Master.scala 
---
@@ -179,6 +181,7 @@ private[deploy] class Master(
 }
 persistenceEngine = persistenceEngine_
 leaderElectionAgent = leaderElectionAgent_
+startupTime = System.currentTimeMillis()
--- End diff --

I understand what you mean.
I do not always query, but developers occasionally to observe.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20287: [SPARK-23121][WEB-UI] When the Spark Streaming ap...

2018-01-16 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20287

[SPARK-23121][WEB-UI] When the Spark Streaming app is running for a period 
of time, the page is incorrectly reported when accessing '/jobs' or 
'/jobs/job?id=13'

## What changes were proposed in this pull request?

When the Spark Streaming app is running for a period of time, the page is 
incorrectly reported when accessing '/ jobs /' or '/ jobs / job /? Id = 13' and 
ui can not be accessed.
 
Test command:
./bin/spark-submit --class 
org.apache.spark.examples.streaming.HdfsWordCount 
./examples/jars/spark-examples_2.11-2.4.0-SNAPSHOT.jar /spark
 
The app is running for a period of time,  ui can not be accessed, please 
see attachment.
fix before:

![1](https://user-images.githubusercontent.com/26266482/35024280-8c06f79e-fb79-11e7-8e5c-b804e06945d2.png)

![2](https://user-images.githubusercontent.com/26266482/35024281-8c353906-fb79-11e7-8f99-4e1bfbac9776.png)

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23121

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20287.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20287


commit 03a84436ef2b6227f8bcfdd0b803c9457c8bd5cd
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-01-17T03:26:19Z

[SPARK-23121][WEB-UI]When the Spark Streaming app is running for a period 
of time, the page is incorrectly reported when accessing '/josb' or 
'/jobs/job?id=13'




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20259: [SPARK-23066][WEB-UI] Master Page increase master...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20259#discussion_r161937136
  
--- Diff: core/src/main/scala/org/apache/spark/deploy/master/Master.scala 
---
@@ -125,6 +125,8 @@ private[deploy] class Master(
   private var restServer: Option[StandaloneRestServer] = None
   private var restServerBoundPort: Option[Int] = None
 
+  var startupTime: Long = 0
--- End diff --

 metricRegistry.register(MetricRegistry.name("startupTime"), new 
Gauge[String] {
override def getValue: String = UIUtils.formatDate(master.startupTime)
  })

if startupTime is private, master.startupTime is unable to call.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20259: [SPARK-23066][WEB-UI] Master Page increase master...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20259#discussion_r161936082
  
--- Diff: core/src/main/scala/org/apache/spark/deploy/master/Master.scala 
---
@@ -179,6 +181,7 @@ private[deploy] class Master(
 }
 persistenceEngine = persistenceEngine_
 leaderElectionAgent = leaderElectionAgent_
+startupTime = System.currentTimeMillis()
--- End diff --

Spark master process zombie, the background has a shell script 
automatically pull the spark master process to ensure high availability, but 
the restart process, there may be some applications such as failure.

If I look at startup time metric today, if the startup time is ten days ago 
or a month ago, I would think the system is relatively stable, there is no 
restart behavior.

If I look at the startup time metric today, if startup time was 1 day ago 
or an hour ago, I would assume that the system is unstable and that a recent 
reboot has occurred, requiring developers to troubleshoot problems and analyze 
them.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-16 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
Yes, it makes the Workers / Apps lists collapsible in the same way as other 
blocks.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-15 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
I agree with your second suggestion, before I did not understand what you 
mean, now I passed the test I understand what you mean.

1.In order for collapsible tables to persist on reload each table much be 
added to the function at the bottom on web.js. When I refresh the page, if it 
is hidden, will still be hidden; if it is displayed, will still be displayed.

2.to ensure user interface consistency.

@ajbozarth @srowen 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20259: [SPARK-23066][WEB-UI] Master Page increase master start-...

2018-01-15 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20259
  
 I set the startup time to a metric. The metric instead of the master page 
display.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20259: [SPARK-23066][WEB-UI] Master Page increase master start-...

2018-01-15 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20259
  
OK, i understand your suggestion.
Can I set the startup time to a metric?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-14 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
@ajbozarth 
The first suggestion, I have already fixed it.

![3](https://user-images.githubusercontent.com/26266482/34931312-27b9e74a-fa09-11e7-89e5-8b7c0f5ad59b.png)


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20259: [SPARK-23066][WEB-UI] Master Page increase master start-...

2018-01-14 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20259
  
1. Concerned about the start-up time is to see if the system is stable.

2. Our system has 50,000 + app running every day, maser will generate a lot 
of app registration, management and other log information, the log volume will 
be great, all must use log4j, org.apache.log4j.DailyRollingFileAppender 
rollback log, if not In this way, the log file will bang the local disk.

3. Production environment lunix server user name and password only super 
managers have, we ordinary developers do not have permission, only through the 
WEB UI to access.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20259: [SPARK-23066][WEB-UI] Master Page increase master start-...

2018-01-14 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20259
  
Sir, I stick to my point for the following reasons:

1.When the spark system is running for some time, the log has been rolled 
back, because we use log4j, simply do not see the beginning of the start time.

2.lunix server not everyone has permission to visit, I do not have 
permission to see, but we are very concerned about the start-up time, we can 
only see the WEB UI.

3. It's already crowded. I can adjust the details, look at the snapshot, 
what do you think?

![2](https://user-images.githubusercontent.com/26266482/34925730-1781348c-f9e6-11e7-9ef0-ad84fff03f04.png)



---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-13 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
1. The first suggestion, I will fix it.
2. The second suggestion, I think it is not necessary. Because spark system 
is small, such as 3 workers, do not need to hide the table from the beginning.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20259: [SPARK-23066][WEB-UI] Master Page increase master...

2018-01-13 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20259

[SPARK-23066][WEB-UI] Master Page increase master start-up time.

## What changes were proposed in this pull request?

When a spark system runs stably for a long time, we do not know how long it 
actually runs and can not get its startup time from the UI.
So, it is necessary to increase the Master start-up time.

## How was this patch tested?

 manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23066

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20259.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20259


commit b985c0f30081c5fdf93b54cccebb6afad462bde1
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-01-13T08:52:46Z

[SPARK-23066][WEB-UI] Master Page increase master start-up time.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
@ajbozarth @srowen 
Fix the code, increase the arrow of the form page, maintain the consistency 
of the function.

after fix:

![4](https://user-images.githubusercontent.com/26266482/34861201-29dcd1e4-f79e-11e7-8015-28c320e4b4bc.png)

![5](https://user-images.githubusercontent.com/26266482/34861202-2a114334-f79e-11e7-8deb-428836770bef.png)

![6](https://user-images.githubusercontent.com/26266482/34861203-2a3f6a3e-f79e-11e7-9c2a-924ea67b12c7.png)

![7](https://user-images.githubusercontent.com/26266482/34861204-2a70296c-f79e-11e7-915a-fc6a64e07108.png)

![8](https://user-images.githubusercontent.com/26266482/34861205-2aa42262-f79e-11e7-8c97-828ec24e0f71.png)




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20194: [SPARK-22999][SQL]'show databases like command' c...

2018-01-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20194#discussion_r161130352
  
--- Diff: 
sql/catalyst/src/main/antlr4/org/apache/spark/sql/catalyst/parser/SqlBase.g4 ---
@@ -141,7 +141,7 @@ statement
 (LIKE? pattern=STRING)?
#showTables
 | SHOW TABLE EXTENDED ((FROM | IN) db=identifier)?
 LIKE pattern=STRING partitionSpec? 
#showTable
-| SHOW DATABASES (LIKE pattern=STRING)?
#showDatabases
+| SHOW DATABASES (LIKE? pattern=STRING)?
#showDatabases
--- End diff --

@gatorsmile  @dongjoon-hyun 
I think we can make LIKE optional.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  

![3](https://user-images.githubusercontent.com/26266482/34856154-87b381b6-f77e-11e7-932e-bb14415dc56a.png)



---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-11 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
No, just hide the table, in fact, the data is already on the page, but we 
can not see.
When we refresh the page, it will re-show all the data.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20194: [SPARK-22999][SQL]'show databases like command' c...

2018-01-10 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20194#discussion_r160869162
  
--- Diff: 
sql/catalyst/src/main/antlr4/org/apache/spark/sql/catalyst/parser/SqlBase.g4 ---
@@ -141,7 +141,7 @@ statement
 (LIKE? pattern=STRING)?
#showTables
 | SHOW TABLE EXTENDED ((FROM | IN) db=identifier)?
 LIKE pattern=STRING partitionSpec? 
#showTable
-| SHOW DATABASES (LIKE pattern=STRING)?
#showDatabases
+| SHOW DATABASES (LIKE? pattern=STRING)?
#showDatabases
--- End diff --

No, I just saw like show tables like can be removed. so I think show 
databases like can also be removed. Just think it is removed, the operation is 
more convenient.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #20216: [SPARK-23024][WEB-UI]Spark ui about the contents of the ...

2018-01-10 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/20216
  
Dear Sir, However, the real spark big data environment, a very large number 
of workers, every day running a very large number of applications, has 
completed a very large number of applications, there are also some failed 
applications. 
When so much data are loaded into the WEB UI, coupled with slow Internet 
access and other factors, if you have a hidden form of the function, not only 
for the overall look and feel of the WEB UI neat, more personal feel more 
convenient than jump access.
If you jump to a table and view the table's records, if you want to see the 
records in other tables, you have to flip it up and down because the jump's 
link is at the top of the page.
Attachment has been uploaded.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20216: [SPARK-23024][WEB-UI]Spark ui about the contents ...

2018-01-10 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20216

[SPARK-23024][WEB-UI]Spark ui about the contents of the form need to have 
hidden and show features, when the table records very much.

## What changes were proposed in this pull request?
Spark ui about the contents of the form need to have hidden and show 
features, when the table records very much. Because sometimes you do not care 
about the record of the table, you just want to see the contents of the next 
table, but you have to scroll the scroll bar for a long time to see the 
contents of the next table.

Currently we have about 500 workers, but I just wanted to see the logs for 
the running applications table. I had to scroll through the scroll bars for a 
long time to see the logs for the running applications table.

In order to ensure functional consistency, I modified the Master Page, 
Worker Page, Job Page, Stage Page, Task Page, Configuration Page, Storage Page, 
Pool Page.

## How was this patch tested?
manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-23024

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20216.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20216


commit 1e4a18616a41dfbb4d6e58134facd5e593846af5
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-01-10T08:16:40Z

[SPARK-23024][WEB-UI]Spark ui about the contents of the form need to have 
hidden and show features, when the table records very much.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20194: [SPARK-22999][SQL]'show databases like command' c...

2018-01-08 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/20194#discussion_r160313038
  
--- Diff: 
sql/catalyst/src/main/antlr4/org/apache/spark/sql/catalyst/parser/SqlBase.g4 ---
@@ -141,7 +141,7 @@ statement
 (LIKE? pattern=STRING)?
#showTables
 | SHOW TABLE EXTENDED ((FROM | IN) db=identifier)?
 LIKE pattern=STRING partitionSpec? 
#showTable
-| SHOW DATABASES (LIKE pattern=STRING)?
#showDatabases
+| SHOW DATABASES (LIKE? pattern=STRING)?
#showDatabases
--- End diff --

Hive did not change, but I think spark can change the next.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #20194: [SPARK-22999][SQL]'show databases like command' c...

2018-01-08 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/20194

[SPARK-22999][SQL]'show databases like command' can remove the like keyword

## What changes were proposed in this pull request?

SHOW DATABASES (LIKE pattern = STRING)? Can be like the back increase?
When using this command, LIKE keyword can be removed. 
You can refer to the SHOW TABLES command, SHOW TABLES 'test *' and SHOW 
TABELS like 'test *' can be used. 
Similarly SHOW DATABASES 'test *' and SHOW DATABASES like 'test *' can be 
used.

## How was this patch tested?
unit tests   manual tests
Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-22999

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/20194.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #20194


commit 26e7c0d22b07144502eb5e05343d6a78824a1f1b
Author: guoxiaolong <guo.xiaolong1@...>
Date:   2018-01-09T02:40:39Z

[SPARK-22999][SQL]'show databases like command' can remove the like keyword




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19841: [SPARK-22642][SQL] the createdTempDir will not be delete...

2017-11-29 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19841
  
+1, avoid unpredictable exceptions that cause the temporary directory or 
file to be deleted.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [DOC]update the API doc and modify the stage API descrip...

2017-11-09 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
@cloud-fan
Help merge the code.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [DOC]update the API doc and modify the stage API descrip...

2017-11-09 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
retest this please


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [DOC]update the API doc and modify the stage API descrip...

2017-11-03 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
retest this please


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [DOC]update the API doc and modify the stage API descrip...

2017-11-03 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
I have updated the title and description.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [CORE]Modify the duration real-time calculation and upda...

2017-11-02 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
Thank you for your review comments, I have to restore the code, not running 
in the code calculation. Now only keep the document changes. Please review 
again.
@srowen @jiangxb1987 @cloud-fan @ajbozarth 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19625: [SPARK-22407][WEB-UI] Add rdd id column on storage page ...

2017-11-02 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19625
  
Please upload the screenshot in PR.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [CORE]Modify the duration real-time calculation and upda...

2017-10-31 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
retest this please


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19532: [CORE]Modify the duration real-time calculation a...

2017-10-30 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/19532#discussion_r147677958
  
--- Diff: core/src/main/scala/org/apache/spark/ui/SparkUI.scala ---
@@ -120,7 +120,7 @@ private[spark] class SparkUI private (
 attemptId = None,
 startTime = new Date(startTime),
 endTime = new Date(-1),
-duration = 0,
+duration = System.currentTimeMillis() - startTime,
--- End diff --

No effect.
In master page, ‘duration ’is calculated using the duration attribute 
of the ApplicationInfo object.

![1](https://user-images.githubusercontent.com/26266482/32168880-2925daf4-bda9-11e7-8e89-925d5d29bf2c.png)


![2](https://user-images.githubusercontent.com/26266482/32168885-2e6c9aca-bda9-11e7-9f77-085c4721ead3.png)

And ApplicationAttemptInfo object does not matter.


 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19520: [SPARK-22298][WEB-UI] url encode APP id before generatin...

2017-10-29 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19520
  
I would like to ask, under what circumstances the application id will 
contain a forward slash?


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19507: [WEB-UI] Add count in fair scheduler pool page

2017-10-29 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19507
  
@srowen 
Help review the code.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [CORE]Modify the duration real-time calculation and upda...

2017-10-29 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
@jiangxb1987  @srowen 
Help review the code.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19507: [WEB-UI] Add count in fair scheduler pool page

2017-10-24 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19507
  
Please refer to  https://github.com/apache/spark/pull/19346 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19532: [CORE]Modify the duration real-time calculation and upda...

2017-10-24 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19532
  
@jiangxb1987 I modified it.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19532: [CORE]stage api modify the description format, ad...

2017-10-19 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/19532#discussion_r145618914
  
--- Diff: core/src/main/scala/org/apache/spark/ui/SparkUI.scala ---
@@ -120,7 +120,7 @@ private[spark] class SparkUI private (
 attemptId = None,
 startTime = new Date(startTime),
 endTime = new Date(-1),
-duration = 0,
+duration = System.currentTimeMillis() - startTime,
--- End diff --


![1](https://user-images.githubusercontent.com/26266482/31758998-5fc3d946-b4e2-11e7-9839-44b1fceb63bf.png)

I would like to ask, running the spark application in master ui,  why 
real-time display how long?
The logic of these two places I think should be the same.



---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19532: [CORE]stage api modify the description format, ad...

2017-10-18 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/19532

[CORE]stage api modify the description format, add version api, modify the 
duration real-time calculation


## What changes were proposed in this pull request?


stage api modify the description format
A list of all stages for a given application.
?status=[active|complete|pending|failed] list only stages 
in the state.
content should be included in  
add version api doc '/api/v1/version'
modify the duration real-time calculation in running appcations

## How was this patch tested?
manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-22311

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/19532.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #19532


commit 8f53eceb9ed3c33388cef09f628dfb7e4f6de70d
Author: guoxiaolong <guo.xiaolo...@zte.com.cn>
Date:   2017-10-19T03:15:13Z

[CORE]stage api modify the description format, add version api, modify the 
duration real-time calculation




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19242: [CORE][DOC]Add event log conf.

2017-10-18 Thread guoxiaolongzte
Github user guoxiaolongzte commented on a diff in the pull request:

https://github.com/apache/spark/pull/19242#discussion_r145346521
  
--- Diff: docs/configuration.md ---
@@ -740,6 +740,20 @@ Apart from these, the following properties are also 
available, and may be useful
   
 
 
+  spark.eventLog.overwrite
+  false
+  
+Whether to overwrite any existing files.
+  
+
+
+  spark.eventLog.buffer.kb
+  100
+  
+Buffer size to use when writing to output streams.Buffer size in KB.
--- End diff --

I have fixed the description and correction unit.
Please check org.apache.spark.internal.config#EVENT_LOG_OUTPUT_BUFFER_SIZE

  private[spark] val EVENT_LOG_OUTPUT_BUFFER_SIZE = 
ConfigBuilder("spark.eventLog.buffer.kb")
.bytesConf(ByteUnit.KiB)
.createWithDefaultString("100k")


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19242: [CORE][DOC]Add event log conf.

2017-10-18 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19242
  
@srowen 
Help to review the code, thanks.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19360: [SPARK-22139][CORE]Remove the variable which is n...

2017-10-18 Thread guoxiaolongzte
Github user guoxiaolongzte closed the pull request at:

https://github.com/apache/spark/pull/19360


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19360: [SPARK-22139][CORE]Remove the variable which is n...

2017-10-18 Thread guoxiaolongzte
GitHub user guoxiaolongzte reopened a pull request:

https://github.com/apache/spark/pull/19360

[SPARK-22139][CORE]Remove the variable which is never used in 
SparkConf.scala

## What changes were proposed in this pull request?

Remove the variable which is never used in SparkConf.scala.
val executorClasspathKey = "spark.executor.extraClassPath"
val driverOptsKey = "spark.driver.extraJavaOptions"
val driverClassPathKey = "spark.driver.extraClassPath"
val sparkExecutorInstances = "spark.executor.instances"
They variables are never used. Because the implementation code for the 
validation rule has been removed in SPARK-17979.

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-22139

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/19360.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #19360


commit f345aa8487a64a0256c6965bc198ba8842cd0a51
Author: guoxiaolong <guo.xiaolo...@zte.com.cn>
Date:   2017-09-27T06:58:37Z

[SPARK-22139] Remove the variable which is never used in SparkConf.scala




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19507: [WEB-UI] Add count in fair scheduler pool page

2017-10-18 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19507
  
@ajbozarth 
Sorry, upload the code before I accidentally withdrew the parenthesis.  I 
rejoined the parenthesis. I have fixed it.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19507: add count in fair scheduler pool page

2017-10-16 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/19507

add count in fair scheduler pool page

## What changes were proposed in this pull request?

Add count in fair scheduler pool page. The purpose is to know the 
statistics clearly.
For specific reasons, please refer to PR of 
https://github.com/apache/spark/pull/18525



## How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration 
tests, manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, 
remove this)

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark 
add_count_in_fair_scheduler_pool_page

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/19507.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #19507


commit 4903200a3bb36ba42e3fb57d4fc160cb637554a3
Author: guoxiaolong <guo.xiaolo...@zte.com.cn>
Date:   2017-10-16T11:18:18Z

add count in fair scheduler pool page




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19399: [SPARK-22175][WEB-UI] Add status column to history page

2017-10-09 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19399
  
Nice, I think it should be merged.


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19360: [SPARK-22139][CORE]Remove the variable which is n...

2017-10-08 Thread guoxiaolongzte
Github user guoxiaolongzte closed the pull request at:

https://github.com/apache/spark/pull/19360


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark issue #19360: [SPARK-22139][CORE]Remove the variable which is never us...

2017-10-08 Thread guoxiaolongzte
Github user guoxiaolongzte commented on the issue:

https://github.com/apache/spark/pull/19360
  
@HyukjinKwon The problem of the PR you follow, I do not care, I will close 
this PR. 


---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



[GitHub] spark pull request #19397: [SPARK-22173] Table CSS style needs to be adjuste...

2017-09-29 Thread guoxiaolongzte
GitHub user guoxiaolongzte opened a pull request:

https://github.com/apache/spark/pull/19397

[SPARK-22173] Table CSS style needs to be adjusted in History Page and in 
Executors Page.

## What changes were proposed in this pull request?

There is a problem with table CSS style.

1. At present, table CSS style is too crowded, and the table width cannot 
adapt itself.

2. Table CSS style is different from job page, stage page, task page, 
master page, worker page, etc. The Spark web UI needs to be consistent.

## How was this patch tested?

manual tests

Please review http://spark.apache.org/contributing.html before opening a 
pull request.


You can merge this pull request into a Git repository by running:

$ git pull https://github.com/guoxiaolongzte/spark SPARK-22173

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/spark/pull/19397.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #19397


commit ff46aa01208b640cabb897f6d5f7bd1fe2dcbccf
Author: guoxiaolong <guo.xiaolo...@zte.com.cn>
Date:   2017-09-30T01:35:36Z

[SPARK-22173] Table CSS style needs to be adjusted in History Page and in 
Executors Page.




---

-
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org



  1   2   3   4   >