[
https://issues.apache.org/jira/browse/HIVE-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14225423#comment-14225423
]
Brock Noland commented on HIVE-8836:
------------------------------------
Similar warning on the other:
{noformat}
2014-11-25 14:39:26,714 INFO ql.Driver (SessionState.java:printInfo(828)) -
Query ID = hiveptest_20141125143939_77f4ba53-fc27-4efe-ad3d-85e23ea54748
2014-11-25 14:39:26,714 INFO ql.Driver (SessionState.java:printInfo(828)) -
Total jobs = 2
2014-11-25 14:39:26,714 INFO ql.Driver (SessionState.java:printInfo(828)) -
Launching Job 1 out of 2
2014-11-25 14:39:26,715 INFO ql.Driver (Driver.java:launchTask(1643)) -
Starting task [Stage-3:MAPRED] in serial mode
2014-11-25 14:39:26,715 INFO exec.Task (SessionState.java:printInfo(828)) - In
order to change the average load for a reducer (in bytes):
2014-11-25 14:39:26,715 INFO exec.Task (SessionState.java:printInfo(828)) -
set hive.exec.reducers.bytes.per.reducer=<number>
2014-11-25 14:39:26,715 INFO exec.Task (SessionState.java:printInfo(828)) - In
order to limit the maximum number of reducers:
2014-11-25 14:39:26,715 INFO exec.Task (SessionState.java:printInfo(828)) -
set hive.exec.reducers.max=<number>
2014-11-25 14:39:26,715 INFO exec.Task (SessionState.java:printInfo(828)) - In
order to set a constant number of reducers:
2014-11-25 14:39:26,715 INFO exec.Task (SessionState.java:printInfo(828)) -
set mapreduce.job.reduces=<number>
2014-11-25 14:39:26,715 DEBUG session.SparkSessionManagerImpl
(SparkSessionManagerImpl.java:getSession(107)) - Existing session
(34e37f91-2cac-4a31-aba7-85b711d8dad3) is reused.
2014-11-25 14:39:26,728 INFO ql.Context (Context.java:getMRScratchDir(266)) -
New scratch dir is
file:/home/hiveptest/50.18.64.184-hiveptest-2/apache-svn-spark-source/itests/qtest-spark/target/tmp/scratchdir/hiveptest/e9216630-f66e-4b31-bc30-58078678a976/hive_2014-11-25_14-39-26_638_5903751068675432492-1
2014-11-25 14:39:26,775 INFO client.SparkClientImpl
(SparkClientImpl.java:onReceive(329)) - Received result for
e288f226-2429-469a-9c53-07fabda12db3
2014-11-25 14:55:54,771 WARN remote.ReliableDeliverySupervisor
(Slf4jLogger.scala:apply$mcV$sp(71)) - Association with remote system
[akka.tcp://[email protected]:38320] has
failed, address is now gated for [5000] ms. Reason is: [Disassociated].
{noformat}
> Enable automatic tests with remote spark client.[Spark Branch]
> --------------------------------------------------------------
>
> Key: HIVE-8836
> URL: https://issues.apache.org/jira/browse/HIVE-8836
> Project: Hive
> Issue Type: Sub-task
> Components: Spark
> Affects Versions: spark-branch
> Reporter: Chengxiang Li
> Assignee: Rui Li
> Labels: Spark-M3
> Fix For: spark-branch
>
> Attachments: HIVE-8836.1-spark.patch, HIVE-8836.2-spark.patch,
> HIVE-8836.3-spark.patch, HIVE-8836.4-spark.patch
>
>
> In real production environment, remote spark client should be used to submit
> spark job for Hive mostly, we should enable automatic test with remote spark
> client to make sure the Hive feature workable with it.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)