[jira] [Updated] (HIVE-15435) HiveCLI: two jobs are submitted consecutively when tez is used

2016-12-19 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15435:
---
Attachment: pom.xml

> HiveCLI: two jobs are submitted consecutively when tez is used
> --
>
> Key: HIVE-15435
> URL: https://issues.apache.org/jira/browse/HIVE-15435
> Project: Hive
>  Issue Type: Bug
>  Components: CLI, Hive, Tez
>Affects Versions: 2.1.1
> Environment: Hadoop 2.7.3, Hive 2.1.1, Tez 0.8.4
>Reporter: Alexandre Linte
> Attachments: pom.xml, rm_tez_query.png
>
>
> Since Hive 2.1.1 (Hive 2.1.0 was OK), the HiveCLI has a strange behavior when 
> submitting a job with Tez as the exection engine.
> Doing a query like: 
> {noformat}
> hive (shfs3453)> set hive.execution.engine=tez;
> hive (shfs3453)> set tez.queue.name=HQ_OLPS;
> hive (shfs3453)> SELECT COUNT(year) FROM chicagocrimes GROUP BY year;
> {noformat}
> The query will submit two jobs: one submitted in the queue defined, one 
> submitted in the queue default. Both finish successfully. But only the second 
> job processes the data.
> You can find attached a screenshot.
> This behavior is not visible when using beeline.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15435) HiveCLI: two jobs are submitted consecutively when tez is used

2016-12-19 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15435?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15763550#comment-15763550
 ] 

Alexandre Linte commented on HIVE-15435:


You can find attached my pom.xml. I compiled with "mvn clean package -Pdist 
-DskipTests -X".
Maven 3.3.9 and Java 1.7 were used for the compilation. I hope this can help to 
reproduce this use case.

> HiveCLI: two jobs are submitted consecutively when tez is used
> --
>
> Key: HIVE-15435
> URL: https://issues.apache.org/jira/browse/HIVE-15435
> Project: Hive
>  Issue Type: Bug
>  Components: CLI, Hive, Tez
>Affects Versions: 2.1.1
> Environment: Hadoop 2.7.3, Hive 2.1.1, Tez 0.8.4
>Reporter: Alexandre Linte
> Attachments: rm_tez_query.png
>
>
> Since Hive 2.1.1 (Hive 2.1.0 was OK), the HiveCLI has a strange behavior when 
> submitting a job with Tez as the exection engine.
> Doing a query like: 
> {noformat}
> hive (shfs3453)> set hive.execution.engine=tez;
> hive (shfs3453)> set tez.queue.name=HQ_OLPS;
> hive (shfs3453)> SELECT COUNT(year) FROM chicagocrimes GROUP BY year;
> {noformat}
> The query will submit two jobs: one submitted in the queue defined, one 
> submitted in the queue default. Both finish successfully. But only the second 
> job processes the data.
> You can find attached a screenshot.
> This behavior is not visible when using beeline.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15435) HiveCLI: two jobs are submitted consecutively when tez is used

2016-12-15 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15435:
---
Attachment: rm_tez_query.png

> HiveCLI: two jobs are submitted consecutively when tez is used
> --
>
> Key: HIVE-15435
> URL: https://issues.apache.org/jira/browse/HIVE-15435
> Project: Hive
>  Issue Type: Bug
>  Components: CLI, Hive, Tez
>Affects Versions: 2.1.1
> Environment: Hadoop 2.7.3, Hive 2.1.1, Tez 0.8.4
>Reporter: Alexandre Linte
> Attachments: rm_tez_query.png
>
>
> Since Hive 2.1.1 (Hive 2.1.0 was OK), the HiveCLI has a strange behavior when 
> submitting a job with Tez as the exection engine.
> Doing a query like: 
> {noformat}
> hive (shfs3453)> set hive.execution.engine=tez;
> hive (shfs3453)> set tez.queue.name=HQ_OLPS;
> hive (shfs3453)> SELECT COUNT(year) FROM chicagocrimes GROUP BY year;
> {noformat}
> The query will submit two jobs: one submitted in the queue defined, one 
> submitted in the queue default. Both finish successfully. But only the second 
> job processes the data.
> You can find attached a screenshot.
> This behavior is not visible when using beeline.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15415) Random "java.util.ConcurrentModificationException"

2016-12-12 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15741905#comment-15741905
 ] 

Alexandre Linte commented on HIVE-15415:


Hi [~rajesh.balamohan], 
the use case seems to be very similar. In all cases, I can confirm that the 
error happens only when doing an INSERT OVERWRITE query.

> Random "java.util.ConcurrentModificationException"
> --
>
> Key: HIVE-15415
> URL: https://issues.apache.org/jira/browse/HIVE-15415
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.3, Hive 2.1.0
>Reporter: Alexandre Linte
>
> I'm regularly facing Hive job failures through Oozie or through the beeline 
> CLI. The jobs exit with an error "FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.MoveTask. 
> java.util.ConcurrentModificationException (state=08S01,code=1)" but not 100% 
> of the time. 
> it's also important to underline that only one user is working on the table 
> when the jobs are running.
> - stderr
> {noformat}
> Connecting to jdbc:hive2://hiveserver2.bigdata.fr:1/default
> Connected to: Apache Hive (version 2.1.0)
> Driver: Hive JDBC (version 2.1.0)
> Transaction isolation: TRANSACTION_REPEATABLE_READ
> No rows affected (1.475 seconds)
> No rows affected (0.004 seconds)
> No rows affected (0.004 seconds)
> No rows affected (58.977 seconds)
> No rows affected (5.524 seconds)
> No rows affected (5.235 seconds)
> Error: Error while processing statement: FAILED: Execution Error, return code 
> 1 from org.apache.hadoop.hive.ql.exec.MoveTask. 
> java.util.ConcurrentModificationException (state=08S01,code=1)
> Closing: 0: jdbc:hive2://hiveserver2.bigdata.fr:1/default
> Intercepting System.exit(2)
> {noformat}
> - stdout
> {noformat}
> Beeline command arguments :
>  -u
>  jdbc:hive2://hiveserver2.bigdata.fr:1/default
>  -n
>  my_user
>  -p
>  DUMMY
>  -d
>  org.apache.hive.jdbc.HiveDriver
>  -f
>  full_job
>  -a
>  delegationToken
>  --hiveconf
>  mapreduce.job.tags=oozie-75b060aacd7ec48c4ed637855e413280
> Fetching child yarn jobs
> tag id : oozie-75b060aacd7ec48c4ed637855e413280
> Child yarn jobs are found -
> =
> >>> Invoking Beeline command line now >>>
> 0: jdbc:hive2://hiveserver2.bigdata.fr> use my_db;
> 0: jdbc:hive2://hiveserver2.bigdata.fr> set hive.execution.engine=tez;
> 0: jdbc:hive2://hiveserver2.bigdata.fr> set tez.queue.name=tez_queue;
> 0: jdbc:hive2://hiveserver2.bigdata.fr>
> 0: jdbc:hive2://hiveserver2.bigdata.fr> insert overwrite table main_table 
> ^M_fd_livcfm
> . . . . . . . . . . . . . . . . . . . . . . .> select
> . . . . . . . . . . . . . . . . . . . . . . .> col.co_cd as co_cd,
> . . . . . . . . . . . . . . . . . . . . . . .> col.line_co_cd as line_co_cd,
> . . . . . . . . . . . . . . . . . . . . . . .> 
> unix_timestamp(min(tt.statut_dt)) ^M as statut_dt
> . . . . . . . . . . . . . . . . . . . . . . .> from 
> dlk_scf_rn_customer_order_li ^Mne col
> . . . . . . . . . . . . . . . . . . . . . . .> join 
> dlk_scf_rn_shipment_handling ^M_utility shu
> . . . . . . . . . . . . . . . . . . . . . . .> on shu.co_cd =col.co_cd
> . . . . . . . . . . . . . . . . . . . . . . .> and shu.line_co_cd = 
> col.line_co_ ^Mcd
> . . . . . . . . . . . . . . . . . . . . . . .> join ( select 
> scaler_internal_ref ^M, statut_dt,recep_number,state,reason
> . . . . . . . . . . . . . . . . . . . . . . .> from 
> dlk_scf_rn_transport_trackin ^Mg where state='LIV' and reason='CFM' ) tt
> . . . . . . . . . . . . . . . . . . . . . . .> on 
> concat('CAL',shu.c_waybill_no) ^M =tt.scaler_internal_ref group by 
> col.co_cd,col.line_co_cd;
> Heart beat
> Heart beat
> 0: jdbc:hive2://hiveserver2.bigdata.fr>
> 0: jdbc:hive2://hiveserver2.bigdata.fr> insert overwrite table main_table 
> ^M_fd_cae
> . . . . . . . . . . . . . . . . . . . . . . .> select
> . . . . . . . . . . . . . . . . . . . . . . .> po_cd as cae, line_po_cd as 
> lcae, ^M origin_co_cd, origin_line_co_cd
> . . . . . . . . . . . . . . . . . . . . . . .> from 
> dlk_scf_rn_purchase_order_li ^Mne
> . . . . . . . . . . . . . . . . . . . . . . .> where instr(po_cd,"7")=1;
> 0: jdbc:hive2://hiveserver2.bigdata.fr>
> 0: jdbc:hive2://hiveserver2.bigdata.fr> insert overwrite table main_table 
> ^M_fd_cai
> . . . . . . . . . . . . . . . . . . . . . . .> select
> . . . . . . . . . . . . . . . . . . . . . . .> po_cd as cai, line_po_cd as 
> lcai, ^M origin_co_cd, origin_line_co_cd
> . . . . . . . . . . . . . . . . . . . . . . .> from 
> dlk_scf_rn_purchase_order_li ^Mne
> . . . . . . . . . . 

[jira] [Commented] (HIVE-15128) "hive.exec.stagingdir" produces an "AccessControlException" when working with temp tables

2016-12-12 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15128?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15741568#comment-15741568
 ] 

Alexandre Linte commented on HIVE-15128:


Nothing new on this ticket?

> "hive.exec.stagingdir" produces an "AccessControlException" when working with 
> temp tables
> -
>
> Key: HIVE-15128
> URL: https://issues.apache.org/jira/browse/HIVE-15128
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline, CLI, Hive
>Affects Versions: 2.1.0
> Environment: Hive 2.1.0, Hadoop 2.7.2
>Reporter: Alexandre Linte
>
> The property "hive.exec.stagingdir" is producing an AccessControlException 
> error, only when working with a temporary table and with an INSERT INTO TABLE 
> statement.
> The configuration used in the hive-site.xml is:
> {noformat}
>   
>   hive.exec.stagingdir
>   /user/${user.name}/.hiveStaging/.staging
>   In Hive >= 0.14, set to 
> ${hive.exec.scratchdir}/${user.name}/.staging
>   
> {noformat}
> Trying to insert a value into a temporary table will produce the following 
> output:
> {noformat}
> hive (shfs3453)> INSERT INTO TABLE test_table_tmp VALUES (11101, 'John', 
> 'Oakland');
> Query ID = shfs3453_20161103180522_5fceae9b-6fb5-4cdd-a0d9-37cc93b27a9f
> Total jobs = 1
> Launching Job 1 out of 1
> Status: Running (Executing on YARN cluster with App id 
> application_1478096256452_0018)
> --
> VERTICES  MODESTATUS  TOTAL  COMPLETED  RUNNING  PENDING  
> FAILED  KILLED
> --
> Map 1 .. container SUCCEEDED  1  100  
>  0   0
> --
> VERTICES: 01/01  [==>>] 100%  ELAPSED TIME: 16.09 s
> --
> Loading data to table shfs3453.test_table_tmp
> Failed with exception org.apache.hadoop.security.AccessControlException: User 
> does not belong to hadoop
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.MoveTask. 
> org.apache.hadoop.security.AccessControlException: User does not belong to 
> hadoop
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.h

[jira] [Resolved] (HIVE-15050) Webhcat fails to start with Syslog appender

2016-12-12 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15050?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte resolved HIVE-15050.

Resolution: Fixed

> Webhcat fails to start with Syslog appender
> ---
>
> Key: HIVE-15050
> URL: https://issues.apache.org/jira/browse/HIVE-15050
> Project: Hive
>  Issue Type: Bug
>  Components: WebHCat
>Affects Versions: 2.1.0
> Environment: Hive 2.1.0, Hadoop 2.7.2
>Reporter: Alexandre Linte
>
> WebHCat fails to start if SyslogAppender is set instead of a more 
> conventional FileAppender or DailyRollingFileAppender. However, Hive itself 
> can work with a syslog appender. 
> The error displayed when the webhcat service is killed is:
> {noformat}
> 2016-10-25 09:45:29,538 main ERROR Unable to write to stream 
> UDP:127.0.0.1:514 for appender SYSLOG
> 2016-10-25 09:45:29,540 main ERROR An exception occurred processing Appender 
> SYSLOG org.apache.logging.log4j.core.appender.AppenderLoggingException: Error 
> flushing stream UDP:127.0.0.1:514
> at 
> org.apache.logging.log4j.core.appender.OutputStreamManager.flush(OutputStreamManager.java:159)
> at 
> org.apache.logging.log4j.core.appender.AbstractOutputStreamAppender.append(AbstractOutputStreamAppender.java:107)
> at 
> org.apache.logging.log4j.core.config.AppenderControl.tryCallAppender(AppenderControl.java:152)
> at 
> org.apache.logging.log4j.core.config.AppenderControl.callAppender0(AppenderControl.java:125)
> at 
> org.apache.logging.log4j.core.config.AppenderControl.callAppenderPreventRecursion(AppenderControl.java:116)
> at 
> org.apache.logging.log4j.core.config.AppenderControl.callAppender(AppenderControl.java:84)
> at 
> org.apache.logging.log4j.core.config.LoggerConfig.callAppenders(LoggerConfig.java:390)
> at 
> org.apache.logging.log4j.core.config.LoggerConfig.processLogEvent(LoggerConfig.java:378)
> at 
> org.apache.logging.log4j.core.config.LoggerConfig.log(LoggerConfig.java:362)
> at 
> org.apache.logging.log4j.core.config.LoggerConfig.log(LoggerConfig.java:352)
> at 
> org.apache.logging.log4j.core.config.AwaitCompletionReliabilityStrategy.log(AwaitCompletionReliabilityStrategy.java:63)
> at org.apache.logging.log4j.core.Logger.logMessage(Logger.java:143)
> at 
> org.apache.logging.log4j.spi.AbstractLogger.logMessage(AbstractLogger.java:1016)
> at 
> org.apache.logging.log4j.spi.AbstractLogger.logIfEnabled(AbstractLogger.java:964)
> at org.apache.logging.slf4j.Log4jLogger.info(Log4jLogger.java:178)
> at 
> org.apache.hive.hcatalog.templeton.AppConfig.init(AppConfig.java:197)
> at 
> org.apache.hive.hcatalog.templeton.AppConfig.(AppConfig.java:176)
> at org.apache.hive.hcatalog.templeton.Main.loadConfig(Main.java:101)
> at org.apache.hive.hcatalog.templeton.Main.init(Main.java:85)
> at org.apache.hive.hcatalog.templeton.Main.(Main.java:80)
> at org.apache.hive.hcatalog.templeton.Main.main(Main.java:305)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
> Caused by: java.io.IOException: Message too long
> at java.net.PlainDatagramSocketImpl.send(Native Method)
> at java.net.DatagramSocket.send(DatagramSocket.java:698)
> at 
> org.apache.logging.log4j.core.net.DatagramOutputStream.flush(DatagramOutputStream.java:103)
> at 
> org.apache.logging.log4j.core.appender.OutputStreamManager.flush(OutputStreamManager.java:156)
> ... 26 more
> {noformat}
> The syslog appender configuration used here is:
> {noformat}
> status = INFO
> name = WebhcatLog4j2
> packages = org.apache.hadoop.hive.ql.log
> # list of properties
> property.webhcat.log.level = INFO
> property.webhcat.root.logger = SYSLOG
> #property.webhcat.log.dir = .
> #property.webhcat.log.file = webhcat.log
> # console appender
> appender.console.type = Console
> appender.console.name = console
> appender.console.target = SYSTEM_ERR
> appender.console.layout.type = PatternLayout
> appender.console.layout.pattern = %d{yy/MM/dd HH:mm:ss} [%t]: %p %c{2}: %m%n
> # syslog appender
> appender.SYSLOG.type = Syslog
> appender.SYSLOG.name = SYSLOG
> appender.SYSLOG.host = 127.0.0.1
> appender.SYSLOG.port = 514
> appender.SYSLOG.protocol = UDP
> appender.SYSLOG.facility = LOCAL1
> # list of all loggers
> loggers = Jersey, WebComponent, Hadoop, HadoopConf, Zookeeper, Je

[jira] [Commented] (HIVE-14631) HiveServer2 regularly fails to connect to metastore

2016-11-04 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15636278#comment-15636278
 ] 

Alexandre Linte commented on HIVE-14631:


Am I the only one with this issue? Nobody to investigate?

> HiveServer2 regularly fails to connect to metastore
> ---
>
> Key: HIVE-14631
> URL: https://issues.apache.org/jira/browse/HIVE-14631
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 1.2.1, 2.0.0, 2.1.0
> Environment: Hive 2.1.0, Hue 3.10.0, Hadoop 2.7.2, Tez 0.8.3
>Reporter: Alexandre Linte
>
> I have a cluster secured with Kerberos and Hive is configured to work with 
> Tez by default. Everything works well through hive-cli and beeline; however, 
> I'm facing a strange behavior through Hue.
> I can have a lot of client connections (these can reach 600) and after a day, 
> the client connections fail. But this is not the case for all clients 
> connection attempts.
> When it fails, I have the following logs on the HiveServer2:
> {noformat}
> Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
> command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
>  INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
> Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
> hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
> Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
> Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
> Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
> parallel
> Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:05 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:05 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:06 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:06 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:06 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:08 hiveserver2.bigdata.fr FAILED: Execution Error, return code 
> -1 from org.apache.hadoop.hive.ql.exec.tez.TezTask
> Aug  3 09:28:08 hiveserver2.bigdata.fr Completed executing 
> command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112);
>  Time taken: 4.002 seconds
> {noformat}
> At the same time I have the following logs on the Metastore are:
> {noformat}
> Aug  3 09:28:03 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:03 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:04 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:04 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:04 metastore01.bigdata.fr Error occurred during processing of 
> message.
> Aug  3 09:28:05 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:05 metastore01.bigdata.fr Error occurred during processing of 
> message.
> Aug  3 09:28:06 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:06 metastore01.bigdata.fr Error occurred during processing of 
> message.
> {noformat}
> To solve the connections issue, I have to restart the HiveServer2.
> Note: I also created a JIRA for Hue: 
> https://issues.cloudera.org/browse/HUE-4748



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15017) Random job failures with MapReduce and Tez

2016-11-04 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15636276#comment-15636276
 ] 

Alexandre Linte commented on HIVE-15017:


Nothing new here?

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: debug_yarn_container_mr_job_datanode03.log, 
> debug_yarn_container_mr_job_datanode05.log, hive-site.xml, hive_cli_mr.txt, 
> hive_cli_tez.txt, nodemanager_logs_mr_job.txt, 
> yarn_container_tez_job_datanode05.txt, yarn_container_tez_job_datanode06.txt, 
> yarn_syslog_mr_job.txt, yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15128) "hive.exec.stagingdir" produces an "AccessControlException" when working with temp tables

2016-11-04 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15128:
---
Description: 
The property "hive.exec.stagingdir" is producing an AccessControlException 
error, only when working with a temporary table and with an INSERT INTO TABLE 
statement.
The configuration used in the hive-site.xml is:
{noformat}
  
  hive.exec.stagingdir
  /user/${user.name}/.hiveStaging/.staging
  In Hive >= 0.14, set to 
${hive.exec.scratchdir}/${user.name}/.staging
  
{noformat}

Trying to insert a value into a temporary table will produce the following 
output:
{noformat}
hive (shfs3453)> INSERT INTO TABLE test_table_tmp VALUES (11101, 'John', 
'Oakland');
Query ID = shfs3453_20161103180522_5fceae9b-6fb5-4cdd-a0d9-37cc93b27a9f
Total jobs = 1
Launching Job 1 out of 1


Status: Running (Executing on YARN cluster with App id 
application_1478096256452_0018)

--
VERTICES  MODESTATUS  TOTAL  COMPLETED  RUNNING  PENDING  
FAILED  KILLED
--
Map 1 .. container SUCCEEDED  1  100
   0   0
--
VERTICES: 01/01  [==>>] 100%  ELAPSED TIME: 16.09 s
--
Loading data to table shfs3453.test_table_tmp
Failed with exception org.apache.hadoop.security.AccessControlException: User 
does not belong to hadoop
at 
org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)

FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.MoveTask. 
org.apache.hadoop.security.AccessControlException: User does not belong to 
hadoop
at 
org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
{noformat}

But if you do an INSERT OVERWRITE TABLE, there will be a warning only. No 
AccessControlException error here and erverything is properly inserted.

{noformat}
hive (shfs3453)> INSERT OVERWRITE TABLE test_table_tmp VALUES (11101, 'John', 
'Oakland');
Query ID = shfs3453_20161103180603_d8d5f689-cc55-44d0-b160-bba0df8d186a
Total jobs = 1
Launching Job 1 out of 1


Status: Running (Executing on YARN cluster with App id 
application_1478096256452_0018)

--
VERTICES  MOD

[jira] [Updated] (HIVE-15128) "hive.exec.stagingdir" produces an "AccessControlException" when working with temp tables

2016-11-04 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15128:
---
Environment: Hive 2.1.0, Hadoop 2.7.2

> "hive.exec.stagingdir" produces an "AccessControlException" when working with 
> temp tables
> -
>
> Key: HIVE-15128
> URL: https://issues.apache.org/jira/browse/HIVE-15128
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline, CLI, Hive
>Affects Versions: 2.1.0
> Environment: Hive 2.1.0, Hadoop 2.7.2
>Reporter: Alexandre Linte
>
> The property "hive.exec.stagingdir" is producing an AccessControlException 
> error, only when working with a temporary table and with an INSERT INTO TABLE 
> statement.
> The configuration used in the hive-site.xml is:
> {noformat}
>   
>   hive.exec.stagingdir
>   /user/${user.name}/.hiveStaging/.staging
>   In Hive >= 0.14, set to 
> ${hive.exec.scratchdir}/${user.name}/.staging
>   
> {noformat}
> Trying to insert a value into a temporary table will produce the following 
> output:
> {noformat}
> hive (shfs3453)> INSERT INTO TABLE test_table_tmp VALUES (11101, 'John', 
> 'Oakland');
> Query ID = shfs3453_20161103180522_5fceae9b-6fb5-4cdd-a0d9-37cc93b27a9f
> Total jobs = 1
> Launching Job 1 out of 1
> Status: Running (Executing on YARN cluster with App id 
> application_1478096256452_0018)
> --
> VERTICES  MODESTATUS  TOTAL  COMPLETED  RUNNING  PENDING  
> FAILED  KILLED
> --
> Map 1 .. container SUCCEEDED  1  100  
>  0   0
> --
> VERTICES: 01/01  [==>>] 100%  ELAPSED TIME: 16.09 s
> --
> Loading data to table shfs3453.test_table_tmp
> Failed with exception org.apache.hadoop.security.AccessControlException: User 
> does not belong to hadoop
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.MoveTask. 
> org.apache.hadoop.security.AccessControlException: User does not belong to 
> hadoop
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(Use

[jira] [Updated] (HIVE-15128) "hive.exec.stagingdir" produces an "AccessControlException" when working with temp tables

2016-11-04 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15128:
---
Environment: (was: The property "hive.exec.stagingdir" is producing an 
AccessControlException error, only when working with a temporary table and with 
an INSERT INTO TABLE statement.
The configuration used in the hive-site.xml is:
{noformat}
  
  hive.exec.stagingdir
  /user/${user.name}/.hiveStaging/.staging
  In Hive >= 0.14, set to 
${hive.exec.scratchdir}/${user.name}/.staging
  
{noformat}

Trying to insert a value into a temporary table will produce the following 
output:
{noformat}
hive (shfs3453)> INSERT INTO TABLE test_table_tmp VALUES (11101, 'John', 
'Oakland');
Query ID = shfs3453_20161103180522_5fceae9b-6fb5-4cdd-a0d9-37cc93b27a9f
Total jobs = 1
Launching Job 1 out of 1


Status: Running (Executing on YARN cluster with App id 
application_1478096256452_0018)

--
VERTICES  MODESTATUS  TOTAL  COMPLETED  RUNNING  PENDING  
FAILED  KILLED
--
Map 1 .. container SUCCEEDED  1  100
   0   0
--
VERTICES: 01/01  [==>>] 100%  ELAPSED TIME: 16.09 s
--
Loading data to table shfs3453.test_table_tmp
Failed with exception org.apache.hadoop.security.AccessControlException: User 
does not belong to hadoop
at 
org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)

FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.MoveTask. 
org.apache.hadoop.security.AccessControlException: User does not belong to 
hadoop
at 
org.apache.hadoop.hdfs.server.namenode.FSDirAttrOp.setOwner(FSDirAttrOp.java:86)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setOwner(FSNamesystem.java:1676)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.setOwner(NameNodeRpcServer.java:702)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.setOwner(ClientNamenodeProtocolServerSideTranslatorPB.java:464)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
{noformat}

But if you do an INSERT OVERWRITE TABLE, there will be a warning only. No 
AccessControlException error here and erverything is properly inserted.

{noformat}
hive (shfs3453)> INSERT OVERWRITE TABLE test_table_tmp VALUES (11101, 'John', 
'Oakland');
Query ID = shfs3453_20161103180603_d8d5f689-cc55-44d0-b160-bba0df8d186a
Total jobs = 1
Launching Job 1 out of 1


Status: Running (Executing on YARN cluster with App id 
application_1478096256452_0018)

--
VERTICES

[jira] [Commented] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-11-02 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15628158#comment-15628158
 ] 

Alexandre Linte commented on HIVE-13830:


I'm still using Spark 1.6.1, Hive 2.1.0 and Hadoop 2.7.2, the error remains 
valid.

> Hive on spark driver crash with Spark 1.6.1
> ---
>
> Key: HIVE-13830
> URL: https://issues.apache.org/jira/browse/HIVE-13830
> Project: Hive
>  Issue Type: Bug
>  Components: Spark, spark-branch
>Affects Versions: 2.0.0, 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> With Hive 1.2.1 I was able to use Hive on  successfully with the use of the 
> -assembly "-assembly-1.4.1-hadoop2.7.1.jar". 
> Today with Hive 2.0.0, I'm unable to use Hive on  whether it be with the 
> -assembly "-assembly-1.4.1-hadoop2.7.1.jar" or the -assembly 
> "-assembly-1.6.1-hadoop2.7.2.jar".
> My configuration is the following:
>   * -. available in HIVE_DIR/
>   *  assembly available in HIVE_DIR/lib
> I gathered several logs below:
> - HQL commands
> {noformat}
> $ hive -v --database shfs3453
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application//-1.6.1/assembly/target/scala-2.10/-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Logging initialized using configuration in 
> file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
> use shfs3453
> OK
> Time taken: 1.425 seconds
> Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
> versions. Consider using a different execution engine (i.e. tez, ) or using 
> Hive 1.X releases.
> hive (shfs3453)> set hive.execution.engine=;
> set hive.execution.engine=
> hive (shfs3453)> set .master=yarn-client;
> set .master=yarn-client
> hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, 
> Day STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
> LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
> District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
> YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
> FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED 
> AS TEXTFILE;
> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
> STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
> STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
> CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
> INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
> FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
> OK
> Time taken: 0.408 seconds
> hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
> chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
> Description = 'FIRST DEGREE MURDER'
> Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting  Job = 79484279-8e75-4b13-8e71-7de463f4d51e
> Status: SENT
> Failed to execute  task, with exception 'java.lang.IllegalStateException(RPC 
> channel is closed.)'
> FAILED: Execution , return code 1 from 
> org.apache.hadoop.hive.ql.exec..SparkTask
> {noformat}
> - Client logs
> {noformat}
> May 24 09:32:19 hive-cli  - org.apache.hive..client.rpc.RpcDispatcherReceived 
>  message:io.netty.handler.codec.DecoderException: 
> java.lang.NoClassDefFoundError: org/apache/hive//client/Job
> at 
> io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:358)
> at 
> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:230)
> at 
> i

[jira] [Commented] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-26 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608809#comment-15608809
 ] 

Alexandre Linte commented on HIVE-15017:


I added DEBUG container logs. 

I don't see a line equal to:
{noformat}
LOG.debug("initApplication: " + Arrays.toString(commandArray));
{noformat}

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: debug_yarn_container_mr_job_datanode03.log, 
> debug_yarn_container_mr_job_datanode05.log, hive-site.xml, hive_cli_mr.txt, 
> hive_cli_tez.txt, nodemanager_logs_mr_job.txt, 
> yarn_container_tez_job_datanode05.txt, yarn_container_tez_job_datanode06.txt, 
> yarn_syslog_mr_job.txt, yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-26 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15017:
---
Attachment: debug_yarn_container_mr_job_datanode05.log
debug_yarn_container_mr_job_datanode03.log

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: debug_yarn_container_mr_job_datanode03.log, 
> debug_yarn_container_mr_job_datanode05.log, hive-site.xml, hive_cli_mr.txt, 
> hive_cli_tez.txt, nodemanager_logs_mr_job.txt, 
> yarn_container_tez_job_datanode05.txt, yarn_container_tez_job_datanode06.txt, 
> yarn_syslog_mr_job.txt, yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-24 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15017:
---
Attachment: hive-site.xml

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: hive-site.xml, hive_cli_mr.txt, hive_cli_tez.txt, 
> nodemanager_logs_mr_job.txt, yarn_container_tez_job_datanode05.txt, 
> yarn_container_tez_job_datanode06.txt, yarn_syslog_mr_job.txt, 
> yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-24 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15601270#comment-15601270
 ] 

Alexandre Linte commented on HIVE-15017:


Which logs do you need in DEBUG mode? (Hadoop, Hive)
I added the hive-site.xml to help.
The HADOOP_YARN_HOME env variable is properly set on every datanode. I checked.
There is no "yarn.nodemanager.linux-container-executor.path" property set on 
any of Hadoop configuration files. The default value must be used.
Yes, bin/container-executor is under Yarn home. Is that wrong?

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: hive_cli_mr.txt, hive_cli_tez.txt, 
> nodemanager_logs_mr_job.txt, yarn_container_tez_job_datanode05.txt, 
> yarn_container_tez_job_datanode06.txt, yarn_syslog_mr_job.txt, 
> yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-20 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15017:
---
Attachment: yarn_container_tez_job_datanode06.txt
yarn_container_tez_job_datanode05.txt

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: hive_cli_mr.txt, hive_cli_tez.txt, 
> nodemanager_logs_mr_job.txt, yarn_container_tez_job_datanode05.txt, 
> yarn_container_tez_job_datanode06.txt, yarn_syslog_mr_job.txt, 
> yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-20 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15591369#comment-15591369
 ] 

Alexandre Linte commented on HIVE-15017:


Hi [~sershe],
The "yarn logs" command doesn't return the logs as you can see below.
{noformat}
[root@namenode01 ~]# yarn logs -applicationId application_1475850791417_0105
/Products/YARN/logs/hdfs/logs/application_1475850791417_0105 does not exist.
Log aggregation has not completed or is not enabled.
{noformat}
So I decided to dig into the logs manually. I found interesting things on both 
datanode05 and datanode06. The error "255" appears regularly, I think this is 
the cause of the container crash.

I uploaded the relevant part of the logs.

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: hive_cli_mr.txt, hive_cli_tez.txt, 
> nodemanager_logs_mr_job.txt, yarn_syslog_mr_job.txt, yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-15017) Random job failures with MapReduce and Tez

2016-10-19 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-15017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-15017:
---
Attachment: nodemanager_logs_mr_job.txt
yarn_syslog_mr_job.txt
yarn_syslog_tez_job.txt
hive_cli_mr.txt
hive_cli_tez.txt

> Random job failures with MapReduce and Tez
> --
>
> Key: HIVE-15017
> URL: https://issues.apache.org/jira/browse/HIVE-15017
> Project: Hive
>  Issue Type: Bug
>  Components: Hive
>Affects Versions: 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0
>Reporter: Alexandre Linte
>Priority: Critical
> Attachments: hive_cli_mr.txt, hive_cli_tez.txt, 
> nodemanager_logs_mr_job.txt, yarn_syslog_mr_job.txt, yarn_syslog_tez_job.txt
>
>
> Since Hive 2.1.0, we are facing a blocking issue on our cluster. All the jobs 
> are failing randomly on mapreduce and tez as well. 
> In both case, we don't have any ERROR or WARN message in the logs. You can 
> find attached:
> - hive cli output errors 
> - yarn logs for a tez and mapreduce job
> - nodemanager logs (mr only, we have the same logs with tez)
> Note: This issue doesn't exist with Pig jobs (mr + tez), Spark jobs (mr), so 
> this cannot be an Hadoop / Yarn issue.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-10-05 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15547919#comment-15547919
 ] 

Alexandre Linte commented on HIVE-13830:


Nothing new here?

> Hive on spark driver crash with Spark 1.6.1
> ---
>
> Key: HIVE-13830
> URL: https://issues.apache.org/jira/browse/HIVE-13830
> Project: Hive
>  Issue Type: Bug
>  Components: Spark, spark-branch
>Affects Versions: 2.0.0, 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> With Hive 1.2.1 I was able to use Hive on  successfully with the use of the 
> -assembly "-assembly-1.4.1-hadoop2.7.1.jar". 
> Today with Hive 2.0.0, I'm unable to use Hive on  whether it be with the 
> -assembly "-assembly-1.4.1-hadoop2.7.1.jar" or the -assembly 
> "-assembly-1.6.1-hadoop2.7.2.jar".
> My configuration is the following:
>   * -. available in HIVE_DIR/
>   *  assembly available in HIVE_DIR/lib
> I gathered several logs below:
> - HQL commands
> {noformat}
> $ hive -v --database shfs3453
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application//-1.6.1/assembly/target/scala-2.10/-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Logging initialized using configuration in 
> file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
> use shfs3453
> OK
> Time taken: 1.425 seconds
> Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
> versions. Consider using a different execution engine (i.e. tez, ) or using 
> Hive 1.X releases.
> hive (shfs3453)> set hive.execution.engine=;
> set hive.execution.engine=
> hive (shfs3453)> set .master=yarn-client;
> set .master=yarn-client
> hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, 
> Day STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
> LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
> District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
> YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
> FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED 
> AS TEXTFILE;
> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
> STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
> STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
> CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
> INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
> FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
> OK
> Time taken: 0.408 seconds
> hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
> chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
> Description = 'FIRST DEGREE MURDER'
> Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting  Job = 79484279-8e75-4b13-8e71-7de463f4d51e
> Status: SENT
> Failed to execute  task, with exception 'java.lang.IllegalStateException(RPC 
> channel is closed.)'
> FAILED: Execution , return code 1 from 
> org.apache.hadoop.hive.ql.exec..SparkTask
> {noformat}
> - Client logs
> {noformat}
> May 24 09:32:19 hive-cli  - org.apache.hive..client.rpc.RpcDispatcherReceived 
>  message:io.netty.handler.codec.DecoderException: 
> java.lang.NoClassDefFoundError: org/apache/hive//client/Job
> at 
> io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:358)
> at 
> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:230)
> at 
> io.netty.handler.codec.ByteToMessageCodec.channelRead(ByteToMessage

[jira] [Commented] (HIVE-14631) HiveServer2 regularly fails to connect to metastore

2016-10-05 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15547915#comment-15547915
 ] 

Alexandre Linte commented on HIVE-14631:


Nothing new here?

> HiveServer2 regularly fails to connect to metastore
> ---
>
> Key: HIVE-14631
> URL: https://issues.apache.org/jira/browse/HIVE-14631
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 1.2.1, 2.0.0, 2.1.0
> Environment: Hive 2.1.0, Hue 3.10.0, Hadoop 2.7.2, Tez 0.8.3
>Reporter: Alexandre Linte
>
> I have a cluster secured with Kerberos and Hive is configured to work with 
> Tez by default. Everything works well through hive-cli and beeline; however, 
> I'm facing a strange behavior through Hue.
> I can have a lot of client connections (these can reach 600) and after a day, 
> the client connections fail. But this is not the case for all clients 
> connection attempts.
> When it fails, I have the following logs on the HiveServer2:
> {noformat}
> Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
> command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
>  INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
> Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
> hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
> Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
> Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
> Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
> parallel
> Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:05 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:05 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:06 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:06 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:06 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:08 hiveserver2.bigdata.fr FAILED: Execution Error, return code 
> -1 from org.apache.hadoop.hive.ql.exec.tez.TezTask
> Aug  3 09:28:08 hiveserver2.bigdata.fr Completed executing 
> command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112);
>  Time taken: 4.002 seconds
> {noformat}
> At the same time I have the following logs on the Metastore are:
> {noformat}
> Aug  3 09:28:03 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:03 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:04 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:04 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:04 metastore01.bigdata.fr Error occurred during processing of 
> message.
> Aug  3 09:28:05 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:05 metastore01.bigdata.fr Error occurred during processing of 
> message.
> Aug  3 09:28:06 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:06 metastore01.bigdata.fr Error occurred during processing of 
> message.
> {noformat}
> To solve the connections issue, I have to restart the HiveServer2.
> Note: I also created a JIRA for Hue: 
> https://issues.cloudera.org/browse/HUE-4748



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-14631) HiveServer2 regularly fails to connect to metastore

2016-08-29 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-14631?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15445078#comment-15445078
 ] 

Alexandre Linte commented on HIVE-14631:


It seems to be an Hive issue. When it happens, all jobs are failing even if you 
are using mapreduce or tez. 
{noformat}
0: jdbc:hive2://hiveserver2.bigdata.fr> SET hive.execution.engine=tez;
No rows affected (0.073 seconds)
0: jdbc:hive2://hiveserver2.bigdata.fr> INSERT INTO TABLE shfs3453.camille_test 
VALUES ('coucou');
Error: Error while processing statement: FAILED: Execution Error, return code 1 
from org.apache.hadoop.hive.ql.exec.tez.TezTask (state=08S01,code=1)
0: jdbc:hive2://hiveserver2.bigdata.fr> SET hive.execution.engine=mr;
No rows affected (0.004 seconds)
0: jdbc:hive2://hiveserver2.bigdata.fr> INSERT INTO TABLE shfs3453.camille_test 
VALUES ('coucou');
Error: Error while processing statement: FAILED: Execution Error, return code 1 
from org.apache.hadoop.hive.ql.exec.mr.MapRedTask. 
org.apache.hadoop.security.authentication.client.AuthenticationException: 
Authentication failed, status: 403, message: Forbidden (state=08S01,code=1)
{noformat}
Moreover, this doesn't only affect Hue / Beeswax; all JDBC connections are more 
generally impacted (eg: beeline).

> HiveServer2 regularly fails to connect to metastore
> ---
>
> Key: HIVE-14631
> URL: https://issues.apache.org/jira/browse/HIVE-14631
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2
>Affects Versions: 1.2.1, 2.0.0, 2.1.0
> Environment: Hive 2.1.0, Hue 3.10.0, Hadoop 2.7.2, Tez 0.8.3
>Reporter: Alexandre Linte
>
> I have a cluster secured with Kerberos and Hive is configured to work with 
> Tez by default. Everything works well through hive-cli and beeline; however, 
> I'm facing a strange behavior through Hue.
> I can have a lot of client connections (these can reach 600) and after a day, 
> the client connections fail. But this is not the case for all clients 
> connection attempts.
> When it fails, I have the following logs on the HiveServer2:
> {noformat}
> Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
> command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
>  INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
> Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
> hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
> Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
> Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
> Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
> parallel
> Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:05 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:05 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:06 hiveserver2.bigdata.fr Trying to connect to metastore with 
> URI thrift://metastore01.bigdata.fr:9083
> Aug  3 09:28:06 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
> Server...
> Aug  3 09:28:06 hiveserver2.bigdata.fr Waiting 1 seconds before next 
> connection attempt.
> Aug  3 09:28:08 hiveserver2.bigdata.fr FAILED: Execution Error, return code 
> -1 from org.apache.hadoop.hive.ql.exec.tez.TezTask
> Aug  3 09:28:08 hiveserver2.bigdata.fr Completed executing 
> command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112);
>  Time taken: 4.002 seconds
> {noformat}
> At the same time I have the following logs on the Metastore are:
> {noformat}
> Aug  3 09:28:03 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:03 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:04 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
> tbl=camille_test
> Aug  3 09:28:04 metastore01.bigdata.fr 
> ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
> tbl=camille_test#011
> Aug  3 09:28:04 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28:04 metastore01.bigdata.fr Error occurred during processing of 
> message.
> Aug  3 09:28:05 metastore01.bigdata.fr SASL negotiation failure
> Aug  3 09:28

[jira] [Updated] (HIVE-14631) HiveServer2 regularly fails to connect to metastore

2016-08-25 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-14631:
---
Description: 
I have a cluster secured with Kerberos and Hive is configured to work with Tez 
by default. Everything works well through hive-cli and beeline; however, I'm 
facing a strange behavior through Hue.
I can have a lot of client connections (these can reach 600) and after a day, 
the client connections fail. But this is not the case for all clients 
connection attempts.

When it fails, I have the following logs on the HiveServer2:
{noformat}
Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
 INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
parallel
Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:05 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:05 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:06 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:06 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:06 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:08 hiveserver2.bigdata.fr FAILED: Execution Error, return code -1 
from org.apache.hadoop.hive.ql.exec.tez.TezTask
Aug  3 09:28:08 hiveserver2.bigdata.fr Completed executing 
command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112);
 Time taken: 4.002 seconds
{noformat}

At the same time I have the following logs on the Metastore are:
{noformat}
Aug  3 09:28:03 metastore01.bigdata.fr 180: get_table : db=shfs3453 
tbl=camille_test
Aug  3 09:28:03 metastore01.bigdata.fr 
ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
tbl=camille_test#011
Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
tbl=camille_test
Aug  3 09:28:04 metastore01.bigdata.fr 
ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
tbl=camille_test#011
Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
tbl=camille_test
Aug  3 09:28:04 metastore01.bigdata.fr 
ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
tbl=camille_test#011
Aug  3 09:28:04 metastore01.bigdata.fr SASL negotiation failure
Aug  3 09:28:04 metastore01.bigdata.fr Error occurred during processing of 
message.
Aug  3 09:28:05 metastore01.bigdata.fr SASL negotiation failure
Aug  3 09:28:05 metastore01.bigdata.fr Error occurred during processing of 
message.
Aug  3 09:28:06 metastore01.bigdata.fr SASL negotiation failure
Aug  3 09:28:06 metastore01.bigdata.fr Error occurred during processing of 
message.
{noformat}

Note: I also created a JIRA for Hue: https://issues.cloudera.org/browse/HUE-4748

  was:
I have a cluster secured with Kerberos and Hive is configured to work with Tez 
by default. Everything works well through hive-cli and beeline; however, I'm 
facing a strange behavior through Hue.
I can have a lot of client connections (these can reach 600) and after a day, 
the client connections fail. But this is not the case for all clients 
connection attempts.

When it fails, I have the following logs on the HiveServer2:
{noformat}
Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
 INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
parallel
Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:05 hiveserver2.bigdat

[jira] [Updated] (HIVE-14631) HiveServer2 regularly fails to connect to metastore

2016-08-25 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-14631:
---
Description: 
I have a cluster secured with Kerberos and Hive is configured to work with Tez 
by default. Everything works well through hive-cli and beeline; however, I'm 
facing a strange behavior through Hue.
I can have a lot of client connections (these can reach 600) and after a day, 
the client connections fail. But this is not the case for all clients 
connection attempts.

When it fails, I have the following logs on the HiveServer2:
{noformat}
Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
 INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
parallel
Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:05 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:05 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:06 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:06 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:06 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:08 hiveserver2.bigdata.fr FAILED: Execution Error, return code -1 
from org.apache.hadoop.hive.ql.exec.tez.TezTask
Aug  3 09:28:08 hiveserver2.bigdata.fr Completed executing 
command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112);
 Time taken: 4.002 seconds
{noformat}

At the same time I have the following logs on the Metastore are:
{noformat}
Aug  3 09:28:03 metastore01.bigdata.fr 180: get_table : db=shfs3453 
tbl=camille_test
Aug  3 09:28:03 metastore01.bigdata.fr 
ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
tbl=camille_test#011
Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
tbl=camille_test
Aug  3 09:28:04 metastore01.bigdata.fr 
ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
tbl=camille_test#011
Aug  3 09:28:04 metastore01.bigdata.fr 180: get_table : db=shfs3453 
tbl=camille_test
Aug  3 09:28:04 metastore01.bigdata.fr 
ugi=shfs3453#011ip=10.77.64.228#011cmd=get_table : db=shfs3453 
tbl=camille_test#011
Aug  3 09:28:04 metastore01.bigdata.fr SASL negotiation failure
Aug  3 09:28:04 metastore01.bigdata.fr Error occurred during processing of 
message.
Aug  3 09:28:05 metastore01.bigdata.fr SASL negotiation failure
Aug  3 09:28:05 metastore01.bigdata.fr Error occurred during processing of 
message.
Aug  3 09:28:06 metastore01.bigdata.fr SASL negotiation failure
Aug  3 09:28:06 metastore01.bigdata.fr Error occurred during processing of 
message.
{noformat}

To solve the connections issue, I have to restart the HiveServer2.

Note: I also created a JIRA for Hue: https://issues.cloudera.org/browse/HUE-4748

  was:
I have a cluster secured with Kerberos and Hive is configured to work with Tez 
by default. Everything works well through hive-cli and beeline; however, I'm 
facing a strange behavior through Hue.
I can have a lot of client connections (these can reach 600) and after a day, 
the client connections fail. But this is not the case for all clients 
connection attempts.

When it fails, I have the following logs on the HiveServer2:
{noformat}
Aug  3 09:28:04 hiveserver2.bigdata.fr Executing 
command(queryId=hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112):
 INSERT INTO TABLE shfs3453.camille_test VALUES ('coucou')
Aug  3 09:28:04 hiveserver2.bigdata.fr Query ID = 
hiveserver2_20160803092803_a216edf1-bb51-43a7-81a6-f40f1574b112
Aug  3 09:28:04 hiveserver2.bigdata.fr Total jobs = 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Launching Job 1 out of 1
Aug  3 09:28:04 hiveserver2.bigdata.fr Starting task [Stage-1:MAPRED] in 
parallel
Aug  3 09:28:04 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thrift://metastore01.bigdata.fr:9083
Aug  3 09:28:04 hiveserver2.bigdata.fr Failed to connect to the MetaStore 
Server...
Aug  3 09:28:04 hiveserver2.bigdata.fr Waiting 1 seconds before next connection 
attempt.
Aug  3 09:28:05 hiveserver2.bigdata.fr Trying to connect to metastore with URI 
thr

[jira] [Updated] (HIVE-13790) log4j2 syslog appender not taking "LoggerFields" and "KeyValuePair" options

2016-07-01 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13790?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13790:
---
Affects Version/s: 2.1.0

> log4j2 syslog appender not taking "LoggerFields" and "KeyValuePair" options
> ---
>
> Key: HIVE-13790
> URL: https://issues.apache.org/jira/browse/HIVE-13790
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, Metastore
>Affects Versions: 2.0.0, 2.1.0
> Environment: Hive 2.0.0, Hadoop 2.7.2, Spark 1.6.1, HBase 1.1.2
>Reporter: Alexandre Linte
>
> I'm trying to use the Syslog appender with log4j2 in Hive 2.0.0. The syslog 
> appender is configured on my hiveserver2 and my metastore.
> With a simple configuration, the logs are well written in the logfile with a 
> generic pattern layout:
> {noformat}
> May 19 10:12:16 myhiveserver2.fr Starting HiveServer2
> May 19 10:12:18 myhiveserver2.fr Connected to metastore.
> May 19 10:12:20 myhiveserver2.fr Service: CLIService is inited.
> May 19 10:12:20 myhiveserver2.fr Service: ThriftBinaryCLIService is inited.
> {noformat}
> I tried to customize this pattern layout by adding the loggerFields parameter 
> in my hive-log4j2.properties. At the end, the configuration file is:
> {noformat}
> status = TRACE
> name = HiveLog4j2
> packages = org.apache.hadoop.hive.ql.log
> property.hive.log.level = INFO
> property.hive.root.logger = SYSLOG
> property.hive.query.id = hadoop
> property.hive.log.dir = /var/log/bigdata
> property.hive.log.file = bigdata.log
> appenders = console, SYSLOG
> appender.console.type = Console
> appender.console.name = console
> appender.console.target = SYSTEM_ERR
> appender.console.layout.type = PatternLayout
> appender.console.layout.pattern = %d{yy/MM/dd HH:mm:ss} [%t]: %p %c{2}: %m%n
> appender.SYSLOG.type = Syslog
> appender.SYSLOG.name = SYSLOG
> appender.SYSLOG.host = 127.0.0.1
> appender.SYSLOG.port = 514
> appender.SYSLOG.protocol = UDP
> appender.SYSLOG.facility = LOCAL1
> appender.SYSLOG.layout.type = loggerFields
> appender.SYSLOG.layout.sdId = test
> appender.SYSLOG.layout.enterpriseId = 18060
> appender.SYSLOG.layout.pairs.type = KeyValuePair
> appender.SYSLOG.layout.pairs.key = service
> appender.SYSLOG.layout.pairs.value = hiveserver2
> appender.SYSLOG.layout.pairs.key = loglevel
> appender.SYSLOG.layout.pairs.value = %p
> appender.SYSLOG.layout.pairs.key = message
> appender.SYSLOG.layout.pairs.value = %c%m%n
> loggers = NIOServerCnxn, ClientCnxnSocketNIO, DataNucleus, Datastore, JPOX
> logger.NIOServerCnxn.name = org.apache.zookeeper.server.NIOServerCnxn
> logger.NIOServerCnxn.level = WARN
> logger.ClientCnxnSocketNIO.name = org.apache.zookeeper.ClientCnxnSocketNIO
> logger.ClientCnxnSocketNIO.level = WARN
> logger.DataNucleus.name = DataNucleus
> logger.DataNucleus.level = ERROR
> logger.Datastore.name = Datastore
> logger.Datastore.level = ERROR
> logger.JPOX.name = JPOX
> logger.JPOX.level = ERROR
> rootLogger.level = ${sys:hive.log.level}
> rootLogger.appenderRefs = root
> rootLogger.appenderRef.root.ref = ${sys:hive.root.logger}
> {noformat}
> Unfortunately, the logs are still written in a generic pattern layout. The 
> KeyValuePairs are not used. The log4j logs are:
> {noformat}
> 2016-05-19 10:36:14,866 main DEBUG Initializing configuration 
> org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5433a329
> 2016-05-19 10:36:16,575 main DEBUG Took 1.706004 seconds to load 3 plugins 
> from package org.apache.hadoop.hive.ql.log
> 2016-05-19 10:36:16,575 main DEBUG PluginManager 'Core' found 80 plugins
> 2016-05-19 10:36:16,576 main DEBUG PluginManager 'Level' found 0 plugins
> 2016-05-19 10:36:16,578 main DEBUG Building Plugin[name=property, 
> class=org.apache.logging.log4j.core.config.Property]. Searching for builder 
> factory method...
> 2016-05-19 10:36:16,583 main DEBUG No builder factory method found in class 
> org.apache.logging.log4j.core.config.Property. Going to try finding a factory 
> method instead.
> 2016-05-19 10:36:16,583 main DEBUG Still building Plugin[name=property, 
> class=org.apache.logging.log4j.core.config.Property]. Searching for factory 
> method...
> 2016-05-19 10:36:16,584 main DEBUG Found factory method [createProperty]: 
> public static org.apache.logging.log4j.core.config.Property 
> org.apache.logging.log4j.core.config.Property.createProperty(java.lang.String,java.lang.String).
> 2016-05-19 10:36:16,611 main DEBUG TypeConverterRegistry initializing.
> 2016-05-19 10:36:16,611 main DEBUG PluginManager 'TypeConverter' found 21 
> plugins
> 2016-05-19 10:36:16,636 main DEBUG Calling createProperty on class 
> org.apache.logging.log4j.core.config.Property for element Property with 
> params(name="hive.log.file", value="bigdata.log")
> 2016-05-19 10:36:16,636 main DEBUG Built Plugin[name=p

[jira] [Commented] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-06-28 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15352653#comment-15352653
 ] 

Alexandre Linte commented on HIVE-13830:


Hi,
I upgraded Hive to Hive 2.1.0. Now I have the following errors:
- HQL commands
{noformat}
hive (shfs3453)> SELECT COUNT(year) FROM chicagocrimes GROUP BY year;
SELECT COUNT(year) FROM chicagocrimes GROUP BY year
FAILED: SemanticException Failed to get a spark session: 
org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark client.
{noformat}
- Client logs
{noformat}
Jun 28 11:02:08 hive-cli INFO - org.apache.hadoop.hive.conf.HiveConfUsing the 
default value passed in for log id: c10f51a3-a72d-40c7-9ff6-26e5fb3732da
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.session.SessionStateUpdating thread name to 
c10f51a3-a72d-40c7-9ff6-26e5fb3732da main
Jun 28 11:02:08 hive-cli INFO - org.apache.hadoop.hive.ql.DriverCompiling 
command(queryId=shfs3453_20160628110208_f0b51237-d391-472d-abe8-f2dd2457a9ed): 
SELECT COUNT(year) FROM chicagocrimes GROUP BY year
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerStarting Semantic Analysis
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerCompleted phase 1 of Semantic 
Analysis
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerGet metadata for source tables
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerGet metadata for subqueries
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerGet metadata for destination 
tables
Jun 28 11:02:08 hive-cli INFO - hive.ql.ContextNew scratch dir is 
hdfs://sandbox/tmp/hive/shfs3453/c10f51a3-a72d-40c7-9ff6-26e5fb3732da/hive_2016-06-28_11-02-08_399_7245611464735028300-1
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerCompleted getting MetaData in 
Semantic Analysis
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerGet metadata for source tables
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerGet metadata for subqueries
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerGet metadata for destination 
tables
Jun 28 11:02:08 hive-cli INFO - hive.ql.ContextNew scratch dir is 
hdfs://sandbox/tmp/hive/shfs3453/c10f51a3-a72d-40c7-9ff6-26e5fb3732da/hive_2016-06-28_11-02-08_399_7245611464735028300-1
Jun 28 11:02:08 hive-cli INFO - org.apache.hadoop.hive.common.FileUtilsCreating 
directory if it doesn't exist: 
hdfs://sandbox/tmp/hive/shfs3453/c10f51a3-a72d-40c7-9ff6-26e5fb3732da/hive_2016-06-28_11-02-08_399_7245611464735028300-1/-mr-10001/.hive-staging_hive_2016-06-28_11-02-08_399_7245611464735028300-1
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.parse.CalcitePlannerCBO Succeeded; optimized logical 
plan.
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for FS(6)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for SEL(5)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for GBY(4)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for RS(3)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for GBY(2)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for SEL(1)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.ppd.OpProcFactoryProcessing for TS(0)
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.optimizer.ColumnPrunerProcFactoryRS 3 oldColExprMap: 
{KEY._col0=Column[_col0], VALUE._col0=Column[_col1]}
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.optimizer.ColumnPrunerProcFactoryRS 3 newColExprMap: 
{KEY._col0=Column[_col0], VALUE._col0=Column[_col1]}
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactoryloading spark 
properties from:spark-defaults.conf
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactoryload spark property 
from spark-defaults.conf (spark.default.parallelism -> 10).
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactoryload spark property 
from spark-defaults.conf (spark.kryoserializer.buffer -> 100m).
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactoryload spark property 
from spark-defaults.conf (spark.executor.memory -> 4g).
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactoryload spark property 
from spark-defaults.conf (spark.driver.memory -> 2g).
Jun 28 11:02:08 hive-cli INFO - 
org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactoryload spark property 
from spark-defaults.conf (spark.kryo.classesToRegister -> 
org.ap

[jira] [Updated] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-06-28 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13830:
---
Environment: Hadoop 2.7.2, Hive 2.1.0, Spark 1.6.1, Kerberos  (was: Hadoop 
2.7.2, Hive 2.0.0, Spark 1.6.1, Kerberos)

> Hive on spark driver crash with Spark 1.6.1
> ---
>
> Key: HIVE-13830
> URL: https://issues.apache.org/jira/browse/HIVE-13830
> Project: Hive
>  Issue Type: Bug
>  Components: Spark, spark-branch
>Affects Versions: 2.0.0, 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> With Hive 1.2.1 I was able to use Hive on Spark successfully with the use of 
> the spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar". 
> Today with Hive 2.0.0, I'm unable to use Hive on Spark whether it be with the 
> spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar" or the spark-assembly 
> "spark-assembly-1.6.1-hadoop2.7.2.jar".
> My configuration is the following:
>   * spark-default.conf available in HIVE_DIR/conf
>   * spark assembly available in HIVE_DIR/lib
> I gathered several logs below:
> - HQL commands
> {noformat}
> $ hive -v --database shfs3453
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Spark/spark-1.6.1/assembly/target/scala-2.10/spark-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Logging initialized using configuration in 
> file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
> use shfs3453
> OK
> Time taken: 1.425 seconds
> Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
> versions. Consider using a different execution engine (i.e. tez, spark) or 
> using Hive 1.X releases.
> hive (shfs3453)> set hive.execution.engine=spark;
> set hive.execution.engine=spark
> hive (shfs3453)> set spark.master=yarn-client;
> set spark.master=yarn-client
> hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, 
> Day STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
> LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
> District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
> YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
> FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED 
> AS TEXTFILE;
> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
> STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
> STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
> CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
> INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
> FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
> OK
> Time taken: 0.408 seconds
> hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
> chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
> Description = 'FIRST DEGREE MURDER'
> Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting Spark Job = 79484279-8e75-4b13-8e71-7de463f4d51e
> Status: SENT
> Failed to execute spark task, with exception 
> 'java.lang.IllegalStateException(RPC channel is closed.)'
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask
> {noformat}
> - Client logs
> {noformat}
> May 24 09:32:19 hive-cli WARN - 
> org.apache.hive.spark.client.rpc.RpcDispatcherReceived error 
> message:io.netty.handler.codec.DecoderException: 
> java.lang.NoClassDefFoundError: org/apache/hive/spark/client/Job
> at 
> io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.jav

[jira] [Updated] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-06-28 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13830:
---
Affects Version/s: 2.1.0

> Hive on spark driver crash with Spark 1.6.1
> ---
>
> Key: HIVE-13830
> URL: https://issues.apache.org/jira/browse/HIVE-13830
> Project: Hive
>  Issue Type: Bug
>  Components: Spark, spark-branch
>Affects Versions: 2.0.0, 2.1.0
> Environment: Hadoop 2.7.2, Hive 2.1.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> With Hive 1.2.1 I was able to use Hive on Spark successfully with the use of 
> the spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar". 
> Today with Hive 2.0.0, I'm unable to use Hive on Spark whether it be with the 
> spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar" or the spark-assembly 
> "spark-assembly-1.6.1-hadoop2.7.2.jar".
> My configuration is the following:
>   * spark-default.conf available in HIVE_DIR/conf
>   * spark assembly available in HIVE_DIR/lib
> I gathered several logs below:
> - HQL commands
> {noformat}
> $ hive -v --database shfs3453
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Spark/spark-1.6.1/assembly/target/scala-2.10/spark-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Logging initialized using configuration in 
> file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
> use shfs3453
> OK
> Time taken: 1.425 seconds
> Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
> versions. Consider using a different execution engine (i.e. tez, spark) or 
> using Hive 1.X releases.
> hive (shfs3453)> set hive.execution.engine=spark;
> set hive.execution.engine=spark
> hive (shfs3453)> set spark.master=yarn-client;
> set spark.master=yarn-client
> hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, 
> Day STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
> LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
> District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
> YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
> FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED 
> AS TEXTFILE;
> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
> STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
> STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
> CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
> INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
> FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
> OK
> Time taken: 0.408 seconds
> hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
> chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
> Description = 'FIRST DEGREE MURDER'
> Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting Spark Job = 79484279-8e75-4b13-8e71-7de463f4d51e
> Status: SENT
> Failed to execute spark task, with exception 
> 'java.lang.IllegalStateException(RPC channel is closed.)'
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask
> {noformat}
> - Client logs
> {noformat}
> May 24 09:32:19 hive-cli WARN - 
> org.apache.hive.spark.client.rpc.RpcDispatcherReceived error 
> message:io.netty.handler.codec.DecoderException: 
> java.lang.NoClassDefFoundError: org/apache/hive/spark/client/Job
> at 
> io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:358)
> at 
> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageD

[jira] [Commented] (HIVE-13950) Beeline: infinite loop during the connection to as remote hiveserver2

2016-06-24 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15347907#comment-15347907
 ] 

Alexandre Linte commented on HIVE-13950:


I think this JIRA is partially solved with Hive 2.1.0. Its also related to the 
following ticket: 
- [HIVE-12834|https://issues.apache.org/jira/browse/HIVE-12834]


> Beeline: infinite loop during the connection to as remote hiveserver2
> -
>
> Key: HIVE-13950
> URL: https://issues.apache.org/jira/browse/HIVE-13950
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 2.0.0, 2.0.1
> Environment: Hadoop 2.7.2, Hive 2.0.1, Tez 0.8.3, Kerberos V
>Reporter: Alexandre Linte
>
> From a hive client machine, I can use beeline to connect to a remote 
> Hiveserver2. The connection is secured with Kerberos.
> During the connection process, I have an infinite loop when a username is 
> entered. The loop prints "Enter username for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:"
>  and only stops if the user pushes down "enter" on its keyboard.
> {noformat}
> [shfs3453@hive-cli01 workspace]$ beeline
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.1-bin/lib/hive-jdbc-2.0.1-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.1-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Spark/spark-1.6.1/hive/assembly/spark-assembly-1.4.1-hadoop2.7.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Beeline version 2.0.1 by Apache Hive
> beeline> !connect 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS
> Connecting to 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS
> Enter username for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive

[jira] [Updated] (HIVE-14087) ALTER TABLE table PARTITION requires write permissions

2016-06-24 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-14087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-14087:
---
Component/s: CLI
 Beeline

> ALTER TABLE table PARTITION requires write permissions
> --
>
> Key: HIVE-14087
> URL: https://issues.apache.org/jira/browse/HIVE-14087
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline, CLI, Hive
>Affects Versions: 2.0.1
> Environment: Hadoop 2.7.2, Hive 2.0.1, Kerberos
>Reporter: Alexandre Linte
>
> I discovered that altering a table will require write permissions when a 
> partition is created. 
> {noformat}
> hive (shfs3453)> ALTER TABLE external_table ADD IF NOT EXISTS 
> PARTITION(address='Idaho') LOCATION 
> "hdfs://sandbox/User/shfs3453/WORK/HIVE_TEST";
> ALTER TABLE external_table ADD IF NOT EXISTS PARTITION(address='Idaho') 
> LOCATION "hdfs://sandbox/User/shfs3453/WORK/HIVE_TEST"
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.DDLTask. 
> MetaException(message:java.security.AccessControlException: Permission 
> denied: user=shfs3453, access=WRITE, 
> inode="/User/shfs3453/WORK/HIVE_TEST":shfs3453:shfs3453:dr-xr-x---
> at 
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:319)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:219)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:190)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1720)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1704)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPathAccess(FSDirectory.java:1678)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAccess(FSNamesystem.java:8178)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.checkAccess(NameNodeRpcServer.java:1911)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.checkAccess(ClientNamenodeProtocolServerSideTranslatorPB.java:1443)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043)
> )
> {noformat}
> This is a strange behavior because nothing is written in 
> "/User/shfs3453/WORK/HIVE_TEST".



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HIVE-13790) log4j2 syslog appender not taking "LoggerFields" and "KeyValuePair" options

2016-06-07 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13790?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15318630#comment-15318630
 ] 

Alexandre Linte commented on HIVE-13790:


Hi [~prasanth_j],

I tried to set the keys in MDC without a better result. Hive seems to ignore 
the MDC configuration.

> log4j2 syslog appender not taking "LoggerFields" and "KeyValuePair" options
> ---
>
> Key: HIVE-13790
> URL: https://issues.apache.org/jira/browse/HIVE-13790
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, Metastore
>Affects Versions: 2.0.0
> Environment: Hive 2.0.0, Hadoop 2.7.2, Spark 1.6.1, HBase 1.1.2
>Reporter: Alexandre Linte
>
> I'm trying to use the Syslog appender with log4j2 in Hive 2.0.0. The syslog 
> appender is configured on my hiveserver2 and my metastore.
> With a simple configuration, the logs are well written in the logfile with a 
> generic pattern layout:
> {noformat}
> May 19 10:12:16 myhiveserver2.fr Starting HiveServer2
> May 19 10:12:18 myhiveserver2.fr Connected to metastore.
> May 19 10:12:20 myhiveserver2.fr Service: CLIService is inited.
> May 19 10:12:20 myhiveserver2.fr Service: ThriftBinaryCLIService is inited.
> {noformat}
> I tried to customize this pattern layout by adding the loggerFields parameter 
> in my hive-log4j2.properties. At the end, the configuration file is:
> {noformat}
> status = TRACE
> name = HiveLog4j2
> packages = org.apache.hadoop.hive.ql.log
> property.hive.log.level = INFO
> property.hive.root.logger = SYSLOG
> property.hive.query.id = hadoop
> property.hive.log.dir = /var/log/bigdata
> property.hive.log.file = bigdata.log
> appenders = console, SYSLOG
> appender.console.type = Console
> appender.console.name = console
> appender.console.target = SYSTEM_ERR
> appender.console.layout.type = PatternLayout
> appender.console.layout.pattern = %d{yy/MM/dd HH:mm:ss} [%t]: %p %c{2}: %m%n
> appender.SYSLOG.type = Syslog
> appender.SYSLOG.name = SYSLOG
> appender.SYSLOG.host = 127.0.0.1
> appender.SYSLOG.port = 514
> appender.SYSLOG.protocol = UDP
> appender.SYSLOG.facility = LOCAL1
> appender.SYSLOG.layout.type = loggerFields
> appender.SYSLOG.layout.sdId = test
> appender.SYSLOG.layout.enterpriseId = 18060
> appender.SYSLOG.layout.pairs.type = KeyValuePair
> appender.SYSLOG.layout.pairs.key = service
> appender.SYSLOG.layout.pairs.value = hiveserver2
> appender.SYSLOG.layout.pairs.key = loglevel
> appender.SYSLOG.layout.pairs.value = %p
> appender.SYSLOG.layout.pairs.key = message
> appender.SYSLOG.layout.pairs.value = %c%m%n
> loggers = NIOServerCnxn, ClientCnxnSocketNIO, DataNucleus, Datastore, JPOX
> logger.NIOServerCnxn.name = org.apache.zookeeper.server.NIOServerCnxn
> logger.NIOServerCnxn.level = WARN
> logger.ClientCnxnSocketNIO.name = org.apache.zookeeper.ClientCnxnSocketNIO
> logger.ClientCnxnSocketNIO.level = WARN
> logger.DataNucleus.name = DataNucleus
> logger.DataNucleus.level = ERROR
> logger.Datastore.name = Datastore
> logger.Datastore.level = ERROR
> logger.JPOX.name = JPOX
> logger.JPOX.level = ERROR
> rootLogger.level = ${sys:hive.log.level}
> rootLogger.appenderRefs = root
> rootLogger.appenderRef.root.ref = ${sys:hive.root.logger}
> {noformat}
> Unfortunately, the logs are still written in a generic pattern layout. The 
> KeyValuePairs are not used. The log4j logs are:
> {noformat}
> 2016-05-19 10:36:14,866 main DEBUG Initializing configuration 
> org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5433a329
> 2016-05-19 10:36:16,575 main DEBUG Took 1.706004 seconds to load 3 plugins 
> from package org.apache.hadoop.hive.ql.log
> 2016-05-19 10:36:16,575 main DEBUG PluginManager 'Core' found 80 plugins
> 2016-05-19 10:36:16,576 main DEBUG PluginManager 'Level' found 0 plugins
> 2016-05-19 10:36:16,578 main DEBUG Building Plugin[name=property, 
> class=org.apache.logging.log4j.core.config.Property]. Searching for builder 
> factory method...
> 2016-05-19 10:36:16,583 main DEBUG No builder factory method found in class 
> org.apache.logging.log4j.core.config.Property. Going to try finding a factory 
> method instead.
> 2016-05-19 10:36:16,583 main DEBUG Still building Plugin[name=property, 
> class=org.apache.logging.log4j.core.config.Property]. Searching for factory 
> method...
> 2016-05-19 10:36:16,584 main DEBUG Found factory method [createProperty]: 
> public static org.apache.logging.log4j.core.config.Property 
> org.apache.logging.log4j.core.config.Property.createProperty(java.lang.String,java.lang.String).
> 2016-05-19 10:36:16,611 main DEBUG TypeConverterRegistry initializing.
> 2016-05-19 10:36:16,611 main DEBUG PluginManager 'TypeConverter' found 21 
> plugins
> 2016-05-19 10:36:16,636 main DEBUG Calling createProperty on class 
> org.apache.logging.log4j.core.config.Pr

[jira] [Commented] (HIVE-13950) Beeline: infinite loop during the connection to as remote hiveserver2

2016-06-06 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15316614#comment-15316614
 ] 

Alexandre Linte commented on HIVE-13950:


I think I found the source of this infinite loop. I edited the beeline 
executable like the following:

{noformat}
#!/usr/bin/env bash

# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements.  See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License.  You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

bin=`dirname "$0"`
bin=`cd "$bin"; pwd`

# If process is backgrounded, don't change terminal settings
#if [[ ! $(ps -o stat= -p $$) == *+ ]]; then
#  export HADOOP_CLIENT_OPTS="$HADOOP_CLIENT_OPTS 
-Djline.terminal=jline.UnsupportedTerminal"
#fi

# Set Hadoop User classpath to true so that httpclient jars are taken from
# hive lib instead of hadoop lib.
export HADOOP_USER_CLASSPATH_FIRST=true

# If process is backgrounded, don't change terminal settings
#if [[ ! $(ps -o stat= -p $$) =~ + ]]; then
#  export HADOOP_CLIENT_OPTS="$HADOOP_CLIENT_OPTS 
-Djline.terminal=jline.UnsupportedTerminal"
#fi

. "$bin"/hive --service beeline "$@"
{noformat}
In addition, by commenting the lines I  also solved a bad access to the command 
history. When pressing arrows "up" and "down" was giving me this:
{noformat}
beeline> ^[[A^[[A^[[A^[[A^[[A^[[A^[[A^[[A^[[A
{noformat}

> Beeline: infinite loop during the connection to as remote hiveserver2
> -
>
> Key: HIVE-13950
> URL: https://issues.apache.org/jira/browse/HIVE-13950
> Project: Hive
>  Issue Type: Bug
>  Components: Beeline
>Affects Versions: 2.0.0, 2.0.1
> Environment: Hadoop 2.7.2, Hive 2.0.1, Tez 0.8.3, Kerberos V
>Reporter: Alexandre Linte
>
> From a hive client machine, I can use beeline to connect to a remote 
> Hiveserver2. The connection is secured with Kerberos.
> During the connection process, I have an infinite loop when a username is 
> entered. The loop prints "Enter username for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:"
>  and only stops if the user pushes down "enter" on its keyboard.
> {noformat}
> [shfs3453@hive-cli01 workspace]$ beeline
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.1-bin/lib/hive-jdbc-2.0.1-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.1-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Spark/spark-1.6.1/hive/assembly/spark-assembly-1.4.1-hadoop2.7.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Beeline version 2.0.1 by Apache Hive
> beeline> !connect 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS
> Connecting to 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS
> Enter username for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserver2.bigdata.fr@REALM.KERBEROS:
>   
> Enter password for 
> jdbc:hive2://hiveserver2.bigdata.fr:1/shfs3453;principal=hiveserver2/hiveserve

[jira] [Commented] (HIVE-13790) log4j2 syslog appender not taking "LoggerFields" and "KeyValuePair" options

2016-06-06 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13790?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15316429#comment-15316429
 ] 

Alexandre Linte commented on HIVE-13790:


Nothing new about this JIRA? Nobody tried to use a SYSLOG appender with Hive 
2.0.0?

> log4j2 syslog appender not taking "LoggerFields" and "KeyValuePair" options
> ---
>
> Key: HIVE-13790
> URL: https://issues.apache.org/jira/browse/HIVE-13790
> Project: Hive
>  Issue Type: Bug
>  Components: HiveServer2, Metastore
>Affects Versions: 2.0.0
> Environment: Hive 2.0.0, Hadoop 2.7.2, Spark 1.6.1, HBase 1.1.2
>Reporter: Alexandre Linte
>
> I'm trying to use the Syslog appender with log4j2 in Hive 2.0.0. The syslog 
> appender is configured on my hiveserver2 and my metastore.
> With a simple configuration, the logs are well written in the logfile with a 
> generic pattern layout:
> {noformat}
> May 19 10:12:16 myhiveserver2.fr Starting HiveServer2
> May 19 10:12:18 myhiveserver2.fr Connected to metastore.
> May 19 10:12:20 myhiveserver2.fr Service: CLIService is inited.
> May 19 10:12:20 myhiveserver2.fr Service: ThriftBinaryCLIService is inited.
> {noformat}
> I tried to customize this pattern layout by adding the loggerFields parameter 
> in my hive-log4j2.properties. At the end, the configuration file is:
> {noformat}
> status = TRACE
> name = HiveLog4j2
> packages = org.apache.hadoop.hive.ql.log
> property.hive.log.level = INFO
> property.hive.root.logger = SYSLOG
> property.hive.query.id = hadoop
> property.hive.log.dir = /var/log/bigdata
> property.hive.log.file = bigdata.log
> appenders = console, SYSLOG
> appender.console.type = Console
> appender.console.name = console
> appender.console.target = SYSTEM_ERR
> appender.console.layout.type = PatternLayout
> appender.console.layout.pattern = %d{yy/MM/dd HH:mm:ss} [%t]: %p %c{2}: %m%n
> appender.SYSLOG.type = Syslog
> appender.SYSLOG.name = SYSLOG
> appender.SYSLOG.host = 127.0.0.1
> appender.SYSLOG.port = 514
> appender.SYSLOG.protocol = UDP
> appender.SYSLOG.facility = LOCAL1
> appender.SYSLOG.layout.type = loggerFields
> appender.SYSLOG.layout.sdId = test
> appender.SYSLOG.layout.enterpriseId = 18060
> appender.SYSLOG.layout.pairs.type = KeyValuePair
> appender.SYSLOG.layout.pairs.key = service
> appender.SYSLOG.layout.pairs.value = hiveserver2
> appender.SYSLOG.layout.pairs.key = loglevel
> appender.SYSLOG.layout.pairs.value = %p
> appender.SYSLOG.layout.pairs.key = message
> appender.SYSLOG.layout.pairs.value = %c%m%n
> loggers = NIOServerCnxn, ClientCnxnSocketNIO, DataNucleus, Datastore, JPOX
> logger.NIOServerCnxn.name = org.apache.zookeeper.server.NIOServerCnxn
> logger.NIOServerCnxn.level = WARN
> logger.ClientCnxnSocketNIO.name = org.apache.zookeeper.ClientCnxnSocketNIO
> logger.ClientCnxnSocketNIO.level = WARN
> logger.DataNucleus.name = DataNucleus
> logger.DataNucleus.level = ERROR
> logger.Datastore.name = Datastore
> logger.Datastore.level = ERROR
> logger.JPOX.name = JPOX
> logger.JPOX.level = ERROR
> rootLogger.level = ${sys:hive.log.level}
> rootLogger.appenderRefs = root
> rootLogger.appenderRef.root.ref = ${sys:hive.root.logger}
> {noformat}
> Unfortunately, the logs are still written in a generic pattern layout. The 
> KeyValuePairs are not used. The log4j logs are:
> {noformat}
> 2016-05-19 10:36:14,866 main DEBUG Initializing configuration 
> org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5433a329
> 2016-05-19 10:36:16,575 main DEBUG Took 1.706004 seconds to load 3 plugins 
> from package org.apache.hadoop.hive.ql.log
> 2016-05-19 10:36:16,575 main DEBUG PluginManager 'Core' found 80 plugins
> 2016-05-19 10:36:16,576 main DEBUG PluginManager 'Level' found 0 plugins
> 2016-05-19 10:36:16,578 main DEBUG Building Plugin[name=property, 
> class=org.apache.logging.log4j.core.config.Property]. Searching for builder 
> factory method...
> 2016-05-19 10:36:16,583 main DEBUG No builder factory method found in class 
> org.apache.logging.log4j.core.config.Property. Going to try finding a factory 
> method instead.
> 2016-05-19 10:36:16,583 main DEBUG Still building Plugin[name=property, 
> class=org.apache.logging.log4j.core.config.Property]. Searching for factory 
> method...
> 2016-05-19 10:36:16,584 main DEBUG Found factory method [createProperty]: 
> public static org.apache.logging.log4j.core.config.Property 
> org.apache.logging.log4j.core.config.Property.createProperty(java.lang.String,java.lang.String).
> 2016-05-19 10:36:16,611 main DEBUG TypeConverterRegistry initializing.
> 2016-05-19 10:36:16,611 main DEBUG PluginManager 'TypeConverter' found 21 
> plugins
> 2016-05-19 10:36:16,636 main DEBUG Calling createProperty on class 
> org.apache.logging.log4j.core.config.Property for element Property with 
>

[jira] [Commented] (HIVE-13819) Read & eXecute permissions on Database allows to ALTER it.

2016-06-02 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15312566#comment-15312566
 ] 

Alexandre Linte commented on HIVE-13819:


Hi !

Nothing new on it ? Is it the expected behavior or did i miss something ?

Thanks and regards.

> Read & eXecute permissions on Database allows to ALTER it.
> --
>
> Key: HIVE-13819
> URL: https://issues.apache.org/jira/browse/HIVE-13819
> Project: Hive
>  Issue Type: Bug
>  Components: Authorization
>Affects Versions: 1.2.1
> Environment: Hadoop 2.7.2, Hive 1.2.1, Kerberos.
>Reporter: Alexandre Linte
>
> Hi,
> As the owner of an Hive database I can modify the Hive database metadata 
> whereas I only has the read and execute permission on the Hive database 
> repository.
> I was expected to not be able to modify these metadata.
> Context:
> - Hive database configured with the Storage Based Authorization strategy.
> - Hive client authorization is disabled.
> - Metastore side security is activated.
> Permission configuration:
> {noformat}
> dr-x--x---   - hive9990hive9990 0 2016-05-20 17:10 
> /path/to/hive/warehouse/p09990.db
> {noformat}
> ALTER command as hive9990 user:
> {noformat}
> hive (p09990)>  ALTER DATABASE p09990 SET DBPROPERTIES ('comment'='database 
> altered');
> OK
> Time taken: 0.277 seconds
> hive (p09990)> DESCRIBE DATABASE EXTENDED p09990;
> OK
> p09990  hdfs://path/to/hive/warehouse/p09990.dbhdfsUSER   
>  {comment=database altered}
> {noformat}
> Configuration of hive-site.xml on the metastore:
> {noformat}
> 
> 
> 
>  
>   
>   hive.security.authorization.enabled
>   false
>   enable or disable the Hive client 
> authorization
>   
>   
>   hive.security.metastore.authorization.manager
>   
> org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider
>   authorization manager class name to be used in the 
> metastore for authorization.
>   The user defined authorization class should implement interface 
> org.apache.hadoop.hive.ql.security.authorization.HiveMetastoreAuthorizationProvider.
>   
>   
>   
>   hive.metastore.pre.event.listeners
>   
> org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener
>   This turns on metastore-side security.
>   
>   
>   
>   hive.security.metastore.authorization.auth.reads
>   true
>   If this is true, the metastore authorizer authorizes read 
> actions on database and table.
>   
>   
>   
>   hive.security.authorization.manager
>   
> org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider
>   The Hive client authorization manager class name.
>   The user defined authorization class should implement interface 
> org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.
>   
>   
>   
>   hive.security.authorization.createtable.owner.grants
>   ALL
>   the privileges automatically granted to the owner whenever 
> a table gets created. 
>An example like "select,drop" will grant select and drop privilege to 
> the owner of the table
>   
>   
>   hive.users.in.admin.role
>   hdfs
>   Comma separated list of users who are in admin role for 
> bootstrapping.
> More users can be added in ADMIN role later.
>   
>   
>   hive.metastore.warehouse.dir
>   /path/to/hive/warehouse/
>   location of default database for the 
> warehouse
>   
>   
>   hive.cli.print.current.db
>   true
>   Whether to include the current database in the Hive 
> prompt.
>   
>   
>   hive.metastore.uris
>   thrift://hiveserver2http01:9083
>   Thrift uri for the remote metastore. Used by metastore 
> client to connect to remote metastore.
>   
>   
>   javax.jdo.option.ConnectionDriverName
>   com.mysql.jdbc.Driver
>   JDBC Driver
>   
>   
>   javax.jdo.option.ConnectionURL
>   jdbc:mysql://hivedb01/metastore
>   JDBC connect string for a JDBC metastore
>   
>   
>   javax.jdo.option.ConnectionUserName
>   metastore
>   username to use against metastore database
>   
>   
>   javax.jdo.option.ConnectionPassword
>   
>   password to use against metastore database
>   
>   
>   datanucleus.autoCreateSchema
>   false
>   creates necessary schema on a startup if one doesn't 
> exist. set this to false, after creating it once
>   
>   
>   hive.metastore.authorization.storage.checks
>   true
>   Should the metastore do authorization checks against the 
> underlying storage
>   for operations like drop-partition (disallow the drop-partition if the user 
> in
>   question doesn't have permissions to delete the corresponding directory
>

[jira] [Commented] (HIVE-13829) Property "hive.mapjoin.optimized.keys" does not exist

2016-05-27 Thread Alexandre Linte (JIRA)

[ 
https://issues.apache.org/jira/browse/HIVE-13829?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15303777#comment-15303777
 ] 

Alexandre Linte commented on HIVE-13829:


Hi [~leftylev], thanks for the feedback. In that case, I think this JIRA can be 
closed. 

> Property "hive.mapjoin.optimized.keys" does not exist
> -
>
> Key: HIVE-13829
> URL: https://issues.apache.org/jira/browse/HIVE-13829
> Project: Hive
>  Issue Type: Bug
>  Components: Configuration
>Affects Versions: 2.0.0
> Environment: Hadoop 2.7.2, Hive 2.0.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> Refering to the documentation 
> (https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties), 
> it is possible to set the following property "hive.mapjoin.optimized.keys". 
> Unfortunately, this property seems to be unknown to Hive.
> Here is an extract of the hive-site.xml which includes the property:
> {noformat}
>   
>   hive.mapjoin.optimized.hashtable
>   true
>   Whether Hive should use a memory-optimized hash table for 
> MapJoin. Only works on Tez, because memory-optimized hash table cannot be 
> serialized.
>   
> {noformat}
> In the logs I have:
> {noformat}
> May 24 09:09:02 hiveserver2.bigdata.fr HiveConf of name 
> hive.mapjoin.optimized.keys does not exist
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-05-27 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13830:
---
Component/s: spark-branch

> Hive on spark driver crash with Spark 1.6.1
> ---
>
> Key: HIVE-13830
> URL: https://issues.apache.org/jira/browse/HIVE-13830
> Project: Hive
>  Issue Type: Bug
>  Components: Spark, spark-branch
>Affects Versions: 2.0.0
> Environment: Hadoop 2.7.2, Hive 2.0.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> With Hive 1.2.1 I was able to use Hive on Spark successfully with the use of 
> the spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar". 
> Today with Hive 2.0.0, I'm unable to use Hive on Spark whether it be with the 
> spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar" or the spark-assembly 
> "spark-assembly-1.6.1-hadoop2.7.2.jar".
> My configuration is the following:
>   * spark-default.conf available in HIVE_DIR/conf
>   * spark assembly available in HIVE_DIR/lib
> I gathered several logs below:
> - HQL commands
> {noformat}
> $ hive -v --database shfs3453
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Spark/spark-1.6.1/assembly/target/scala-2.10/spark-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> Logging initialized using configuration in 
> file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
> use shfs3453
> OK
> Time taken: 1.425 seconds
> Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
> versions. Consider using a different execution engine (i.e. tez, spark) or 
> using Hive 1.X releases.
> hive (shfs3453)> set hive.execution.engine=spark;
> set hive.execution.engine=spark
> hive (shfs3453)> set spark.master=yarn-client;
> set spark.master=yarn-client
> hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, 
> Day STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
> LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
> District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
> YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
> FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED 
> AS TEXTFILE;
> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
> STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
> STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
> CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
> INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
> FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
> OK
> Time taken: 0.408 seconds
> hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
> chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
> Description = 'FIRST DEGREE MURDER'
> Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting Spark Job = 79484279-8e75-4b13-8e71-7de463f4d51e
> Status: SENT
> Failed to execute spark task, with exception 
> 'java.lang.IllegalStateException(RPC channel is closed.)'
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask
> {noformat}
> - Client logs
> {noformat}
> May 24 09:32:19 hive-cli WARN - 
> org.apache.hive.spark.client.rpc.RpcDispatcherReceived error 
> message:io.netty.handler.codec.DecoderException: 
> java.lang.NoClassDefFoundError: org/apache/hive/spark/client/Job
> at 
> io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:358)
> at 
> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder

[jira] [Updated] (HIVE-13829) Property "hive.mapjoin.optimized.keys" does not exist

2016-05-24 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13829:
---
Component/s: Configuration

> Property "hive.mapjoin.optimized.keys" does not exist
> -
>
> Key: HIVE-13829
> URL: https://issues.apache.org/jira/browse/HIVE-13829
> Project: Hive
>  Issue Type: Bug
>  Components: Configuration
>Affects Versions: 2.0.0
> Environment: Hadoop 2.7.2, Hive 2.0.0, Spark 1.6.1, Kerberos
>Reporter: Alexandre Linte
>
> Refering to the documentation 
> (https://cwiki.apache.org/confluence/display/Hive/Configuration+Properties), 
> it is possible to set the following property "hive.mapjoin.optimized.keys". 
> Unfortunately, this property seems to be unknown to Hive.
> Here is an extract of the hive-site.xml which includes the property:
> {noformat}
>   
>   hive.mapjoin.optimized.hashtable
>   true
>   Whether Hive should use a memory-optimized hash table for 
> MapJoin. Only works on Tez, because memory-optimized hash table cannot be 
> serialized.
>   
> {noformat}
> In the logs I have:
> {noformat}
> May 24 09:09:02 hiveserver2.bigdata.fr HiveConf of name 
> hive.mapjoin.optimized.keys does not exist
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-05-24 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13830:
---
Description: 
With Hive 1.2.1 I was able to use Hive on Spark successfully with the use of 
the spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar". 
Today with Hive 2.0.0, I'm unable to use Hive on Spark whether it be with the 
spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar" or the spark-assembly 
"spark-assembly-1.6.1-hadoop2.7.2.jar".

My configuration is the following:
  * spark-default.conf available in HIVE_DIR/conf
  * spark assembly available in HIVE_DIR/lib

I gathered several logs below:
- HQL commands
{noformat}
$ hive -v --database shfs3453
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/application/Spark/spark-1.6.1/assembly/target/scala-2.10/spark-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]

Logging initialized using configuration in 
file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
use shfs3453
OK
Time taken: 1.425 seconds
Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
versions. Consider using a different execution engine (i.e. tez, spark) or 
using Hive 1.X releases.
hive (shfs3453)> set hive.execution.engine=spark;
set hive.execution.engine=spark
hive (shfs3453)> set spark.master=yarn-client;
set spark.master=yarn-client
hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day 
STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS 
TEXTFILE;
CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
OK
Time taken: 0.408 seconds
hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
Description = 'FIRST DEGREE MURDER'
Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
Total jobs = 1
Launching Job 1 out of 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=
In order to set a constant number of reducers:
  set mapreduce.job.reduces=
Starting Spark Job = 79484279-8e75-4b13-8e71-7de463f4d51e
Status: SENT
Failed to execute spark task, with exception 
'java.lang.IllegalStateException(RPC channel is closed.)'
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask
{noformat}

- Client logs
{noformat}
May 24 09:32:19 hive-cli WARN - 
org.apache.hive.spark.client.rpc.RpcDispatcherReceived error 
message:io.netty.handler.codec.DecoderException: 
java.lang.NoClassDefFoundError: org/apache/hive/spark/client/Job
at 
io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:358)
at 
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:230)
at 
io.netty.handler.codec.ByteToMessageCodec.channelRead(ByteToMessageCodec.java:103)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
at 
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
at 
io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
at 
io.netty.channel.AbstractChannelHand

[jira] [Updated] (HIVE-13830) Hive on spark driver crash with Spark 1.6.1

2016-05-24 Thread Alexandre Linte (JIRA)

 [ 
https://issues.apache.org/jira/browse/HIVE-13830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexandre Linte updated HIVE-13830:
---
Description: 
With Hive 1.2.1 I was able to use Hive on Spark successfully with the use of 
the spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar". 
Today with Hive 2.0.0, I'm unable to use Hive on Spark whether it be with the 
spark-assembly "spark-assembly-1.4.1-hadoop2.7.1.jar" or the spark-assembly 
"spark-assembly-1.6.1-hadoop2.7.2.jar".

My configuration is the following:
  * spark-default.conf available in HIVE_DIR/conf
  * spark assembly available in HIVE_DIR/lib

I gathered several logs below:
- HQL commands
{noformat}
$ hive -v --database shfs3453
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/hive-jdbc-2.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/application/Hive/apache-hive-2.0.0-bin/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/application/Spark/spark-1.6.1/assembly/target/scala-2.10/spark-assembly-1.6.1-hadoop2.7.2.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/opt/application/Hadoop/hadoop-2.7.2/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]

Logging initialized using configuration in 
file:/opt/application/Hive/apache-hive-2.0.0-bin/conf/hive-log4j2.properties
use shfs3453
OK
Time taken: 1.425 seconds
Hive-on-MR is deprecated in Hive 2 and may not be available in the future 
versions. Consider using a different execution engine (i.e. tez, spark) or 
using Hive 1.X releases.
hive (shfs3453)> set hive.execution.engine=spark;
set hive.execution.engine=spark
hive (shfs3453)> set spark.master=yarn-client;
set spark.master=yarn-client
hive (shfs3453)> CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day 
STRING, Block STRING, IUCR INT, PrimaryType STRING, Description STRING, 
LocationDescription STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, 
District INT, Ward INT, CommunityArea INT, FBICode INT, XCoordinate BIGINT, 
YCoordinate BIGINT, Year INT, UpdatedOn STRING, Latitude FLOAT, Longitude 
FLOAT, Location STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS 
TEXTFILE;
CREATE TABLE chicagoCrimes2 (ID BIGINT, CaseNumber STRING, Day STRING, Block 
STRING, IUCR INT, PrimaryType STRING, Description STRING, LocationDescription 
STRING, Arrest BOOLEAN, Domestic BOOLEAN, Beat INT, District INT, Ward INT, 
CommunityArea INT, FBICode INT, XCoordinate BIGINT, YCoordinate BIGINT, Year 
INT, UpdatedOn STRING, Latitude FLOAT, Longitude FLOAT, Location STRING) ROW 
FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE
OK
Time taken: 0.408 seconds
hive (shfs3453)> INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM 
chicagocrimes WHERE Description = 'FIRST DEGREE MURDER';
INSERT OVERWRITE TABLE chicagocrimes2 SELECT * FROM chicagocrimes WHERE 
Description = 'FIRST DEGREE MURDER'
Query ID = shfs3453_20160524092714_41c89aec-2c6f-49e9-98c7-d227ca144f73
Total jobs = 1
Launching Job 1 out of 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=
In order to set a constant number of reducers:
  set mapreduce.job.reduces=
Starting Spark Job = 79484279-8e75-4b13-8e71-7de463f4d51e
Status: SENT
Failed to execute spark task, with exception 
'java.lang.IllegalStateException(RPC channel is closed.)'
FAILED: Execution Error, return code 1 from 
org.apache.hadoop.hive.ql.exec.spark.SparkTask
{noformat}

- Client logs
{noformat}
May 24 09:32:19 hive-cli WARN - 
org.apache.hive.spark.client.rpc.RpcDispatcherReceived error 
message:io.netty.handler.codec.DecoderException: 
java.lang.NoClassDefFoundError: org/apache/hive/spark/client/Job
at 
io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:358)
at 
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:230)
at 
io.netty.handler.codec.ByteToMessageCodec.channelRead(ByteToMessageCodec.java:103)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
at 
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:294)
at 
io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:86)
at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:308)
at 
io.netty.channel.AbstractChannelHand