RE: Updating hive metadata
Thank you Dev and Ryan. Repair table worked. One last question. Are there any known side effects for repair table? Thanks Prasanth Chagarlamudi From: Ryan Harris [mailto:ryan.har...@zionsbancorp.com] Sent: Thursday, June 18, 2015 11:54 PM To: user@hive.apache.org Subject: RE: Updating hive metadata you *should* be able to do: create my_table_2 like my_table; dfs -cp /user/hive/warehouse/my_table/* /user/hive/warehouse/my_table_2/ MSCK repair table my_table_2; From: Devopam Mittra [mailto:devo...@gmail.com]mailto:[mailto:devo...@gmail.com] Sent: Thursday, June 18, 2015 10:12 PM To: user@hive.apache.orgmailto:user@hive.apache.org Subject: Re: Updating hive metadata hi Prasanth, I would not suggest tweaking hive metastore info unless you have full knowledge of the entire tables that will get impacted due to such a change. And such things break a lot with upgrades since this is quite unmanageable manually. Why don't you create my_managed_table_2 as type EXTERNAL and link it to the copied data in hdfs layer .. regards Dev On Thu, Jun 18, 2015 at 11:40 PM, Chagarlamudi, Prasanth prasanth.chagarlam...@epsilon.commailto:prasanth.chagarlam...@epsilon.com wrote: Hello, Is there a way to update metadata in hive? Created database mydb; Created my_point_table; Created my_managed_table; Insert into my_managed_table from my_point_table; Now, Create my_point_table_2; //Copy data from hive managed_table to managed_table_2’s location hdfs dfs –cp /user/hive/warehouse/mydb.db/my_managed_table /user/hive/warehouse/mydb.db/my_managed_table_2 At this point, I am expecting the following query Select * from my_managed_table_2; to give me all the data I just copied from my_managed_table; How do I update the hive metastore to consider the data that I copied to my_managed_table_2? Is that even possible? Thanks in advance Prasanth Chagarlamudi This e-mail and files transmitted with it are confidential, and are intended solely for the use of the individual or entity to whom this e-mail is addressed. If you are not the intended recipient, or the employee or agent responsible to deliver it to the intended recipient, you are hereby notified that any dissemination, distribution or copying of this communication is strictly prohibited. If you are not one of the named recipient(s) or otherwise have reason to believe that you received this message in error, please immediately notify sender by e-mail, and destroy the original message. Thank You. -- Devopam Mittra Life and Relations are not binary THIS ELECTRONIC MESSAGE, INCLUDING ANY ACCOMPANYING DOCUMENTS, IS CONFIDENTIAL and may contain information that is privileged and exempt from disclosure under applicable law. If you are neither the intended recipient nor responsible for delivering the message to the intended recipient, please note that any dissemination, distribution, copying or the taking of any action in reliance upon the message is strictly prohibited. If you have received this communication in error, please notify the sender immediately. Thank you. This e-mail and files transmitted with it are confidential, and are intended solely for the use of the individual or entity to whom this e-mail is addressed. If you are not the intended recipient, or the employee or agent responsible to deliver it to the intended recipient, you are hereby notified that any dissemination, distribution or copying of this communication is strictly prohibited. If you are not one of the named recipient(s) or otherwise have reason to believe that you received this message in error, please immediately notify sender by e-mail, and destroy the original message. Thank You.
Re: Re: delta file compact take no effect
My hive version is 1.1.0 r7raul1...@163.com From: Alan Gates Date: 2015-06-18 23:25 To: user Subject: Re: delta file compact take no effect Which version of Hive are you running? A number of deadlock issues were resolved in HIVE-10500 which was released in Hive 1.2. Based on your log it appears it recovered properly from the deadlocks and did manage to compact. Alan. r7raul1...@163.com June 17, 2015 at 18:09 It's work~~ But I see some ERROR and Deadlock . 2015-06-18 09:06:06,509 ERROR [test.oracle-22]: txn.CompactionTxnHandler (CompactionTxnHandler.java:findNextToCompact(194)) - Unable to select next element for compaction, ERROR: could not serialize access due to concurrent update 2015-06-18 09:06:06,509 ERROR [test.oracle-27]: txn.CompactionTxnHandler (CompactionTxnHandler.java:findNextToCompact(194)) - Unable to select next element for compaction, ERROR: could not serialize access due to concurrent update 2015-06-18 09:06:06,509 ERROR [test.oracle-28]: txn.CompactionTxnHandler (CompactionTxnHandler.java:findNextToCompact(194)) - Unable to select next element for compaction, ERROR: could not serialize access due to concurrent update 2015-06-18 09:06:06,509 WARN [test.oracle-22]: txn.TxnHandler (TxnHandler.java:checkRetryable(916)) - Deadlock detected in findNextToCompact, trying again. 2015-06-18 09:06:06,509 WARN [test.oracle-27]: txn.TxnHandler (TxnHandler.java:checkRetryable(916)) - Deadlock detected in findNextToCompact, trying again. 2015-06-18 09:06:06,509 WARN [test.oracle-28]: txn.TxnHandler (TxnHandler.java:checkRetryable(916)) - Deadlock detected in findNextToCompact, trying again. 2015-06-18 09:06:06,544 INFO [test.oracle-26]: compactor.Worker (Worker.java:run(140)) - Starting MAJOR compaction for default.u_data_txn 2015-06-18 09:06:06,874 INFO [test.oracle-26]: impl.TimelineClientImpl (TimelineClientImpl.java:serviceInit(123)) - Timeline service address: http://192.168.117.117:8188/ws/v1/timeline/ 2015-06-18 09:06:06,960 INFO [test.oracle-26]: client.RMProxy (RMProxy.java:createRMProxy(92)) - Connecting to ResourceManager at localhost/127.0.0.1:8032 2015-06-18 09:06:07,175 INFO [test.oracle-26]: impl.TimelineClientImpl (TimelineClientImpl.java:serviceInit(123)) - Timeline service address: http://192.168.117.117:8188/ws/v1/timeline/ 2015-06-18 09:06:07,176 INFO [test.oracle-26]: client.RMProxy (RMProxy.java:createRMProxy(92)) - Connecting to ResourceManager at localhost/127.0.0.1:8032 2015-06-18 09:06:07,298 WARN [test.oracle-26]: mapreduce.JobSubmitter (JobSubmitter.java:copyAndConfigureFiles(150)) - Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this. 2015-06-18 09:06:07,777 INFO [test.oracle-26]: mapreduce.JobSubmitter (JobSubmitter.java:submitJobInternal(401)) - number of splits:2 2015-06-18 09:06:07,876 INFO [test.oracle-26]: mapreduce.JobSubmitter (JobSubmitter.java:printTokens(484)) - Submitting tokens for job: job_1433398549746_0035 2015-06-18 09:06:08,021 INFO [test.oracle-26]: impl.YarnClientImpl (YarnClientImpl.java:submitApplication(236)) - Submitted application application_1433398549746_0035 2015-06-18 09:06:08,052 INFO [test.oracle-26]: mapreduce.Job (Job.java:submit(1299)) - The url to track the job: http://localhost:8088/proxy/application_1433398549746_0035/ 2015-06-18 09:06:08,052 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1344)) - Running job: job_1433398549746_0035 2015-06-18 09:06:18,174 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1365)) - Job job_1433398549746_0035 running in uber mode : false 2015-06-18 09:06:18,176 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1372)) - map 0% reduce 0% 2015-06-18 09:06:23,232 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1372)) - map 50% reduce 0% 2015-06-18 09:06:28,262 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1372)) - map 100% reduce 0% 2015-06-18 09:06:28,273 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1383)) - Job job_1433398549746_0035 completed successfully 2015-06-18 09:06:28,327 INFO [test.oracle-26]: mapreduce.Job (Job.java:monitorAndPrintJob(1390)) - Counters: 30 r7raul1...@163.com r7raul1...@163.com June 10, 2015 at 22:10 I use hive 1.1.0 on hadoop 2.5.0 After I do some update operation on table u_data_txn. My table create many delta file like: drwxr-xr-x - hdfs hive 0 2015-02-06 22:52 /user/hive/warehouse/u_data_txn/delta_001_001 -rw-r--r-- 3 hdfs supergroup 346453 2015-02-06 22:52 /user/hive/warehouse/u_data_txn/delta_001_001/bucket_0 -rw-r--r-- 3 hdfs supergroup 415924 2015-02-06 22:52 /user/hive/warehouse/u_data_txn/delta_001_001/bucket_1 drwxr-xr-x - hdfs hive 0 2015-02-06 22:58 /user/hive/warehouse/u_data_txn/delta_002_002 -rw-r--r-- 3 hdfs supergroup 807 2015-02-06 22:58
Re: hive-1.1.0 - alter table rename issue
i solved this issue by creating tables in metastore manually. Thanks Mahesh On Mon, Jun 22, 2015 at 10:48 AM, Mahesh Sankaran sankarmahes...@gmail.com wrote: Yes i upgraded my hive and HMS recently.but there is no mismatch between hive and HMS. On Mon, Jun 22, 2015 at 10:30 AM, Vishnu Viswanath vishnu.viswanat...@gmail.com wrote: from the JIRA comments it looks like there is a miss match between your hive and HMS. Did you upgrade hive recently? On Mon, Jun 22, 2015 at 10:23 AM, Mahesh Sankaran sankarmahes...@gmail.com wrote: Any thoughts regarding this problem?? Mahesh On Sat, Jun 20, 2015 at 10:55 AM, Mahesh Sankaran sankarmahes...@gmail.com wrote: Hi all, I am using hive-1.1.0 and hadoop-2.6.0. am trying to rename a existing table in hive but it shows following error. hive *altertable testrename rename to testrenamebkp ;* *FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Unable to alter table. Invalid method name: 'alter_table_with_cascade'.* *From this error i followed this jira,* https://issues.apache.org/jira/browse/HIVE-10552 to connect hive with beeline but stills it shows the same error. HADOOP_CONF_DIR=/opt/hadoop-2.6.0/etc/hadoop YARN_CONF_DIR=/opt/hadoop-2.6.0/etc/hadoop /opt/hive/bin/beeline -u jdbc:hive2://hostname:1 username password org.apache.hive.jdbc.HiveDriver Connected to: Apache Hive (version 1.1.0) Driver: Hive JDBC (version 1.1.0) Transaction isolation: TRANSACTION_REPEATABLE_READ Beeline version 1.1.0 by Apache Hive 0: jdbc:hive2://hostname:1 *altertable testrename rename to testrenamebkp ;* *FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Unable to alter table. Invalid method name: 'alter_table_with_cascade'* Thanks Mahesh.S
Re: Query Timeout
Hi Ibrar What is the value given for --hiveconf *hbase.master*=# OR --hiveconf *hbase.zookeeper.quorum*= It seems from the error that HBase Server configuration are not correct. Regards Sanjiv Singh Mob : +091 9990-447-339 On Wed, Jun 17, 2015 at 4:31 PM, Ibrar Ahmed ibrar.ah...@gmail.com wrote: I am able to fix that issue, but got another error [127.0.0.1:1] hive CREATE TABLE IF NOT EXISTS pagecounts_hbase (rowkey STRING, pageviews STRING, bytes STRING) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ('hbase.columns.mapping' = ':key,f:c1,f:c2') TBLPROPERTIES (' hbase.table.name' = 'pagecounts'); [Hive Error]: Query returned non-zero code: 1, cause: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:java.lang.IllegalArgumentException: Not a host:port pair: PBUF ibrar-virtual-machine �� �߯��) �� at org.apache.hadoop.hbase.util.Addressing.parseHostname(Addressing.java:60) at org.apache.hadoop.hbase.ServerName.init(ServerName.java:96) at org.apache.hadoop.hbase.ServerName.parseVersionedServerName(ServerName.java:278) at org.apache.hadoop.hbase.MasterAddressTracker.bytesToServerName(MasterAddressTracker.java:77) at org.apache.hadoop.hbase.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:61) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getMaster(HConnectionManager.java:631) at org.apache.hadoop.hbase.client.HBaseAdmin.init(HBaseAdmin.java:106) at org.apache.hadoop.hive.hbase.HBaseStorageHandler.getHBaseAdmin(HBaseStorageHandler.java:84) at org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:162) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:554) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:547) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:89) at com.sun.proxy.$Proxy7.createTable(Unknown Source) at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:613) at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4194) at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:281) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1472) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1239) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1057) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:880) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:870) at org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:198) at org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:644) at org.apache.hadoop.hive.service.ThriftHive$Processor$execute.getResult(ThriftHive.java:628) at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:206) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) On Wed, Jun 17, 2015 at 3:51 PM, Ibrar Ahmed ibrar.ah...@gmail.com wrote: Hi, Whats wrong with my settings? [127.0.0.1:1] hive CREATE TABLE IF NOT EXISTS pagecounts_hbase (rowkey STRING, pageviews STRING, bytes STRING) STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES ('hbase.columns.mapping' = ':key,f:c1,f:c2') TBLPROPERTIES (' hbase.table.name' = 'pagecounts'); [Hive Error]: Query returned non-zero code: 1, cause: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:MetaException(message:org.apache.hadoop.hbase.MasterNotRunningException: Retried 10 times at org.apache.hadoop.hbase.client.HBaseAdmin.init(HBaseAdmin.java:127) at org.apache.hadoop.hive.hbase.HBaseStorageHandler.getHBaseAdmin(HBaseStorageHandler.java:84) at org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:162) at
Re: Hive counters for records read and written
+ Dev Regards Sanjiv Singh Mob : +091 9990-447-339 On Wed, Jun 17, 2015 at 7:46 PM, Hemanth Meka hemanth.m...@datametica.com wrote: Hi, I can see that two new counters have been added for hive (RECORDS_IN/RECORDS_OUT) in hive 0.14. Prior to this release which counters could be used to get the records read by hive job and records written. Because i noticed that in hive 0.14 for a few hive jobs i see map_input_records but the map_output_records counter is 0 but the job actually writes something to output table and the hive log also gives that count correctly. In this case how else can we get records read and records written in releases before 0.14. Regards Hemanth
Re: hive cbo calciteplanner
+ Dev Regards Sanjiv Singh Mob : +091 9990-447-339 On Wed, Jun 17, 2015 at 6:42 AM, wangzhenhua (G) wangzhen...@huawei.com wrote: Hi all, I'm reading the source code of Hive cbo (CalcaitePlanner), but I find it hard to follow. Listed below are some of the questions: 1. What's the relationship between HepPlanner and HiveVolcanoPlanner? 2. I don't have a clue about these concepts: clusters, traitDef and collectGarbage(). Thanks for any help. -- best regards, -zhenhua
Re: HBase and Hive integration
Thanks Sanjiv. I've updated the Hive config setting the hbase.zookeeper.quorum to point to the appropriate zookeeper. On Tue, Jun 23, 2015 at 10:53 AM, Buntu Dev buntu...@gmail.com wrote: Thanks Sanjiv. On 6/23/15, @Sanjiv Singh sanjiv.is...@gmail.com wrote: Hi Buntu, Hive config to provide zookeeper quorum for the HBase cluster --hiveconf hbase.zookeeper.quorum=## Regards Sanjiv Singh Mob : +091 9990-447-339 On Fri, Jun 12, 2015 at 10:04 PM, Buntu Dev buntu...@gmail.com wrote: Thanks Nick for the write up. It was quite helpful for a newbie like me. Is there any Hive config to provide the zookeeper quorum for the HBase cluster since I got Hive and HBase on separate clusters? Thanks! On Tue, Jun 9, 2015 at 12:03 AM, Nick Dimiduk ndimi...@gmail.com wrote: Hi there. I go through a complete example in this pair of blog posts [0], [1]. Basically, create the table with the storage handler, without EXTERNAL and it's lifecycle will be managed by hive. [0]: http://www.n10k.com/blog/hbase-via-hive-pt1/ [1]: http://www.n10k.com/blog/hbase-via-hive-pt2/ On Fri, Jun 5, 2015 at 10:56 AM, Sean Busbey bus...@cloudera.com wrote: +user@hive -user@hbase to bcc Hi! This question is better handled by the hive user list, so I've copied them in and moved the hbase user list to bcc. On Fri, Jun 5, 2015 at 12:54 PM, Buntu Dev buntu...@gmail.com wrote: Hi - Newbie question: I got Hive and HBase on different clusters and say all the appropriate ports are open to connect Hive to HBase, then how to create a Hive managed HBase table? Thanks! -- Sean
Re: HBase and Hive integration
Thanks Sanjiv. On 6/23/15, @Sanjiv Singh sanjiv.is...@gmail.com wrote: Hi Buntu, Hive config to provide zookeeper quorum for the HBase cluster --hiveconf hbase.zookeeper.quorum=## Regards Sanjiv Singh Mob : +091 9990-447-339 On Fri, Jun 12, 2015 at 10:04 PM, Buntu Dev buntu...@gmail.com wrote: Thanks Nick for the write up. It was quite helpful for a newbie like me. Is there any Hive config to provide the zookeeper quorum for the HBase cluster since I got Hive and HBase on separate clusters? Thanks! On Tue, Jun 9, 2015 at 12:03 AM, Nick Dimiduk ndimi...@gmail.com wrote: Hi there. I go through a complete example in this pair of blog posts [0], [1]. Basically, create the table with the storage handler, without EXTERNAL and it's lifecycle will be managed by hive. [0]: http://www.n10k.com/blog/hbase-via-hive-pt1/ [1]: http://www.n10k.com/blog/hbase-via-hive-pt2/ On Fri, Jun 5, 2015 at 10:56 AM, Sean Busbey bus...@cloudera.com wrote: +user@hive -user@hbase to bcc Hi! This question is better handled by the hive user list, so I've copied them in and moved the hbase user list to bcc. On Fri, Jun 5, 2015 at 12:54 PM, Buntu Dev buntu...@gmail.com wrote: Hi - Newbie question: I got Hive and HBase on different clusters and say all the appropriate ports are open to connect Hive to HBase, then how to create a Hive managed HBase table? Thanks! -- Sean
Re: Left function
Or You can wrrite UDF to avail LEFT function in Hive. Follow : http://snowplowanalytics.com/blog/2013/02/08/writing-hive-udfs-and-serdes/ Regards Sanjiv Singh Mob : +091 9990-447-339 On Tue, Jun 16, 2015 at 3:13 PM, Nitin Pawar nitinpawar...@gmail.com wrote: try using substr function On Tue, Jun 16, 2015 at 3:03 PM, Ravisankar Mani rrav...@gmail.com wrote: Hi every one, how to get leftmost length of characters from the string in hive? In Mysql or sq has specific function LEFT(string,length) Could you please help any other way to achieve this scenario? Regards Ravisnkar -- Nitin Pawar
hive -e run tez query error
When I use hive 1.1.0 on tez 0.53 in hadoop 2.3.0: hive -v -e set hive.execution.engine=tez;set mapred.job.queue.name=bi_etl;drop table TESTTMP.a_start;create table TESTTMP.a_start(id bigint);insert overwrite table TESTTMP.a_start select id from tandem.p_city;drop table TESTTMP.a_end;create table TESTTMP.a_end(id bigint);insert overwrite table TESTTMP.a_end select id from TESTTMP.a_start; Logging initialized using configuration in jar:file:/usr/local/src/apache-hive/lib/hive-common-1.1.0.jar!/hive-log4j.properties SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/usr/local/src/apache-hive/lib/hive-jdbc-1.1.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/usr/lib/zookeeper/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] set hive.execution.engine=tez set mapred.job.queue.name=bi_etl drop table TESTTMP.a_start OK Time taken: 1.702 seconds create table TESTTMP.a_start(id bigint) OK Time taken: 0.311 seconds insert overwrite table TESTTMP.a_start select id from tandem.p_city Query ID = lujian_20150623134848_31ba3183-74b8-4c3a-96ae-5c8b650b99df Total jobs = 1 Launching Job 1 out of 1 Status: Running (Executing on YARN cluster with App id application_1433219182593_252390) VERTICES STATUS TOTAL COMPLETED RUNNING PENDING FAILED KILLED Map 1 .. SUCCEEDED 1 1 0 0 0 0 VERTICES: 01/01 [==] 100% ELAPSED TIME: 13.55 s Loading data to table testtmp.a_start Moved: 'hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testtmp.db/a_start/00_0' to trash at: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/.Trash/Current Table testtmp.a_start stats: [numFiles=1, numRows=531, totalSize=2154, rawDataSize=1623] OK Time taken: 25.586 seconds drop table TESTTMP.a_end OK Time taken: 0.232 seconds create table TESTTMP.a_end(id bigint) OK Time taken: 0.068 seconds insert overwrite table TESTTMP.a_end select id from TESTTMP.a_start Query ID = lujian_20150623134949_bff735c9-6abc-47e7-a9f7-f7e2be7e43e9 Total jobs = 1 Launching Job 1 out of 1 Status: Running (Executing on YARN cluster with App id application_1433219182593_252390) VERTICES STATUS TOTAL COMPLETED RUNNING PENDING FAILED KILLED Map 1 FAILED 1 0 0 1 4 0 VERTICES: 00/01 [--] 0% ELAPSED TIME: 15.46 s Status: Failed Vertex failed, vertexName=Map 1, vertexId=vertex_1433219182593_252390_2_00, diagnostics=[Task failed, taskId=task_1433219182593_252390_2_00_00, diagnostics=[TaskAttempt 0 failed, info=[Container container_1433219182593_252390_01_03 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb ]], TaskAttempt 1 failed, info=[Container container_1433219182593_252390_01_04 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb ]], TaskAttempt 2 failed, info=[Container container_1433219182593_252390_01_05 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb ]], TaskAttempt 3 failed, info=[Container container_1433219182593_252390_01_06 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb ]]], Vertex failed as one or more tasks failed. failedTasks:1, Vertex vertex_1433219182593_252390_2_00 [Map 1] killed/failed due to:null] DAG failed due to vertex failure. failedVertices:1 killedVertices:0 FAILED: Execution Error, return code 2 from