Re: Container launch failed Error

2014-11-24 Thread Amit Behera
Hi Daniel,

Thank you , Its running fine.

*Another question:*
 could you please tell me what to do If I will get *Shuffle Error*.
one time I got this type of error while running a join query on 300GB data
with 20GB data


Thanks
Amit

On Mon, Nov 24, 2014 at 11:13 PM, Daniel Haviv <
daniel.ha...@veracity-group.com> wrote:

> Good luck
> Share your results with us
>
> Daniel
>
> On 24 בנוב׳ 2014, at 19:36, Amit Behera  wrote:
>
> Hi Daniel,
>
> Thanks a lot,
>
>
> I will do that and rerun the query. :)
>
> On Mon, Nov 24, 2014 at 10:59 PM, Daniel Haviv <
> daniel.ha...@veracity-group.com> wrote:
>
>> It is a problem as the application master needs to contact the other nodes
>>
>> Try updating the hosts file on all the machines and try again.
>>
>> Daniel
>>
>> On 24 בנוב׳ 2014, at 19:26, Amit Behera  wrote:
>>
>> I did not modify in all the slaves. except slave
>>
>> will it be a problem ?
>>
>> But for small data (up to 20 GB table) it is running and for 300GB table
>> only count(*) running sometimes and sometimes failed
>>
>> Thanks
>> Amit
>>
>> On Mon, Nov 24, 2014 at 10:37 PM, Daniel Haviv <
>> daniel.ha...@veracity-group.com> wrote:
>>
>>> did you copy the hosts file to all the nodes?
>>>
>>> Daniel
>>>
>>> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
>>>
>>> hi Daniel,
>>>
>>>
>>> this stacktrace same for other query .
>>> for different run I am getting slave7 sometime slave8...
>>>
>>> And also I registered all machine IPs in /etc/hosts
>>>
>>> Regards
>>> Amit
>>>
>>>
>>>
>>> On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv <
>>> daniel.ha...@veracity-group.com> wrote:
>>>
 It seems that the application master can't resolve slave6's name to an
 IP

 Daniel

 On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:

 Hi Users,

 *my cluster(1+8) configuration*:

 RAM  : 32 GB each
 HDFS : 1.5 TB SSD
 CPU   : 8 core each

 ---

 I am trying to query on 300GB of table but I am able to run only select
 query.

 Except select query , for all other query I am getting following
 exception.





 Total jobs = 1

 Stage-1 is selected by condition resolver.

 Launching Job 1 out of 1

 Number of reduce tasks not specified. Estimated
 from input data size: 183

 In order to change the average load for a
 reducer (in bytes):

   set
 hive.exec.reducers.bytes.per.reducer=

 In order to limit the maximum number of
 reducers:

   set hive.exec.reducers.max=

 In order to set a constant number of reducers:

   set mapreduce.job.reduces=

 Starting Job = job_1416831990090_0005, Tracking
 URL = http://master:8088/proxy/application_1416831990090_0005/

 Kill Command = /root/hadoop/bin/hadoop job
 -kill job_1416831990090_0005

 Hadoop job information for Stage-1: number of
 mappers: 679; number of reducers: 183

 2014-11-24 19:43:01,523 Stage-1 map = 0%,
 reduce = 0%

 2014-11-24 19:43:22,730 Stage-1 map = 53%,
 reduce = 0%, Cumulative CPU 625.19 sec

 2014-11-24 19:43:23,778 Stage-1 map = 100%,
 reduce = 100%

 MapReduce Total cumulative CPU time: 10 minutes
 25 seconds 190 msec

 Ended Job = job_1416831990090_0005 with errors

 Error during job, obtaining debugging
 information...

 Examining task ID:
 task_1416831990090_0005_m_05 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_42 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_35 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_65 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_02 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_07 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_58 (and more) from job
 job_1416831990090_0005

 Examining task ID:
 task_1416831990090_0005_m_43 (and more) from job
 job_1416831990090_0005


  Task with the most failures(4):

 -

 Task ID:

   task_1416831990090_0005_m_05


  URL:

  
 http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05

 -

 Diagnostic Messages for this Task:

 Container launch failed for
 container_1416831990090_0005_01_000112 :
 java.lang.IllegalArgumentException: java.net.UnknownHostException:
 slave6

at
 org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)

 

Re: Container launch failed Error

2014-11-24 Thread Daniel Haviv
Good luck
Share your results with us

Daniel

> On 24 בנוב׳ 2014, at 19:36, Amit Behera  wrote:
> 
> Hi Daniel,
> 
> Thanks a lot,
> 
> 
> I will do that and rerun the query. :)
> 
>> On Mon, Nov 24, 2014 at 10:59 PM, Daniel Haviv 
>>  wrote:
>> It is a problem as the application master needs to contact the other nodes
>> 
>> Try updating the hosts file on all the machines and try again.
>> 
>> Daniel
>> 
>>> On 24 בנוב׳ 2014, at 19:26, Amit Behera  wrote:
>>> 
>>> I did not modify in all the slaves. except slave 
>>> 
>>> will it be a problem ?
>>> 
>>> But for small data (up to 20 GB table) it is running and for 300GB table 
>>> only count(*) running sometimes and sometimes failed  
>>> 
>>> Thanks
>>> Amit  
>>> 
 On Mon, Nov 24, 2014 at 10:37 PM, Daniel Haviv 
  wrote:
 did you copy the hosts file to all the nodes?
 
 Daniel
 
> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
> 
> hi Daniel,
> 
> 
> this stacktrace same for other query .
> for different run I am getting slave7 sometime slave8... 
> 
> And also I registered all machine IPs in /etc/hosts 
> 
> Regards
> Amit
> 
> 
> 
>> On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv 
>>  wrote:
>> It seems that the application master can't resolve slave6's name to an IP
>> 
>> Daniel
>> 
>>> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
>>> 
>>> Hi Users,
>>> 
>>> my cluster(1+8) configuration:
>>> 
>>> RAM  : 32 GB each
>>> HDFS : 1.5 TB SSD
>>> CPU   : 8 core each
>>> 
>>> ---
>>> 
>>> I am trying to query on 300GB of table but I am able to run only select 
>>> query.
>>> 
>>> Except select query , for all other query I am getting following 
>>> exception.
>>> 
>>> 
>>> 
>>> 
>>> 
>>> Total jobs = 1
>>> Stage-1 is selected by condition resolver.
>>> Launching Job 1 out of 1
>>> Number of reduce tasks not specified. Estimated
>>> from input data size: 183
>>> In order to change the average load for a
>>> reducer (in bytes):
>>>   set
>>> hive.exec.reducers.bytes.per.reducer=
>>> In order to limit the maximum number of
>>> reducers:
>>>   set hive.exec.reducers.max=
>>> In order to set a constant number of reducers:
>>>   set mapreduce.job.reduces=
>>> Starting Job = job_1416831990090_0005, Tracking
>>> URL = http://master:8088/proxy/application_1416831990090_0005/
>>> Kill Command = /root/hadoop/bin/hadoop job 
>>> -kill job_1416831990090_0005
>>> Hadoop job information for Stage-1: number of
>>> mappers: 679; number of reducers: 183
>>> 2014-11-24 19:43:01,523 Stage-1 map = 0%, 
>>> reduce = 0%
>>> 2014-11-24 19:43:22,730 Stage-1 map = 53%, 
>>> reduce = 0%, Cumulative CPU 625.19 sec
>>> 2014-11-24 19:43:23,778 Stage-1 map = 100%, 
>>> reduce = 100%
>>> MapReduce Total cumulative CPU time: 10 minutes
>>> 25 seconds 190 msec
>>> Ended Job = job_1416831990090_0005 with errors
>>> Error during job, obtaining debugging
>>> information...
>>> Examining task ID:
>>> task_1416831990090_0005_m_05 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_42 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_35 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_65 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_02 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_07 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_58 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_43 (and more) from job
>>> job_1416831990090_0005
>>> 
>>> 
>>> Task with the most failures(4): 
>>> -
>>> Task ID:
>>>   task_1416831990090_0005_m_05
>>> 
>>> 
>>> URL:
>>>  
>>> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
>>> -
>>> Diagnostic Messages for this Task:
>>> Container launch failed for
>>> container_1416831990090_0005_01_000112 :
>>> java.lang.IllegalArgumentException: java.net.UnknownHostException:
>>> slave6
>>> at
>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>>> at
>>> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>>> at
>>> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)

Re: Container launch failed Error

2014-11-24 Thread Amit Behera
Hi Daniel,

Thanks a lot,


I will do that and rerun the query. :)

On Mon, Nov 24, 2014 at 10:59 PM, Daniel Haviv <
daniel.ha...@veracity-group.com> wrote:

> It is a problem as the application master needs to contact the other nodes
>
> Try updating the hosts file on all the machines and try again.
>
> Daniel
>
> On 24 בנוב׳ 2014, at 19:26, Amit Behera  wrote:
>
> I did not modify in all the slaves. except slave
>
> will it be a problem ?
>
> But for small data (up to 20 GB table) it is running and for 300GB table
> only count(*) running sometimes and sometimes failed
>
> Thanks
> Amit
>
> On Mon, Nov 24, 2014 at 10:37 PM, Daniel Haviv <
> daniel.ha...@veracity-group.com> wrote:
>
>> did you copy the hosts file to all the nodes?
>>
>> Daniel
>>
>> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
>>
>> hi Daniel,
>>
>>
>> this stacktrace same for other query .
>> for different run I am getting slave7 sometime slave8...
>>
>> And also I registered all machine IPs in /etc/hosts
>>
>> Regards
>> Amit
>>
>>
>>
>> On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv <
>> daniel.ha...@veracity-group.com> wrote:
>>
>>> It seems that the application master can't resolve slave6's name to an IP
>>>
>>> Daniel
>>>
>>> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
>>>
>>> Hi Users,
>>>
>>> *my cluster(1+8) configuration*:
>>>
>>> RAM  : 32 GB each
>>> HDFS : 1.5 TB SSD
>>> CPU   : 8 core each
>>>
>>> ---
>>>
>>> I am trying to query on 300GB of table but I am able to run only select
>>> query.
>>>
>>> Except select query , for all other query I am getting following
>>> exception.
>>>
>>>
>>>
>>>
>>>
>>> Total jobs = 1
>>>
>>> Stage-1 is selected by condition resolver.
>>>
>>> Launching Job 1 out of 1
>>>
>>> Number of reduce tasks not specified. Estimated
>>> from input data size: 183
>>>
>>> In order to change the average load for a
>>> reducer (in bytes):
>>>
>>>   set
>>> hive.exec.reducers.bytes.per.reducer=
>>>
>>> In order to limit the maximum number of
>>> reducers:
>>>
>>>   set hive.exec.reducers.max=
>>>
>>> In order to set a constant number of reducers:
>>>
>>>   set mapreduce.job.reduces=
>>>
>>> Starting Job = job_1416831990090_0005, Tracking
>>> URL = http://master:8088/proxy/application_1416831990090_0005/
>>>
>>> Kill Command = /root/hadoop/bin/hadoop job
>>> -kill job_1416831990090_0005
>>>
>>> Hadoop job information for Stage-1: number of
>>> mappers: 679; number of reducers: 183
>>>
>>> 2014-11-24 19:43:01,523 Stage-1 map = 0%,
>>> reduce = 0%
>>>
>>> 2014-11-24 19:43:22,730 Stage-1 map = 53%,
>>> reduce = 0%, Cumulative CPU 625.19 sec
>>>
>>> 2014-11-24 19:43:23,778 Stage-1 map = 100%,
>>> reduce = 100%
>>>
>>> MapReduce Total cumulative CPU time: 10 minutes
>>> 25 seconds 190 msec
>>>
>>> Ended Job = job_1416831990090_0005 with errors
>>>
>>> Error during job, obtaining debugging
>>> information...
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_05 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_42 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_35 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_65 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_02 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_07 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_58 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_43 (and more) from job
>>> job_1416831990090_0005
>>>
>>>
>>>  Task with the most failures(4):
>>>
>>> -
>>>
>>> Task ID:
>>>
>>>   task_1416831990090_0005_m_05
>>>
>>>
>>>  URL:
>>>
>>>  
>>> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
>>>
>>> -
>>>
>>> Diagnostic Messages for this Task:
>>>
>>> Container launch failed for
>>> container_1416831990090_0005_01_000112 :
>>> java.lang.IllegalArgumentException: java.net.UnknownHostException:
>>> slave6
>>>
>>> at
>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>>>
>>> at
>>> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>>>
>>> at
>>> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>>>
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>>>
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>>>
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocol

Re: Container launch failed Error

2014-11-24 Thread Daniel Haviv
It is a problem as the application master needs to contact the other nodes

Try updating the hosts file on all the machines and try again.

Daniel

> On 24 בנוב׳ 2014, at 19:26, Amit Behera  wrote:
> 
> I did not modify in all the slaves. except slave 
> 
> will it be a problem ?
> 
> But for small data (up to 20 GB table) it is running and for 300GB table only 
> count(*) running sometimes and sometimes failed  
> 
> Thanks
> Amit  
> 
>> On Mon, Nov 24, 2014 at 10:37 PM, Daniel Haviv 
>>  wrote:
>> did you copy the hosts file to all the nodes?
>> 
>> Daniel
>> 
>>> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
>>> 
>>> hi Daniel,
>>> 
>>> 
>>> this stacktrace same for other query .
>>> for different run I am getting slave7 sometime slave8... 
>>> 
>>> And also I registered all machine IPs in /etc/hosts 
>>> 
>>> Regards
>>> Amit
>>> 
>>> 
>>> 
 On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv 
  wrote:
 It seems that the application master can't resolve slave6's name to an IP
 
 Daniel
 
> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
> 
> Hi Users,
> 
> my cluster(1+8) configuration:
> 
> RAM  : 32 GB each
> HDFS : 1.5 TB SSD
> CPU   : 8 core each
> 
> ---
> 
> I am trying to query on 300GB of table but I am able to run only select 
> query.
> 
> Except select query , for all other query I am getting following 
> exception.
> 
> 
> 
> 
> 
> Total jobs = 1
> Stage-1 is selected by condition resolver.
> Launching Job 1 out of 1
> Number of reduce tasks not specified. Estimated
> from input data size: 183
> In order to change the average load for a
> reducer (in bytes):
>   set
> hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of
> reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting Job = job_1416831990090_0005, Tracking
> URL = http://master:8088/proxy/application_1416831990090_0005/
> Kill Command = /root/hadoop/bin/hadoop job 
> -kill job_1416831990090_0005
> Hadoop job information for Stage-1: number of
> mappers: 679; number of reducers: 183
> 2014-11-24 19:43:01,523 Stage-1 map = 0%, 
> reduce = 0%
> 2014-11-24 19:43:22,730 Stage-1 map = 53%, 
> reduce = 0%, Cumulative CPU 625.19 sec
> 2014-11-24 19:43:23,778 Stage-1 map = 100%, 
> reduce = 100%
> MapReduce Total cumulative CPU time: 10 minutes
> 25 seconds 190 msec
> Ended Job = job_1416831990090_0005 with errors
> Error during job, obtaining debugging
> information...
> Examining task ID:
> task_1416831990090_0005_m_05 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_42 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_35 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_65 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_02 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_07 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_58 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_43 (and more) from job
> job_1416831990090_0005
> 
> 
> Task with the most failures(4): 
> -
> Task ID:
>   task_1416831990090_0005_m_05
> 
> 
> URL:
>  
> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
> -
> Diagnostic Messages for this Task:
> Container launch failed for
> container_1416831990090_0005_01_000112 :
> java.lang.IllegalArgumentException: java.net.UnknownHostException:
> slave6
>   at
> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>   at
> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>   at
> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.Contain

Re: Container launch failed Error

2014-11-24 Thread Amit Behera
* except slave6, slave7, slave8

On Mon, Nov 24, 2014 at 10:56 PM, Amit Behera  wrote:

> I did not modify in all the slaves. except slave
>
> will it be a problem ?
>
> But for small data (up to 20 GB table) it is running and for 300GB table
> only count(*) running sometimes and sometimes failed
>
> Thanks
> Amit
>
> On Mon, Nov 24, 2014 at 10:37 PM, Daniel Haviv <
> daniel.ha...@veracity-group.com> wrote:
>
>> did you copy the hosts file to all the nodes?
>>
>> Daniel
>>
>> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
>>
>> hi Daniel,
>>
>>
>> this stacktrace same for other query .
>> for different run I am getting slave7 sometime slave8...
>>
>> And also I registered all machine IPs in /etc/hosts
>>
>> Regards
>> Amit
>>
>>
>>
>> On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv <
>> daniel.ha...@veracity-group.com> wrote:
>>
>>> It seems that the application master can't resolve slave6's name to an IP
>>>
>>> Daniel
>>>
>>> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
>>>
>>> Hi Users,
>>>
>>> *my cluster(1+8) configuration*:
>>>
>>> RAM  : 32 GB each
>>> HDFS : 1.5 TB SSD
>>> CPU   : 8 core each
>>>
>>> ---
>>>
>>> I am trying to query on 300GB of table but I am able to run only select
>>> query.
>>>
>>> Except select query , for all other query I am getting following
>>> exception.
>>>
>>>
>>>
>>>
>>>
>>> Total jobs = 1
>>>
>>> Stage-1 is selected by condition resolver.
>>>
>>> Launching Job 1 out of 1
>>>
>>> Number of reduce tasks not specified. Estimated
>>> from input data size: 183
>>>
>>> In order to change the average load for a
>>> reducer (in bytes):
>>>
>>>   set
>>> hive.exec.reducers.bytes.per.reducer=
>>>
>>> In order to limit the maximum number of
>>> reducers:
>>>
>>>   set hive.exec.reducers.max=
>>>
>>> In order to set a constant number of reducers:
>>>
>>>   set mapreduce.job.reduces=
>>>
>>> Starting Job = job_1416831990090_0005, Tracking
>>> URL = http://master:8088/proxy/application_1416831990090_0005/
>>>
>>> Kill Command = /root/hadoop/bin/hadoop job
>>> -kill job_1416831990090_0005
>>>
>>> Hadoop job information for Stage-1: number of
>>> mappers: 679; number of reducers: 183
>>>
>>> 2014-11-24 19:43:01,523 Stage-1 map = 0%,
>>> reduce = 0%
>>>
>>> 2014-11-24 19:43:22,730 Stage-1 map = 53%,
>>> reduce = 0%, Cumulative CPU 625.19 sec
>>>
>>> 2014-11-24 19:43:23,778 Stage-1 map = 100%,
>>> reduce = 100%
>>>
>>> MapReduce Total cumulative CPU time: 10 minutes
>>> 25 seconds 190 msec
>>>
>>> Ended Job = job_1416831990090_0005 with errors
>>>
>>> Error during job, obtaining debugging
>>> information...
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_05 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_42 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_35 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_65 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_02 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_07 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_58 (and more) from job
>>> job_1416831990090_0005
>>>
>>> Examining task ID:
>>> task_1416831990090_0005_m_43 (and more) from job
>>> job_1416831990090_0005
>>>
>>>
>>>  Task with the most failures(4):
>>>
>>> -
>>>
>>> Task ID:
>>>
>>>   task_1416831990090_0005_m_05
>>>
>>>
>>>  URL:
>>>
>>>  
>>> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
>>>
>>> -
>>>
>>> Diagnostic Messages for this Task:
>>>
>>> Container launch failed for
>>> container_1416831990090_0005_01_000112 :
>>> java.lang.IllegalArgumentException: java.net.UnknownHostException:
>>> slave6
>>>
>>> at
>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>>>
>>> at
>>> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>>>
>>> at
>>> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>>>
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>>>
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>>>
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)
>>>
>>> at
>>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl.getCMProxy(ContainerLauncherImpl.java:403)
>>>
>>> at
>>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Containe

Re: Container launch failed Error

2014-11-24 Thread Amit Behera
I did not modify in all the slaves. except slave

will it be a problem ?

But for small data (up to 20 GB table) it is running and for 300GB table
only count(*) running sometimes and sometimes failed

Thanks
Amit

On Mon, Nov 24, 2014 at 10:37 PM, Daniel Haviv <
daniel.ha...@veracity-group.com> wrote:

> did you copy the hosts file to all the nodes?
>
> Daniel
>
> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
>
> hi Daniel,
>
>
> this stacktrace same for other query .
> for different run I am getting slave7 sometime slave8...
>
> And also I registered all machine IPs in /etc/hosts
>
> Regards
> Amit
>
>
>
> On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv <
> daniel.ha...@veracity-group.com> wrote:
>
>> It seems that the application master can't resolve slave6's name to an IP
>>
>> Daniel
>>
>> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
>>
>> Hi Users,
>>
>> *my cluster(1+8) configuration*:
>>
>> RAM  : 32 GB each
>> HDFS : 1.5 TB SSD
>> CPU   : 8 core each
>>
>> ---
>>
>> I am trying to query on 300GB of table but I am able to run only select
>> query.
>>
>> Except select query , for all other query I am getting following
>> exception.
>>
>>
>>
>>
>>
>> Total jobs = 1
>>
>> Stage-1 is selected by condition resolver.
>>
>> Launching Job 1 out of 1
>>
>> Number of reduce tasks not specified. Estimated
>> from input data size: 183
>>
>> In order to change the average load for a
>> reducer (in bytes):
>>
>>   set
>> hive.exec.reducers.bytes.per.reducer=
>>
>> In order to limit the maximum number of
>> reducers:
>>
>>   set hive.exec.reducers.max=
>>
>> In order to set a constant number of reducers:
>>
>>   set mapreduce.job.reduces=
>>
>> Starting Job = job_1416831990090_0005, Tracking
>> URL = http://master:8088/proxy/application_1416831990090_0005/
>>
>> Kill Command = /root/hadoop/bin/hadoop job
>> -kill job_1416831990090_0005
>>
>> Hadoop job information for Stage-1: number of
>> mappers: 679; number of reducers: 183
>>
>> 2014-11-24 19:43:01,523 Stage-1 map = 0%,
>> reduce = 0%
>>
>> 2014-11-24 19:43:22,730 Stage-1 map = 53%,
>> reduce = 0%, Cumulative CPU 625.19 sec
>>
>> 2014-11-24 19:43:23,778 Stage-1 map = 100%,
>> reduce = 100%
>>
>> MapReduce Total cumulative CPU time: 10 minutes
>> 25 seconds 190 msec
>>
>> Ended Job = job_1416831990090_0005 with errors
>>
>> Error during job, obtaining debugging
>> information...
>>
>> Examining task ID:
>> task_1416831990090_0005_m_05 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_42 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_35 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_65 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_02 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_07 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_58 (and more) from job
>> job_1416831990090_0005
>>
>> Examining task ID:
>> task_1416831990090_0005_m_43 (and more) from job
>> job_1416831990090_0005
>>
>>
>>  Task with the most failures(4):
>>
>> -
>>
>> Task ID:
>>
>>   task_1416831990090_0005_m_05
>>
>>
>>  URL:
>>
>>  
>> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
>>
>> -
>>
>> Diagnostic Messages for this Task:
>>
>> Container launch failed for
>> container_1416831990090_0005_01_000112 :
>> java.lang.IllegalArgumentException: java.net.UnknownHostException:
>> slave6
>>
>>  at
>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>>
>>  at
>> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>>
>>  at
>> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>>
>>  at
>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>>
>>  at
>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>>
>>  at
>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)
>>
>>  at
>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl.getCMProxy(ContainerLauncherImpl.java:403)
>>
>>  at
>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.launch(ContainerLauncherImpl.java:138)
>>
>>  at
>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:369)
>>
>>  at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>
>> 

Re: Container launch failed Error

2014-11-24 Thread Daniel Haviv
did you copy the hosts file to all the nodes?

Daniel

> On 24 בנוב׳ 2014, at 19:04, Amit Behera  wrote:
> 
> hi Daniel,
> 
> 
> this stacktrace same for other query .
> for different run I am getting slave7 sometime slave8... 
> 
> And also I registered all machine IPs in /etc/hosts 
> 
> Regards
> Amit
> 
> 
> 
>> On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv 
>>  wrote:
>> It seems that the application master can't resolve slave6's name to an IP
>> 
>> Daniel
>> 
>>> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
>>> 
>>> Hi Users,
>>> 
>>> my cluster(1+8) configuration:
>>> 
>>> RAM  : 32 GB each
>>> HDFS : 1.5 TB SSD
>>> CPU   : 8 core each
>>> 
>>> ---
>>> 
>>> I am trying to query on 300GB of table but I am able to run only select 
>>> query.
>>> 
>>> Except select query , for all other query I am getting following exception.
>>> 
>>> 
>>> 
>>> 
>>> 
>>> Total jobs = 1
>>> Stage-1 is selected by condition resolver.
>>> Launching Job 1 out of 1
>>> Number of reduce tasks not specified. Estimated
>>> from input data size: 183
>>> In order to change the average load for a
>>> reducer (in bytes):
>>>   set
>>> hive.exec.reducers.bytes.per.reducer=
>>> In order to limit the maximum number of
>>> reducers:
>>>   set hive.exec.reducers.max=
>>> In order to set a constant number of reducers:
>>>   set mapreduce.job.reduces=
>>> Starting Job = job_1416831990090_0005, Tracking
>>> URL = http://master:8088/proxy/application_1416831990090_0005/
>>> Kill Command = /root/hadoop/bin/hadoop job 
>>> -kill job_1416831990090_0005
>>> Hadoop job information for Stage-1: number of
>>> mappers: 679; number of reducers: 183
>>> 2014-11-24 19:43:01,523 Stage-1 map = 0%, 
>>> reduce = 0%
>>> 2014-11-24 19:43:22,730 Stage-1 map = 53%, 
>>> reduce = 0%, Cumulative CPU 625.19 sec
>>> 2014-11-24 19:43:23,778 Stage-1 map = 100%, 
>>> reduce = 100%
>>> MapReduce Total cumulative CPU time: 10 minutes
>>> 25 seconds 190 msec
>>> Ended Job = job_1416831990090_0005 with errors
>>> Error during job, obtaining debugging
>>> information...
>>> Examining task ID:
>>> task_1416831990090_0005_m_05 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_42 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_35 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_65 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_02 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_07 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_58 (and more) from job
>>> job_1416831990090_0005
>>> Examining task ID:
>>> task_1416831990090_0005_m_43 (and more) from job
>>> job_1416831990090_0005
>>> 
>>> 
>>> Task with the most failures(4): 
>>> -
>>> Task ID:
>>>   task_1416831990090_0005_m_05
>>> 
>>> 
>>> URL:
>>>  
>>> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
>>> -
>>> Diagnostic Messages for this Task:
>>> Container launch failed for
>>> container_1416831990090_0005_01_000112 :
>>> java.lang.IllegalArgumentException: java.net.UnknownHostException:
>>> slave6
>>> at
>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>>> at
>>> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>>> at
>>> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>>> at
>>> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)
>>> at
>>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl.getCMProxy(ContainerLauncherImpl.java:403)
>>> at
>>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.launch(ContainerLauncherImpl.java:138)
>>> at
>>> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:369)
>>> at
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>> at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>> at java.lang.Thread.run(Thread.java:745)
>>> Caused by: java.net.UnknownHostException: slave6
>>> ... 12 more
>>> 
>>> 
>>> 
>>> 
>>> FAILED: Execution Error, return code 2 from
>>> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>>> MapReduce J

Re: Container launch failed Error

2014-11-24 Thread Amit Behera
hi Daniel,


this stacktrace same for other query .
for different run I am getting slave7 sometime slave8...

And also I registered all machine IPs in /etc/hosts

Regards
Amit



On Mon, Nov 24, 2014 at 10:22 PM, Daniel Haviv <
daniel.ha...@veracity-group.com> wrote:

> It seems that the application master can't resolve slave6's name to an IP
>
> Daniel
>
> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
>
> Hi Users,
>
> *my cluster(1+8) configuration*:
>
> RAM  : 32 GB each
> HDFS : 1.5 TB SSD
> CPU   : 8 core each
>
> ---
>
> I am trying to query on 300GB of table but I am able to run only select
> query.
>
> Except select query , for all other query I am getting following exception.
>
>
>
>
>
> Total jobs = 1
>
> Stage-1 is selected by condition resolver.
>
> Launching Job 1 out of 1
>
> Number of reduce tasks not specified. Estimated
> from input data size: 183
>
> In order to change the average load for a
> reducer (in bytes):
>
>   set
> hive.exec.reducers.bytes.per.reducer=
>
> In order to limit the maximum number of
> reducers:
>
>   set hive.exec.reducers.max=
>
> In order to set a constant number of reducers:
>
>   set mapreduce.job.reduces=
>
> Starting Job = job_1416831990090_0005, Tracking
> URL = http://master:8088/proxy/application_1416831990090_0005/
>
> Kill Command = /root/hadoop/bin/hadoop job
> -kill job_1416831990090_0005
>
> Hadoop job information for Stage-1: number of
> mappers: 679; number of reducers: 183
>
> 2014-11-24 19:43:01,523 Stage-1 map = 0%,
> reduce = 0%
>
> 2014-11-24 19:43:22,730 Stage-1 map = 53%,
> reduce = 0%, Cumulative CPU 625.19 sec
>
> 2014-11-24 19:43:23,778 Stage-1 map = 100%,
> reduce = 100%
>
> MapReduce Total cumulative CPU time: 10 minutes
> 25 seconds 190 msec
>
> Ended Job = job_1416831990090_0005 with errors
>
> Error during job, obtaining debugging
> information...
>
> Examining task ID:
> task_1416831990090_0005_m_05 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_42 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_35 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_65 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_02 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_07 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_58 (and more) from job
> job_1416831990090_0005
>
> Examining task ID:
> task_1416831990090_0005_m_43 (and more) from job
> job_1416831990090_0005
>
>
>  Task with the most failures(4):
>
> -
>
> Task ID:
>
>   task_1416831990090_0005_m_05
>
>
>  URL:
>
>  
> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
>
> -
>
> Diagnostic Messages for this Task:
>
> Container launch failed for
> container_1416831990090_0005_01_000112 :
> java.lang.IllegalArgumentException: java.net.UnknownHostException:
> slave6
>
>   at
> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>
>   at
> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>
>   at
> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)
>
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl.getCMProxy(ContainerLauncherImpl.java:403)
>
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.launch(ContainerLauncherImpl.java:138)
>
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:369)
>
>   at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>
>   at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>
>   at java.lang.Thread.run(Thread.java:745)
>
> Caused by: java.net.UnknownHostException: slave6
>
>   ... 12 more
>
>
>
>  FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>
> MapReduce Jobs Launched:
>
> Job 0: Map: 679  Reduce: 183   Cumulative CPU:
> 625.19 sec   HDFS Read: 0 HDFS Write: 0 FAIL
>
> Total MapReduce CPU Time Spent: 10 minutes 25
> seconds 190 mse
>
>
>
>
> Please help me to fix the issue.
>
> Thanks
> Amit
>
>


Re: Container launch failed Error

2014-11-24 Thread Daniel Haviv
It seems that the application master can't resolve slave6's name to an IP

Daniel

> On 24 בנוב׳ 2014, at 18:49, Amit Behera  wrote:
> 
> Hi Users,
> 
> my cluster(1+8) configuration:
> 
> RAM  : 32 GB each
> HDFS : 1.5 TB SSD
> CPU   : 8 core each
> 
> ---
> 
> I am trying to query on 300GB of table but I am able to run only select query.
> 
> Except select query , for all other query I am getting following exception.
> 
> 
> 
> 
> 
> Total jobs = 1
> Stage-1 is selected by condition resolver.
> Launching Job 1 out of 1
> Number of reduce tasks not specified. Estimated
> from input data size: 183
> In order to change the average load for a
> reducer (in bytes):
>   set
> hive.exec.reducers.bytes.per.reducer=
> In order to limit the maximum number of
> reducers:
>   set hive.exec.reducers.max=
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=
> Starting Job = job_1416831990090_0005, Tracking
> URL = http://master:8088/proxy/application_1416831990090_0005/
> Kill Command = /root/hadoop/bin/hadoop job 
> -kill job_1416831990090_0005
> Hadoop job information for Stage-1: number of
> mappers: 679; number of reducers: 183
> 2014-11-24 19:43:01,523 Stage-1 map = 0%, 
> reduce = 0%
> 2014-11-24 19:43:22,730 Stage-1 map = 53%, 
> reduce = 0%, Cumulative CPU 625.19 sec
> 2014-11-24 19:43:23,778 Stage-1 map = 100%, 
> reduce = 100%
> MapReduce Total cumulative CPU time: 10 minutes
> 25 seconds 190 msec
> Ended Job = job_1416831990090_0005 with errors
> Error during job, obtaining debugging
> information...
> Examining task ID:
> task_1416831990090_0005_m_05 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_42 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_35 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_65 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_02 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_07 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_58 (and more) from job
> job_1416831990090_0005
> Examining task ID:
> task_1416831990090_0005_m_43 (and more) from job
> job_1416831990090_0005
> 
> 
> Task with the most failures(4): 
> -
> Task ID:
>   task_1416831990090_0005_m_05
> 
> 
> URL:
>  
> http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05
> -
> Diagnostic Messages for this Task:
> Container launch failed for
> container_1416831990090_0005_01_000112 :
> java.lang.IllegalArgumentException: java.net.UnknownHostException:
> slave6
>   at
> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)
>   at
> org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)
>   at
> org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)
>   at
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl.getCMProxy(ContainerLauncherImpl.java:403)
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.launch(ContainerLauncherImpl.java:138)
>   at
> org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:369)
>   at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>   at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>   at java.lang.Thread.run(Thread.java:745)
> Caused by: java.net.UnknownHostException: slave6
>   ... 12 more
> 
> 
> 
> 
> FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> MapReduce Jobs Launched: 
> Job 0: Map: 679  Reduce: 183   Cumulative CPU:
> 625.19 sec   HDFS Read: 0 HDFS Write: 0 FAIL
> Total MapReduce CPU Time Spent: 10 minutes 25
> seconds 190 mse
>
> 
> 
> Please help me to fix the issue.
> 
> Thanks
> Amit


Container launch failed Error

2014-11-24 Thread Amit Behera
Hi Users,

*my cluster(1+8) configuration*:

RAM  : 32 GB each
HDFS : 1.5 TB SSD
CPU   : 8 core each

---

I am trying to query on 300GB of table but I am able to run only select
query.

Except select query , for all other query I am getting following exception.





Total jobs = 1

Stage-1 is selected by condition resolver.

Launching Job 1 out of 1

Number of reduce tasks not specified. Estimated
from input data size: 183

In order to change the average load for a
reducer (in bytes):

  set
hive.exec.reducers.bytes.per.reducer=

In order to limit the maximum number of
reducers:

  set hive.exec.reducers.max=

In order to set a constant number of reducers:

  set mapreduce.job.reduces=

Starting Job = job_1416831990090_0005, Tracking
URL = http://master:8088/proxy/application_1416831990090_0005/

Kill Command = /root/hadoop/bin/hadoop job
-kill job_1416831990090_0005

Hadoop job information for Stage-1: number of
mappers: 679; number of reducers: 183

2014-11-24 19:43:01,523 Stage-1 map = 0%,
reduce = 0%

2014-11-24 19:43:22,730 Stage-1 map = 53%,
reduce = 0%, Cumulative CPU 625.19 sec

2014-11-24 19:43:23,778 Stage-1 map = 100%,
reduce = 100%

MapReduce Total cumulative CPU time: 10 minutes
25 seconds 190 msec

Ended Job = job_1416831990090_0005 with errors

Error during job, obtaining debugging
information...

Examining task ID:
task_1416831990090_0005_m_05 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_42 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_35 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_65 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_02 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_07 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_58 (and more) from job
job_1416831990090_0005

Examining task ID:
task_1416831990090_0005_m_43 (and more) from job
job_1416831990090_0005


 Task with the most failures(4):

-

Task ID:

  task_1416831990090_0005_m_05


 URL:

 
http://master:8088/taskdetails.jsp?jobid=job_1416831990090_0005&tipid=task_1416831990090_0005_m_05

-

Diagnostic Messages for this Task:

Container launch failed for
container_1416831990090_0005_01_000112 :
java.lang.IllegalArgumentException: java.net.UnknownHostException:
slave6

at
org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:418)

at
org.apache.hadoop.security.SecurityUtil.setTokenService(SecurityUtil.java:397)

at
org.apache.hadoop.yarn.util.ConverterUtils.convertFromYarn(ConverterUtils.java:233)

at
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.newProxy(ContainerManagementProtocolProxy.java:211)

at
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy$ContainerManagementProtocolProxyData.(ContainerManagementProtocolProxy.java:189)

at
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy.getProxy(ContainerManagementProtocolProxy.java:110)

at
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl.getCMProxy(ContainerLauncherImpl.java:403)

at
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.launch(ContainerLauncherImpl.java:138)

at
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:369)

at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)

at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)

at java.lang.Thread.run(Thread.java:745)

Caused by: java.net.UnknownHostException: slave6

... 12 more



 FAILED: Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapRedTask

MapReduce Jobs Launched:

Job 0: Map: 679  Reduce: 183   Cumulative CPU:
625.19 sec   HDFS Read: 0 HDFS Write: 0 FAIL

Total MapReduce CPU Time Spent: 10 minutes 25
seconds 190 mse




Please help me to fix the issue.

Thanks
Amit


Re: UPDATE in Hive -0.14.0

2014-11-24 Thread Mahesh Kumar
hi unmesha sreeveni,

  As i told earlier create the table with ACID ouput format
support.



On Mon, Nov 24, 2014 at 3:09 PM, unmesha sreeveni 
wrote:

> Created a Table in Hive
>
> create external table HiveTest (EmployeeID Int,FirstName
> String,Designation String,Salary Int,Department String) row format
> delimited fields terminated by "," location '/user/aibladmin/Hive';
>
> And set all the properties in hive-site.xml
> hive.support.concurrency – true
> hive.enforce.bucketing – true
> hive.exec.dynamic.partition.mode – nonstrict
> hive.txn.manager –org.apache.hadoop.hive.ql.lockmgr.DbTxnManager
> hive.compactor.initiator.on – true
> hive.compactor.worker.threads – 1
>
> The when I tried
> hive>
> >
> > UPDATE HiveTest SET salary = 5 WHERE employeeid = 19;
> FAILED: SemanticException [Error 10297]: Attempt to do update or delete on
> table default.HiveTest that does not use an AcidOutputFormat or is not
> bucketed
>
>  Is it beacause of hive.enforce.bucketing – true set in hive-site.xml
> Is bucketing like partition?
>
>
>
>
>
> On Mon, Nov 24, 2014 at 2:55 PM, Mahesh Kumar 
> wrote:
>
>> Hi unmesha sreevani,
>>
>> *Create metastore in mysql and create the tables as
>> per the below link.*
>> https://github.com/apache/hive/blob/trunk/metastore/scripts/upCreate
>> metastore in mysql and create the tables as per the below
>> link.grade/mysql/hive-schema-0.14.0.mysql.sql
>> 
>> .
>>
>> *And add these properties in hive-site.xml.*
>>
>> 
>>  hive.support.concurrency
>>  true
>> 
>>
>> 
>>  hive.enforce.bucketing
>>  true
>> 
>>
>> 
>>  hive.exec.dynamic.partition.mode
>>  nonstrict
>> 
>>
>> 
>>  hive.txn.manager
>>  org.apache.hadoop.hive.ql.lockmgr.DbTxnManager
>> 
>>
>> 
>>  hive.compactor.initiator.on
>>  true
>> 
>>
>> 
>>  hive.compactor.worker.threads
>>  1
>> 
>>
>> *Make sure your table creation supports ACID ouput format.Create like
>> following*.
>>
>> create table test(id int, name varchar(128)) clustered by (id) into 2
>> buckets stored as orc TBLPROPERTIES ('transactional'='true')
>>
>>
>> Regards,
>>
>> Mahesh.S
>>
>>
>
>
> --
> *Thanks & Regards *
>
>
> *Unmesha Sreeveni U.B*
> *Hadoop, Bigdata Developer*
> *Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
> http://www.unmeshasreeveni.blogspot.in/
>
>
>


Re: UPDATE in Hive -0.14.0

2014-11-24 Thread unmesha sreeveni
Created a Table in Hive

create external table HiveTest (EmployeeID Int,FirstName String,Designation
String,Salary Int,Department String) row format delimited fields terminated
by "," location '/user/aibladmin/Hive';

And set all the properties in hive-site.xml
hive.support.concurrency – true
hive.enforce.bucketing – true
hive.exec.dynamic.partition.mode – nonstrict
hive.txn.manager –org.apache.hadoop.hive.ql.lockmgr.DbTxnManager
hive.compactor.initiator.on – true
hive.compactor.worker.threads – 1

The when I tried
hive>
>
> UPDATE HiveTest SET salary = 5 WHERE employeeid = 19;
FAILED: SemanticException [Error 10297]: Attempt to do update or delete on
table default.HiveTest that does not use an AcidOutputFormat or is not
bucketed

 Is it beacause of hive.enforce.bucketing – true set in hive-site.xml
Is bucketing like partition?





On Mon, Nov 24, 2014 at 2:55 PM, Mahesh Kumar 
wrote:

> Hi unmesha sreevani,
>
> *Create metastore in mysql and create the tables as
> per the below link.*
> https://github.com/apache/hive/blob/trunk/metastore/scripts/upCreate
> metastore in mysql and create the tables as per the below
> link.grade/mysql/hive-schema-0.14.0.mysql.sql
> 
> .
>
> *And add these properties in hive-site.xml.*
>
> 
>  hive.support.concurrency
>  true
> 
>
> 
>  hive.enforce.bucketing
>  true
> 
>
> 
>  hive.exec.dynamic.partition.mode
>  nonstrict
> 
>
> 
>  hive.txn.manager
>  org.apache.hadoop.hive.ql.lockmgr.DbTxnManager
> 
>
> 
>  hive.compactor.initiator.on
>  true
> 
>
> 
>  hive.compactor.worker.threads
>  1
> 
>
> *Make sure your table creation supports ACID ouput format.Create like
> following*.
>
> create table test(id int, name varchar(128)) clustered by (id) into 2
> buckets stored as orc TBLPROPERTIES ('transactional'='true')
>
>
> Regards,
>
> Mahesh.S
>
>


-- 
*Thanks & Regards *


*Unmesha Sreeveni U.B*
*Hadoop, Bigdata Developer*
*Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
http://www.unmeshasreeveni.blogspot.in/


Re: UPDATE in Hive -0.14.0

2014-11-24 Thread Mahesh Kumar
Hi unmesha sreevani,

*Create metastore in mysql and create the tables as per
the below link.*
https://github.com/apache/hive/blob/trunk/metastore/scripts/upCreate
metastore in mysql and create the tables as per the below
link.grade/mysql/hive-schema-0.14.0.mysql.sql

.

*And add these properties in hive-site.xml.*


 hive.support.concurrency
 true



 hive.enforce.bucketing
 true



 hive.exec.dynamic.partition.mode
 nonstrict



 hive.txn.manager
 org.apache.hadoop.hive.ql.lockmgr.DbTxnManager



 hive.compactor.initiator.on
 true



 hive.compactor.worker.threads
 1


*Make sure your table creation supports ACID ouput format.Create like
following*.

create table test(id int, name varchar(128)) clustered by (id) into 2
buckets stored as orc TBLPROPERTIES ('transactional'='true')


Regards,

Mahesh.S


Re: UPDATE in Hive -0.14.0

2014-11-24 Thread @Sanjiv Singh
Make sure you have configured these properties hive-site.xml

hive.support.concurrency
hive.enforce.bucketing
hive.exec.dynamic.partition.mode
hive.txn.manager
hive.compactor.initiator.on
hive.compactor.worker.threads


also following property has been set at table level
TBLPROPERTIES('transactional'='true')

Regards
Sanjiv Singh
Mob :  +091 9990-447-339

On Mon, Nov 24, 2014 at 1:58 PM, Nitin Pawar 
wrote:

> whats your create table DDL?
> On 24 Nov 2014 13:43, "unmesha sreeveni"  wrote:
>
>> Hi
>>
>> I am using hive -0.14.0 which support UPDATE statement
>>
>> but I am getting an error once I did this Command
>> UPDATE Emp SET salary = 5 WHERE employeeid = 19;
>>
>> FAILED: SemanticException [Error 10294]: Attempt to do update or delete
>> using transaction manager that does not support these operations.
>> hive>
>>
>>
>> Am I doing anything wrong?
>>
>> --
>> *Thanks & Regards *
>>
>>
>> *Unmesha Sreeveni U.B*
>> *Hadoop, Bigdata Developer*
>> *Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
>> http://www.unmeshasreeveni.blogspot.in/
>>
>>
>>


Re: UPDATE in Hive -0.14.0

2014-11-24 Thread Nitin Pawar
whats your create table DDL?
On 24 Nov 2014 13:43, "unmesha sreeveni"  wrote:

> Hi
>
> I am using hive -0.14.0 which support UPDATE statement
>
> but I am getting an error once I did this Command
> UPDATE Emp SET salary = 5 WHERE employeeid = 19;
>
> FAILED: SemanticException [Error 10294]: Attempt to do update or delete
> using transaction manager that does not support these operations.
> hive>
>
>
> Am I doing anything wrong?
>
> --
> *Thanks & Regards *
>
>
> *Unmesha Sreeveni U.B*
> *Hadoop, Bigdata Developer*
> *Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
> http://www.unmeshasreeveni.blogspot.in/
>
>
>


UPDATE in Hive -0.14.0

2014-11-24 Thread unmesha sreeveni
Hi

I am using hive -0.14.0 which support UPDATE statement

but I am getting an error once I did this Command
UPDATE Emp SET salary = 5 WHERE employeeid = 19;

FAILED: SemanticException [Error 10294]: Attempt to do update or delete
using transaction manager that does not support these operations.
hive>


Am I doing anything wrong?

-- 
*Thanks & Regards *


*Unmesha Sreeveni U.B*
*Hadoop, Bigdata Developer*
*Centre for Cyber Security | Amrita Vishwa Vidyapeetham*
http://www.unmeshasreeveni.blogspot.in/