any thoughts from the logs and config I have shared?

On Aug 21, 2016 8:32 AM, "rammohan ganapavarapu" <rammohanga...@gmail.com>
wrote:

> so in job.properties what is the jobtracker property, is it RM ip: port or
> scheduler port which is 8030, if I use 8030 I am getting unknown protocol
> proto buffer error.
>
> On Aug 21, 2016 7:37 AM, "Sunil Govind" <sunil.gov...@gmail.com> wrote:
>
>> Hi.
>>
>> It seems its an oozie issue. From conf, RM scheduler is running at port
>> 8030.
>> But your job.properties is taking 8032. I suggest you could double
>> confirm your oozie configuration and see the configurations are intact to
>> contact RM. Sharing a link also
>> https://discuss.zendesk.com/hc/en-us/articles/203355837-How-
>> to-run-a-MapReduce-jar-using-Oozie-workflow
>>
>> Thanks
>> Sunil
>>
>>
>> On Sun, Aug 21, 2016 at 8:41 AM rammohan ganapavarapu <
>> rammohanga...@gmail.com> wrote:
>>
>>> Please find the attached config that i got from yarn ui and  AM,RM logs.
>>> I only see that connecting to 0.0.0.0:8030 when i submit job using
>>> oozie, but if i submit as yarn jar its working fine as i posted in my
>>> previous posts.
>>>
>>> Here is my oozie job.properties file, i have a java class that just
>>> prints
>>>
>>> nameNode=hdfs://master01:8020
>>> jobTracker=master01:8032
>>> workflowName=EchoJavaJob
>>> oozie.use.system.libpath=true
>>>
>>> queueName=default
>>> hdfsWorkflowHome=/user/uap/oozieWorkflows
>>>
>>> workflowPath=${nameNode}${hdfsWorkflowHome}/${workflowName}
>>> oozie.wf.application.path=${workflowPath}
>>>
>>> Please let me know if you guys find any clue why its trying to connect
>>> to 0.0.0.:8030.
>>>
>>> Thanks,
>>> Ram
>>>
>>>
>>> On Fri, Aug 19, 2016 at 11:54 PM, Sunil Govind <sunil.gov...@gmail.com>
>>> wrote:
>>>
>>>> Hi Ram
>>>>
>>>> From the console log, as Rohith said, AM is looking for AM at 8030. So
>>>> pls confirm the RM port once.
>>>> Could you please share AM and RM logs.
>>>>
>>>> Thanks
>>>> Sunil
>>>>
>>>> On Sat, Aug 20, 2016 at 10:36 AM rammohan ganapavarapu <
>>>> rammohanga...@gmail.com> wrote:
>>>>
>>>>> yes, I did configured.
>>>>>
>>>>> On Aug 19, 2016 7:22 PM, "Rohith Sharma K S" <ksrohithsha...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi
>>>>>>
>>>>>> From below discussion and AM logs, I see that AM container has
>>>>>> launched but not able to connect to RM.
>>>>>>
>>>>>> This looks like your configuration issue. Would you check your
>>>>>> job.xml jar that does *yarn.resourcemanager.scheduler.address *has
>>>>>> been configured?
>>>>>>
>>>>>> Essentially, this address required by MRAppMaster for connecting to
>>>>>> RM for heartbeats. If you don’t not configure, default value will be 
>>>>>> taken
>>>>>> i.e 8030.
>>>>>>
>>>>>>
>>>>>> Thanks & Regards
>>>>>> Rohith Sharma K S
>>>>>>
>>>>>> On Aug 20, 2016, at 7:02 AM, rammohan ganapavarapu <
>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>
>>>>>> Even if  the cluster dont have enough resources it should connect to "
>>>>>>
>>>>>> /0.0.0.0:8030" right? it should connect to my <RM_HOST:8030>, not sure 
>>>>>> why its trying to connect to 0.0.0.0:8030.
>>>>>>
>>>>>> I have verified the config and i removed traces of 0.0.0.0 still no luck.
>>>>>>
>>>>>> org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at 
>>>>>> /0.0.0.0:8030
>>>>>>
>>>>>> If an one has any clue please share.
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> Ram
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Aug 19, 2016 at 2:32 PM, rammohan ganapavarapu <
>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>
>>>>>>> When i submit a job using yarn its seems working only with oozie its
>>>>>>> failing i guess, not sure what is missing.
>>>>>>>
>>>>>>> yarn jar /uap/hadoop/share/hadoop/mapre
>>>>>>> duce/hadoop-mapreduce-examples-2.7.1.jar pi 20 1000
>>>>>>> Number of Maps  = 20
>>>>>>> Samples per Map = 1000
>>>>>>> .
>>>>>>> .
>>>>>>> .
>>>>>>> Job Finished in 19.622 seconds
>>>>>>> Estimated value of Pi is 3.14280000000000000000
>>>>>>>
>>>>>>> Ram
>>>>>>>
>>>>>>> On Fri, Aug 19, 2016 at 11:46 AM, rammohan ganapavarapu <
>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Ok, i have used yarn-utils.py to get the correct values for my
>>>>>>>> cluster and update those properties and restarted RM and NM but still 
>>>>>>>> no
>>>>>>>> luck not sure what i am missing, any other insights will help me.
>>>>>>>>
>>>>>>>> Below are my properties from yarn-site.xml and map-site.xml.
>>>>>>>>
>>>>>>>> python yarn-utils.py -c 24 -m 63 -d 3 -k False
>>>>>>>>  Using cores=24 memory=63GB disks=3 hbase=False
>>>>>>>>  Profile: cores=24 memory=63488MB reserved=1GB usableMem=62GB
>>>>>>>> disks=3
>>>>>>>>  Num Container=6
>>>>>>>>  Container Ram=10240MB
>>>>>>>>  Used Ram=60GB
>>>>>>>>  Unused Ram=1GB
>>>>>>>>  yarn.scheduler.minimum-allocation-mb=10240
>>>>>>>>  yarn.scheduler.maximum-allocation-mb=61440
>>>>>>>>  yarn.nodemanager.resource.memory-mb=61440
>>>>>>>>  mapreduce.map.memory.mb=5120
>>>>>>>>  mapreduce.map.java.opts=-Xmx4096m
>>>>>>>>  mapreduce.reduce.memory.mb=10240
>>>>>>>>  mapreduce.reduce.java.opts=-Xmx8192m
>>>>>>>>  yarn.app.mapreduce.am.resource.mb=5120
>>>>>>>>  yarn.app.mapreduce.am.command-opts=-Xmx4096m
>>>>>>>>  mapreduce.task.io.sort.mb=1024
>>>>>>>>
>>>>>>>>
>>>>>>>>     <property>
>>>>>>>>       <name>mapreduce.map.memory.mb</name>
>>>>>>>>       <value>5120</value>
>>>>>>>>     </property>
>>>>>>>>     <property>
>>>>>>>>       <name>mapreduce.map.java.opts</name>
>>>>>>>>       <value>-Xmx4096m</value>
>>>>>>>>     </property>
>>>>>>>>     <property>
>>>>>>>>       <name>mapreduce.reduce.memory.mb</name>
>>>>>>>>       <value>10240</value>
>>>>>>>>     </property>
>>>>>>>>     <property>
>>>>>>>>       <name>mapreduce.reduce.java.opts</name>
>>>>>>>>       <value>-Xmx8192m</value>
>>>>>>>>     </property>
>>>>>>>>     <property>
>>>>>>>>       <name>yarn.app.mapreduce.am.resource.mb</name>
>>>>>>>>       <value>5120</value>
>>>>>>>>     </property>
>>>>>>>>     <property>
>>>>>>>>       <name>yarn.app.mapreduce.am.command-opts</name>
>>>>>>>>       <value>-Xmx4096m</value>
>>>>>>>>     </property>
>>>>>>>>     <property>
>>>>>>>>       <name>mapreduce.task.io.sort.mb</name>
>>>>>>>>       <value>1024</value>
>>>>>>>>     </property>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>      <property>
>>>>>>>>       <name>yarn.scheduler.minimum-allocation-mb</name>
>>>>>>>>       <value>10240</value>
>>>>>>>>     </property>
>>>>>>>>
>>>>>>>>      <property>
>>>>>>>>       <name>yarn.scheduler.maximum-allocation-mb</name>
>>>>>>>>       <value>61440</value>
>>>>>>>>     </property>
>>>>>>>>
>>>>>>>>      <property>
>>>>>>>>       <name>yarn.nodemanager.resource.memory-mb</name>
>>>>>>>>       <value>61440</value>
>>>>>>>>     </property>
>>>>>>>>
>>>>>>>>
>>>>>>>> Ram
>>>>>>>>
>>>>>>>> On Thu, Aug 18, 2016 at 11:14 PM, tkg_cangkul <
>>>>>>>> yuza.ras...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> maybe this link can be some reference to tune up the cluster:
>>>>>>>>>
>>>>>>>>> http://jason4zhu.blogspot.co.id/2014/10/memory-configuration
>>>>>>>>> -in-hadoop.html
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On 19/08/16 11:13, rammohan ganapavarapu wrote:
>>>>>>>>>
>>>>>>>>> Do you know what properties to tune?
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Ram
>>>>>>>>>
>>>>>>>>> On Thu, Aug 18, 2016 at 9:11 PM, tkg_cangkul <
>>>>>>>>> yuza.ras...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> i think that's because you don't have enough resource.  u can
>>>>>>>>>> tune your cluster config to maximize your resource.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On 19/08/16 11:03, rammohan ganapavarapu wrote:
>>>>>>>>>>
>>>>>>>>>> I dont see any thing odd except this not sure if i have to worry
>>>>>>>>>> about it or not.
>>>>>>>>>>
>>>>>>>>>> 2016-08-19 03:29:26,621 INFO [main] 
>>>>>>>>>> org.apache.hadoop.yarn.client.RMProxy:
>>>>>>>>>> Connecting to ResourceManager at /0.0.0.0:8030
>>>>>>>>>> 2016-08-19 03:29:27,646 INFO [main] org.apache.hadoop.ipc.Client:
>>>>>>>>>> Retrying connect to server: 0.0.0.0/0.0.0.0:8030. Already tried
>>>>>>>>>> 0 time(s); retry policy is RetryUpToMaximumCo
>>>>>>>>>> untWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
>>>>>>>>>> 2016-08-19 03:29:28,647 INFO [main] org.apache.hadoop.ipc.Client:
>>>>>>>>>> Retrying connect to server: 0.0.0.0/0.0.0.0:8030. Already tried
>>>>>>>>>> 1 time(s); retry policy is 
>>>>>>>>>> RetryUpToMaximumCountWithFixedSleep(maxRetries=10,
>>>>>>>>>> sleepTime=1000 MILLISECONDS)
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> its keep printing this log ..in app container logs.
>>>>>>>>>>
>>>>>>>>>> On Thu, Aug 18, 2016 at 8:20 PM, tkg_cangkul <
>>>>>>>>>> yuza.ras...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> maybe u can check the logs from port 8088 on your browser. that
>>>>>>>>>>> was RM UI. just choose your job id and then check the logs.
>>>>>>>>>>>
>>>>>>>>>>> On 19/08/16 10:14, rammohan ganapavarapu wrote:
>>>>>>>>>>>
>>>>>>>>>>> Sunil,
>>>>>>>>>>>
>>>>>>>>>>> Thanks you for your input, below are my server metrics for RM.
>>>>>>>>>>> Also attached RM UI for capacity scheduler resources. How else i 
>>>>>>>>>>> can find?
>>>>>>>>>>>
>>>>>>>>>>> {
>>>>>>>>>>>       "name": "Hadoop:service=ResourceManage
>>>>>>>>>>> r,name=QueueMetrics,q0=root",
>>>>>>>>>>>       "modelerType": "QueueMetrics,q0=root",
>>>>>>>>>>>       "tag.Queue": "root",
>>>>>>>>>>>       "tag.Context": "yarn",
>>>>>>>>>>>       "tag.Hostname": "hadoop001",
>>>>>>>>>>>       "running_0": 0,
>>>>>>>>>>>       "running_60": 0,
>>>>>>>>>>>       "running_300": 0,
>>>>>>>>>>>       "running_1440": 0,
>>>>>>>>>>>       "AppsSubmitted": 1,
>>>>>>>>>>>       "AppsRunning": 0,
>>>>>>>>>>>       "AppsPending": 0,
>>>>>>>>>>>       "AppsCompleted": 0,
>>>>>>>>>>>       "AppsKilled": 0,
>>>>>>>>>>>       "AppsFailed": 1,
>>>>>>>>>>>       "AllocatedMB": 0,
>>>>>>>>>>>       "AllocatedVCores": 0,
>>>>>>>>>>>       "AllocatedContainers": 0,
>>>>>>>>>>>       "AggregateContainersAllocated": 2,
>>>>>>>>>>>       "AggregateContainersReleased": 2,
>>>>>>>>>>>       "AvailableMB": 64512,
>>>>>>>>>>>       "AvailableVCores": 24,
>>>>>>>>>>>       "PendingMB": 0,
>>>>>>>>>>>       "PendingVCores": 0,
>>>>>>>>>>>       "PendingContainers": 0,
>>>>>>>>>>>       "ReservedMB": 0,
>>>>>>>>>>>       "ReservedVCores": 0,
>>>>>>>>>>>       "ReservedContainers": 0,
>>>>>>>>>>>       "ActiveUsers": 0,
>>>>>>>>>>>       "ActiveApplications": 0
>>>>>>>>>>>     },
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Aug 18, 2016 at 6:49 PM, Sunil Govind <
>>>>>>>>>>> sunil.gov...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi
>>>>>>>>>>>>
>>>>>>>>>>>> It could be because of many of reasons. Also I am not sure
>>>>>>>>>>>> about which scheduler your are using, pls share more details such 
>>>>>>>>>>>> as RM log
>>>>>>>>>>>> etc.
>>>>>>>>>>>>
>>>>>>>>>>>> I could point out few reasons
>>>>>>>>>>>>  - Such as "Not enough resource is cluster" can cause this
>>>>>>>>>>>>  - If using Capacity Scheduler, if queue capacity is maxed out,
>>>>>>>>>>>> such case can happen.
>>>>>>>>>>>>  - Similarly if max-am-resource-percent is crossed per queue
>>>>>>>>>>>> level, then also AM container may not be launched.
>>>>>>>>>>>>
>>>>>>>>>>>> you could check RM log to get more information if AM container
>>>>>>>>>>>> is laucnhed.
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks
>>>>>>>>>>>> Sunil
>>>>>>>>>>>>
>>>>>>>>>>>> On Fri, Aug 19, 2016 at 5:37 AM rammohan ganapavarapu <
>>>>>>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>
>>>>>>>>>>>>> When i submit a MR job, i am getting this from AM UI but it
>>>>>>>>>>>>> never get finished, what am i missing ?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Ram
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>> To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org
>>>>>>>>>>> For additional commands, e-mail: user-h...@hadoop.apache.org
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>

Reply via email to