yes, I did configured.

On Aug 19, 2016 7:22 PM, "Rohith Sharma K S" <ksrohithsha...@gmail.com>
wrote:

> Hi
>
> From below discussion and AM logs, I see that AM container has launched
> but not able to connect to RM.
>
> This looks like your configuration issue. Would you check your job.xml jar
> that does *yarn.resourcemanager.scheduler.address *has been configured?
>
> Essentially, this address required by MRAppMaster for connecting to RM for
> heartbeats. If you don’t not configure, default value will be taken i.e
> 8030.
>
>
> Thanks & Regards
> Rohith Sharma K S
>
> On Aug 20, 2016, at 7:02 AM, rammohan ganapavarapu <
> rammohanga...@gmail.com> wrote:
>
> Even if  the cluster dont have enough resources it should connect to "
>
> /0.0.0.0:8030" right? it should connect to my <RM_HOST:8030>, not sure why 
> its trying to connect to 0.0.0.0:8030.
>
> I have verified the config and i removed traces of 0.0.0.0 still no luck.
>
> org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager at 
> /0.0.0.0:8030
>
> If an one has any clue please share.
>
> Thanks,
>
> Ram
>
>
>
> On Fri, Aug 19, 2016 at 2:32 PM, rammohan ganapavarapu <
> rammohanga...@gmail.com> wrote:
>
>> When i submit a job using yarn its seems working only with oozie its
>> failing i guess, not sure what is missing.
>>
>> yarn jar 
>> /uap/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.1.jar
>> pi 20 1000
>> Number of Maps  = 20
>> Samples per Map = 1000
>> .
>> .
>> .
>> Job Finished in 19.622 seconds
>> Estimated value of Pi is 3.14280000000000000000
>>
>> Ram
>>
>> On Fri, Aug 19, 2016 at 11:46 AM, rammohan ganapavarapu <
>> rammohanga...@gmail.com> wrote:
>>
>>> Ok, i have used yarn-utils.py to get the correct values for my cluster
>>> and update those properties and restarted RM and NM but still no luck not
>>> sure what i am missing, any other insights will help me.
>>>
>>> Below are my properties from yarn-site.xml and map-site.xml.
>>>
>>> python yarn-utils.py -c 24 -m 63 -d 3 -k False
>>>  Using cores=24 memory=63GB disks=3 hbase=False
>>>  Profile: cores=24 memory=63488MB reserved=1GB usableMem=62GB disks=3
>>>  Num Container=6
>>>  Container Ram=10240MB
>>>  Used Ram=60GB
>>>  Unused Ram=1GB
>>>  yarn.scheduler.minimum-allocation-mb=10240
>>>  yarn.scheduler.maximum-allocation-mb=61440
>>>  yarn.nodemanager.resource.memory-mb=61440
>>>  mapreduce.map.memory.mb=5120
>>>  mapreduce.map.java.opts=-Xmx4096m
>>>  mapreduce.reduce.memory.mb=10240
>>>  mapreduce.reduce.java.opts=-Xmx8192m
>>>  yarn.app.mapreduce.am.resource.mb=5120
>>>  yarn.app.mapreduce.am.command-opts=-Xmx4096m
>>>  mapreduce.task.io.sort.mb=1024
>>>
>>>
>>>     <property>
>>>       <name>mapreduce.map.memory.mb</name>
>>>       <value>5120</value>
>>>     </property>
>>>     <property>
>>>       <name>mapreduce.map.java.opts</name>
>>>       <value>-Xmx4096m</value>
>>>     </property>
>>>     <property>
>>>       <name>mapreduce.reduce.memory.mb</name>
>>>       <value>10240</value>
>>>     </property>
>>>     <property>
>>>       <name>mapreduce.reduce.java.opts</name>
>>>       <value>-Xmx8192m</value>
>>>     </property>
>>>     <property>
>>>       <name>yarn.app.mapreduce.am.resource.mb</name>
>>>       <value>5120</value>
>>>     </property>
>>>     <property>
>>>       <name>yarn.app.mapreduce.am.command-opts</name>
>>>       <value>-Xmx4096m</value>
>>>     </property>
>>>     <property>
>>>       <name>mapreduce.task.io.sort.mb</name>
>>>       <value>1024</value>
>>>     </property>
>>>
>>>
>>>
>>>      <property>
>>>       <name>yarn.scheduler.minimum-allocation-mb</name>
>>>       <value>10240</value>
>>>     </property>
>>>
>>>      <property>
>>>       <name>yarn.scheduler.maximum-allocation-mb</name>
>>>       <value>61440</value>
>>>     </property>
>>>
>>>      <property>
>>>       <name>yarn.nodemanager.resource.memory-mb</name>
>>>       <value>61440</value>
>>>     </property>
>>>
>>>
>>> Ram
>>>
>>> On Thu, Aug 18, 2016 at 11:14 PM, tkg_cangkul <yuza.ras...@gmail.com>
>>> wrote:
>>>
>>>> maybe this link can be some reference to tune up the cluster:
>>>>
>>>> http://jason4zhu.blogspot.co.id/2014/10/memory-configuration
>>>> -in-hadoop.html
>>>>
>>>>
>>>> On 19/08/16 11:13, rammohan ganapavarapu wrote:
>>>>
>>>> Do you know what properties to tune?
>>>>
>>>> Thanks,
>>>> Ram
>>>>
>>>> On Thu, Aug 18, 2016 at 9:11 PM, tkg_cangkul <yuza.ras...@gmail.com>
>>>> wrote:
>>>>
>>>>> i think that's because you don't have enough resource.  u can tune
>>>>> your cluster config to maximize your resource.
>>>>>
>>>>>
>>>>> On 19/08/16 11:03, rammohan ganapavarapu wrote:
>>>>>
>>>>> I dont see any thing odd except this not sure if i have to worry about
>>>>> it or not.
>>>>>
>>>>> 2016-08-19 03:29:26,621 INFO [main] org.apache.hadoop.yarn.client.RMProxy:
>>>>> Connecting to ResourceManager at /0.0.0.0:8030
>>>>> 2016-08-19 03:29:27,646 INFO [main] org.apache.hadoop.ipc.Client:
>>>>> Retrying connect to server: 0.0.0.0/0.0.0.0:8030. Already tried 0
>>>>> time(s); retry policy is RetryUpToMaximumCo
>>>>> untWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
>>>>> 2016-08-19 03:29:28,647 INFO [main] org.apache.hadoop.ipc.Client:
>>>>> Retrying connect to server: 0.0.0.0/0.0.0.0:8030. Already tried 1
>>>>> time(s); retry policy is 
>>>>> RetryUpToMaximumCountWithFixedSleep(maxRetries=10,
>>>>> sleepTime=1000 MILLISECONDS)
>>>>>
>>>>>
>>>>> its keep printing this log ..in app container logs.
>>>>>
>>>>> On Thu, Aug 18, 2016 at 8:20 PM, tkg_cangkul <yuza.ras...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> maybe u can check the logs from port 8088 on your browser. that was
>>>>>> RM UI. just choose your job id and then check the logs.
>>>>>>
>>>>>> On 19/08/16 10:14, rammohan ganapavarapu wrote:
>>>>>>
>>>>>> Sunil,
>>>>>>
>>>>>> Thanks you for your input, below are my server metrics for RM. Also
>>>>>> attached RM UI for capacity scheduler resources. How else i can find?
>>>>>>
>>>>>> {
>>>>>>       "name": "Hadoop:service=ResourceManage
>>>>>> r,name=QueueMetrics,q0=root",
>>>>>>       "modelerType": "QueueMetrics,q0=root",
>>>>>>       "tag.Queue": "root",
>>>>>>       "tag.Context": "yarn",
>>>>>>       "tag.Hostname": "hadoop001",
>>>>>>       "running_0": 0,
>>>>>>       "running_60": 0,
>>>>>>       "running_300": 0,
>>>>>>       "running_1440": 0,
>>>>>>       "AppsSubmitted": 1,
>>>>>>       "AppsRunning": 0,
>>>>>>       "AppsPending": 0,
>>>>>>       "AppsCompleted": 0,
>>>>>>       "AppsKilled": 0,
>>>>>>       "AppsFailed": 1,
>>>>>>       "AllocatedMB": 0,
>>>>>>       "AllocatedVCores": 0,
>>>>>>       "AllocatedContainers": 0,
>>>>>>       "AggregateContainersAllocated": 2,
>>>>>>       "AggregateContainersReleased": 2,
>>>>>>       "AvailableMB": 64512,
>>>>>>       "AvailableVCores": 24,
>>>>>>       "PendingMB": 0,
>>>>>>       "PendingVCores": 0,
>>>>>>       "PendingContainers": 0,
>>>>>>       "ReservedMB": 0,
>>>>>>       "ReservedVCores": 0,
>>>>>>       "ReservedContainers": 0,
>>>>>>       "ActiveUsers": 0,
>>>>>>       "ActiveApplications": 0
>>>>>>     },
>>>>>>
>>>>>> On Thu, Aug 18, 2016 at 6:49 PM, Sunil Govind <sunil.gov...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> Hi
>>>>>>>
>>>>>>> It could be because of many of reasons. Also I am not sure about
>>>>>>> which scheduler your are using, pls share more details such as RM log 
>>>>>>> etc.
>>>>>>>
>>>>>>> I could point out few reasons
>>>>>>>  - Such as "Not enough resource is cluster" can cause this
>>>>>>>  - If using Capacity Scheduler, if queue capacity is maxed out, such
>>>>>>> case can happen.
>>>>>>>  - Similarly if max-am-resource-percent is crossed per queue level,
>>>>>>> then also AM container may not be launched.
>>>>>>>
>>>>>>> you could check RM log to get more information if AM container is
>>>>>>> laucnhed.
>>>>>>>
>>>>>>> Thanks
>>>>>>> Sunil
>>>>>>>
>>>>>>> On Fri, Aug 19, 2016 at 5:37 AM rammohan ganapavarapu <
>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> When i submit a MR job, i am getting this from AM UI but it never
>>>>>>>> get finished, what am i missing ?
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Ram
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> ---------------------------------------------------------------------
>>>>>> To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org
>>>>>> For additional commands, e-mail: user-h...@hadoop.apache.org
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>
>

Reply via email to