Guys,

I was able to fix this issue but trail and error not sure which property
made it work :) but its working and i have to use 8032 as jobtracker. I
also restarted all the components before i was only restarting nodemanager
and resource manager after property update.

You get this error if you use wrong port

Socket Reader #1 for port 8030: readAndProcess from client 10.16.3.51 threw
exception [org.apache.hadoop.security.AccessControlException: SIMPLE
authentication is not enabled.  Available:[TOKEN]]
org.apache.hadoop.security.AccessControlException: SIMPLE authentication is
not enabled.  Available:[TOKEN]

Thanks a lot for all your help,

Ram

On Mon, Aug 22, 2016 at 10:27 AM, rammohan ganapavarapu <
rammohanga...@gmail.com> wrote:

> Thank you all, I have updated my oozie job.properties to use 8030 and now
> i am getting below error
>
>
>
>
>
>     2016-08-22 17:22:02,893 INFO org.apache.hadoop.yarn.server.
> resourcemanager.ResourceTrackerService: NodeManager from node
> slave03(cmPort: 40511 httpPort: 8042) registered with capability:
> <memory:8192, vCores:8>, assigned nodeId slave03:40511
> 2016-08-22 17:22:02,893 INFO org.apache.hadoop.yarn.server.
> resourcemanager.rmnode.RMNodeImpl: slave03:40511 Node Transitioned from
> NEW to RUNNING
> 2016-08-22 17:22:02,893 INFO org.apache.hadoop.yarn.server.
> resourcemanager.scheduler.capacity.CapacityScheduler: Added node
> slave03:40511 clusterResource: <memory:24576, vCores:24>
> 2016-08-22 17:23:14,258 INFO org.apache.hadoop.ipc.Server: Socket Reader
> #1 for port 8030: readAndProcess from client 10.16.3.51 threw exception
> [org.apache.hadoop.security.AccessControlException: SIMPLE authentication
> is not enabled.  Available:[TOKEN]]
> org.apache.hadoop.security.AccessControlException: SIMPLE authentication
> is not enabled.  Available:[TOKEN]
>         at org.apache.hadoop.ipc.Server$Connection.
> initializeAuthContext(Server.java:1564)
>         at org.apache.hadoop.ipc.Server$Connection.readAndProcess(
> Server.java:1520)
>         at org.apache.hadoop.ipc.Server$Listener.doRead(Server.java:771)
>         at org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(
> Server.java:637)
>         at org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.
> java:608)
>
>
> So i did enabled simple auth by below config in core-site.xml and
> restarted namenode,datanode,rm and nm but still getting same error do i
> have to do any thing else to enable simple auth?
>
>
>     <property>
>       <name>hadoop.security.authentication</name>
>       <value>simple</value>
>     </property>
>
>
> Ram
>
> On Mon, Aug 22, 2016 at 9:43 AM, Sunil Govind <sunil.gov...@gmail.com>
> wrote:
>
>> HI Ram
>>
>> RM logs looks fine and as per config it looks like RM is running on 8030
>> itself.
>> I am not very sure about the oozie end config which you mentioned. I
>> suggest you could check the config end more and debug there.
>> Also will let other community folks to pitch in if they have some other
>> opinion.
>>
>> Thanks
>> Sunil
>>
>> On Mon, Aug 22, 2016 at 8:57 PM rammohan ganapavarapu <
>> rammohanga...@gmail.com> wrote:
>>
>>> any thoughts from the logs and config I have shared?
>>>
>>> On Aug 21, 2016 8:32 AM, "rammohan ganapavarapu" <
>>> rammohanga...@gmail.com> wrote:
>>>
>>>> so in job.properties what is the jobtracker property, is it RM ip: port
>>>> or scheduler port which is 8030, if I use 8030 I am getting unknown
>>>> protocol proto buffer error.
>>>>
>>>> On Aug 21, 2016 7:37 AM, "Sunil Govind" <sunil.gov...@gmail.com> wrote:
>>>>
>>>>> Hi.
>>>>>
>>>>> It seems its an oozie issue. From conf, RM scheduler is running at
>>>>> port 8030.
>>>>> But your job.properties is taking 8032. I suggest you could double
>>>>> confirm your oozie configuration and see the configurations are intact to
>>>>> contact RM. Sharing a link also
>>>>> https://discuss.zendesk.com/hc/en-us/articles/203355837-How-
>>>>> to-run-a-MapReduce-jar-using-Oozie-workflow
>>>>>
>>>>> Thanks
>>>>> Sunil
>>>>>
>>>>>
>>>>> On Sun, Aug 21, 2016 at 8:41 AM rammohan ganapavarapu <
>>>>> rammohanga...@gmail.com> wrote:
>>>>>
>>>>>> Please find the attached config that i got from yarn ui and  AM,RM
>>>>>> logs. I only see that connecting to 0.0.0.0:8030 when i submit job
>>>>>> using oozie, but if i submit as yarn jar its working fine as i posted in 
>>>>>> my
>>>>>> previous posts.
>>>>>>
>>>>>> Here is my oozie job.properties file, i have a java class that just
>>>>>> prints
>>>>>>
>>>>>> nameNode=hdfs://master01:8020
>>>>>> jobTracker=master01:8032
>>>>>> workflowName=EchoJavaJob
>>>>>> oozie.use.system.libpath=true
>>>>>>
>>>>>> queueName=default
>>>>>> hdfsWorkflowHome=/user/uap/oozieWorkflows
>>>>>>
>>>>>> workflowPath=${nameNode}${hdfsWorkflowHome}/${workflowName}
>>>>>> oozie.wf.application.path=${workflowPath}
>>>>>>
>>>>>> Please let me know if you guys find any clue why its trying to
>>>>>> connect to 0.0.0.:8030.
>>>>>>
>>>>>> Thanks,
>>>>>> Ram
>>>>>>
>>>>>>
>>>>>> On Fri, Aug 19, 2016 at 11:54 PM, Sunil Govind <
>>>>>> sunil.gov...@gmail.com> wrote:
>>>>>>
>>>>>>> Hi Ram
>>>>>>>
>>>>>>> From the console log, as Rohith said, AM is looking for AM at 8030.
>>>>>>> So pls confirm the RM port once.
>>>>>>> Could you please share AM and RM logs.
>>>>>>>
>>>>>>> Thanks
>>>>>>> Sunil
>>>>>>>
>>>>>>> On Sat, Aug 20, 2016 at 10:36 AM rammohan ganapavarapu <
>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>
>>>>>>>> yes, I did configured.
>>>>>>>>
>>>>>>>> On Aug 19, 2016 7:22 PM, "Rohith Sharma K S" <
>>>>>>>> ksrohithsha...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hi
>>>>>>>>>
>>>>>>>>> From below discussion and AM logs, I see that AM container has
>>>>>>>>> launched but not able to connect to RM.
>>>>>>>>>
>>>>>>>>> This looks like your configuration issue. Would you check your
>>>>>>>>> job.xml jar that does *yarn.resourcemanager.scheduler.address *has
>>>>>>>>> been configured?
>>>>>>>>>
>>>>>>>>> Essentially, this address required by MRAppMaster for connecting
>>>>>>>>> to RM for heartbeats. If you don’t not configure, default value will 
>>>>>>>>> be
>>>>>>>>> taken i.e 8030.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thanks & Regards
>>>>>>>>> Rohith Sharma K S
>>>>>>>>>
>>>>>>>>> On Aug 20, 2016, at 7:02 AM, rammohan ganapavarapu <
>>>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>> Even if  the cluster dont have enough resources it should connect
>>>>>>>>> to "
>>>>>>>>>
>>>>>>>>> /0.0.0.0:8030" right? it should connect to my <RM_HOST:8030>, not 
>>>>>>>>> sure why its trying to connect to 0.0.0.0:8030.
>>>>>>>>>
>>>>>>>>> I have verified the config and i removed traces of 0.0.0.0 still no 
>>>>>>>>> luck.
>>>>>>>>>
>>>>>>>>> org.apache.hadoop.yarn.client.RMProxy: Connecting to ResourceManager 
>>>>>>>>> at /0.0.0.0:8030
>>>>>>>>>
>>>>>>>>> If an one has any clue please share.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>>
>>>>>>>>> Ram
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Fri, Aug 19, 2016 at 2:32 PM, rammohan ganapavarapu <
>>>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> When i submit a job using yarn its seems working only with oozie
>>>>>>>>>> its failing i guess, not sure what is missing.
>>>>>>>>>>
>>>>>>>>>> yarn jar /uap/hadoop/share/hadoop/mapre
>>>>>>>>>> duce/hadoop-mapreduce-examples-2.7.1.jar pi 20 1000
>>>>>>>>>> Number of Maps  = 20
>>>>>>>>>> Samples per Map = 1000
>>>>>>>>>> .
>>>>>>>>>> .
>>>>>>>>>> .
>>>>>>>>>> Job Finished in 19.622 seconds
>>>>>>>>>> Estimated value of Pi is 3.14280000000000000000
>>>>>>>>>>
>>>>>>>>>> Ram
>>>>>>>>>>
>>>>>>>>>> On Fri, Aug 19, 2016 at 11:46 AM, rammohan ganapavarapu <
>>>>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Ok, i have used yarn-utils.py to get the correct values for my
>>>>>>>>>>> cluster and update those properties and restarted RM and NM but 
>>>>>>>>>>> still no
>>>>>>>>>>> luck not sure what i am missing, any other insights will help me.
>>>>>>>>>>>
>>>>>>>>>>> Below are my properties from yarn-site.xml and map-site.xml.
>>>>>>>>>>>
>>>>>>>>>>> python yarn-utils.py -c 24 -m 63 -d 3 -k False
>>>>>>>>>>>  Using cores=24 memory=63GB disks=3 hbase=False
>>>>>>>>>>>  Profile: cores=24 memory=63488MB reserved=1GB usableMem=62GB
>>>>>>>>>>> disks=3
>>>>>>>>>>>  Num Container=6
>>>>>>>>>>>  Container Ram=10240MB
>>>>>>>>>>>  Used Ram=60GB
>>>>>>>>>>>  Unused Ram=1GB
>>>>>>>>>>>  yarn.scheduler.minimum-allocation-mb=10240
>>>>>>>>>>>  yarn.scheduler.maximum-allocation-mb=61440
>>>>>>>>>>>  yarn.nodemanager.resource.memory-mb=61440
>>>>>>>>>>>  mapreduce.map.memory.mb=5120
>>>>>>>>>>>  mapreduce.map.java.opts=-Xmx4096m
>>>>>>>>>>>  mapreduce.reduce.memory.mb=10240
>>>>>>>>>>>  mapreduce.reduce.java.opts=-Xmx8192m
>>>>>>>>>>>  yarn.app.mapreduce.am.resource.mb=5120
>>>>>>>>>>>  yarn.app.mapreduce.am.command-opts=-Xmx4096m
>>>>>>>>>>>  mapreduce.task.io.sort.mb=1024
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>mapreduce.map.memory.mb</name>
>>>>>>>>>>>       <value>5120</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>mapreduce.map.java.opts</name>
>>>>>>>>>>>       <value>-Xmx4096m</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>mapreduce.reduce.memory.mb</name>
>>>>>>>>>>>       <value>10240</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>mapreduce.reduce.java.opts</name>
>>>>>>>>>>>       <value>-Xmx8192m</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>yarn.app.mapreduce.am.resource.mb</name>
>>>>>>>>>>>       <value>5120</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>yarn.app.mapreduce.am.command-opts</name>
>>>>>>>>>>>       <value>-Xmx4096m</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>     <property>
>>>>>>>>>>>       <name>mapreduce.task.io.sort.mb</name>
>>>>>>>>>>>       <value>1024</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>      <property>
>>>>>>>>>>>       <name>yarn.scheduler.minimum-allocation-mb</name>
>>>>>>>>>>>       <value>10240</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>
>>>>>>>>>>>      <property>
>>>>>>>>>>>       <name>yarn.scheduler.maximum-allocation-mb</name>
>>>>>>>>>>>       <value>61440</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>
>>>>>>>>>>>      <property>
>>>>>>>>>>>       <name>yarn.nodemanager.resource.memory-mb</name>
>>>>>>>>>>>       <value>61440</value>
>>>>>>>>>>>     </property>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Ram
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Aug 18, 2016 at 11:14 PM, tkg_cangkul <
>>>>>>>>>>> yuza.ras...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> maybe this link can be some reference to tune up the cluster:
>>>>>>>>>>>>
>>>>>>>>>>>> http://jason4zhu.blogspot.co.id/2014/10/memory-configuration
>>>>>>>>>>>> -in-hadoop.html
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 19/08/16 11:13, rammohan ganapavarapu wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Do you know what properties to tune?
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> Ram
>>>>>>>>>>>>
>>>>>>>>>>>> On Thu, Aug 18, 2016 at 9:11 PM, tkg_cangkul <
>>>>>>>>>>>> yuza.ras...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> i think that's because you don't have enough resource.  u can
>>>>>>>>>>>>> tune your cluster config to maximize your resource.
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On 19/08/16 11:03, rammohan ganapavarapu wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>> I dont see any thing odd except this not sure if i have to
>>>>>>>>>>>>> worry about it or not.
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2016-08-19 03:29:26,621 INFO [main]
>>>>>>>>>>>>> org.apache.hadoop.yarn.client.RMProxy: Connecting to
>>>>>>>>>>>>> ResourceManager at /0.0.0.0:8030
>>>>>>>>>>>>> 2016-08-19 03:29:27,646 INFO [main]
>>>>>>>>>>>>> org.apache.hadoop.ipc.Client: Retrying connect to server:
>>>>>>>>>>>>> 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy
>>>>>>>>>>>>> is RetryUpToMaximumCo
>>>>>>>>>>>>> untWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
>>>>>>>>>>>>> 2016-08-19 03:29:28,647 INFO [main]
>>>>>>>>>>>>> org.apache.hadoop.ipc.Client: Retrying connect to server:
>>>>>>>>>>>>> 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy
>>>>>>>>>>>>> is RetryUpToMaximumCountWithFixedSleep(maxRetries=10,
>>>>>>>>>>>>> sleepTime=1000 MILLISECONDS)
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> its keep printing this log ..in app container logs.
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Thu, Aug 18, 2016 at 8:20 PM, tkg_cangkul <
>>>>>>>>>>>>> yuza.ras...@gmail.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> maybe u can check the logs from port 8088 on your browser.
>>>>>>>>>>>>>> that was RM UI. just choose your job id and then check the logs.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 19/08/16 10:14, rammohan ganapavarapu wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Sunil,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks you for your input, below are my server metrics for
>>>>>>>>>>>>>> RM. Also attached RM UI for capacity scheduler resources. How 
>>>>>>>>>>>>>> else i can
>>>>>>>>>>>>>> find?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> {
>>>>>>>>>>>>>>       "name": "Hadoop:service=ResourceManage
>>>>>>>>>>>>>> r,name=QueueMetrics,q0=root",
>>>>>>>>>>>>>>       "modelerType": "QueueMetrics,q0=root",
>>>>>>>>>>>>>>       "tag.Queue": "root",
>>>>>>>>>>>>>>       "tag.Context": "yarn",
>>>>>>>>>>>>>>       "tag.Hostname": "hadoop001",
>>>>>>>>>>>>>>       "running_0": 0,
>>>>>>>>>>>>>>       "running_60": 0,
>>>>>>>>>>>>>>       "running_300": 0,
>>>>>>>>>>>>>>       "running_1440": 0,
>>>>>>>>>>>>>>       "AppsSubmitted": 1,
>>>>>>>>>>>>>>       "AppsRunning": 0,
>>>>>>>>>>>>>>       "AppsPending": 0,
>>>>>>>>>>>>>>       "AppsCompleted": 0,
>>>>>>>>>>>>>>       "AppsKilled": 0,
>>>>>>>>>>>>>>       "AppsFailed": 1,
>>>>>>>>>>>>>>       "AllocatedMB": 0,
>>>>>>>>>>>>>>       "AllocatedVCores": 0,
>>>>>>>>>>>>>>       "AllocatedContainers": 0,
>>>>>>>>>>>>>>       "AggregateContainersAllocated": 2,
>>>>>>>>>>>>>>       "AggregateContainersReleased": 2,
>>>>>>>>>>>>>>       "AvailableMB": 64512,
>>>>>>>>>>>>>>       "AvailableVCores": 24,
>>>>>>>>>>>>>>       "PendingMB": 0,
>>>>>>>>>>>>>>       "PendingVCores": 0,
>>>>>>>>>>>>>>       "PendingContainers": 0,
>>>>>>>>>>>>>>       "ReservedMB": 0,
>>>>>>>>>>>>>>       "ReservedVCores": 0,
>>>>>>>>>>>>>>       "ReservedContainers": 0,
>>>>>>>>>>>>>>       "ActiveUsers": 0,
>>>>>>>>>>>>>>       "ActiveApplications": 0
>>>>>>>>>>>>>>     },
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Thu, Aug 18, 2016 at 6:49 PM, Sunil Govind <
>>>>>>>>>>>>>> sunil.gov...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> It could be because of many of reasons. Also I am not sure
>>>>>>>>>>>>>>> about which scheduler your are using, pls share more details 
>>>>>>>>>>>>>>> such as RM log
>>>>>>>>>>>>>>> etc.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I could point out few reasons
>>>>>>>>>>>>>>>  - Such as "Not enough resource is cluster" can cause this
>>>>>>>>>>>>>>>  - If using Capacity Scheduler, if queue capacity is maxed
>>>>>>>>>>>>>>> out, such case can happen.
>>>>>>>>>>>>>>>  - Similarly if max-am-resource-percent is crossed per queue
>>>>>>>>>>>>>>> level, then also AM container may not be launched.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> you could check RM log to get more information if AM
>>>>>>>>>>>>>>> container is laucnhed.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>>>> Sunil
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Fri, Aug 19, 2016 at 5:37 AM rammohan ganapavarapu <
>>>>>>>>>>>>>>> rammohanga...@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> When i submit a MR job, i am getting this from AM UI but it
>>>>>>>>>>>>>>>> never get finished, what am i missing ?
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>> Ram
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> ---------------------------------------------------------------------
>>>>>>>>>>>>>> To unsubscribe, e-mail: user-unsubscr...@hadoop.apache.org
>>>>>>>>>>>>>> For additional commands, e-mail: user-h...@hadoop.apache.org
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>
>

Reply via email to