org.apache.hadoop.ipc.StandbyException occurs at the thirty of per hour in standby NN

2014-01-24 Thread Francis . Hu
hello, All

 

Installed 2 NN and 3 DN in my hadoop-2.2.0 cluster,and implemented HDFS HA
with QJM. Currently, looking at the log of standby NN ,it throws below
exception at a regular interval, one hour:

 

2014-01-24 03:30:01,245 ERROR
org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
as:ubuntu (auth:SIMPLE) cause:org.apache.hadoop.ipc.StandbyException:
Operation category READ is not supported in state standby

 

Actually, the active NN is working ,and no applications access HDFS with
error.

 

Does anyone know what the problem is ?

 

 

 

Thanks,

Francis.Hu



答复: which web ui can track yarn job status?

2013-09-11 Thread Francis . Hu
what the value do you set for yarn.resourcemanager.webapp.address in
yarn-site.xml ?

 

Francis.Hu

 

发件人: ch huang [mailto:justlo...@gmail.com] 
发送时间: Wednesday, September 11, 2013 15:34
收件人: user@hadoop.apache.org
主题: which web ui can track yarn job status?

 

hi,all: 

  i use yarn so 50030 is not available for job status,thanks



答复: Is there any way to set Reducer to output to multi-places?

2013-09-02 Thread Francis . Hu
Thanks, Binglin

 

I found the class below that can do it :).

http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/lib/MultipleOutputs.html
 

 

 

 

发件人: Binglin Chang [mailto:decst...@gmail.com] 
发送时间: Monday, September 02, 2013 17:37
收件人: user@hadoop.apache.org
主题: Re: Is there any way to set Reducer to output to multi-places?

 

MultipleOutputFormat allows you to write multiple files in one reducer, but 
can't write output to HDFS and Database concurrently, but I is a good example 
to show how you can write a customized OutputFormat to achieve this.

Please note that for fault tolerance, a reducer may run multiple times, this 
may generate redundant data, hadoop handles files using FileOutputCommitter, 
you need to handle database case by yourself(e.g. insert record only if record 
doesn't exists). 

 

On Mon, Sep 2, 2013 at 5:11 PM, Rahul Bhattacharjee  
wrote:

This might help

http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/lib/MultipleOutputFormat.html

Thanks,
Rahul

 

On Mon, Sep 2, 2013 at 2:38 PM, Francis.Hu  wrote:

hi, All

 

Is there any way to set Reducer to output to multi-places ?  For example: a 
reducer's result can be output to HDFS and Database concurrently.

 

Thanks,

Francis.Hu

 

 



答复: Is there any way to set Reducer to output to multi-places?

2013-09-02 Thread Francis . Hu
Hi, Rahul

 

I found the class MultipleOutputs which has to do with what I want .  It can 
make Reducers to write to additional outputs(File and Database in my Env.) 
other than the job default output. 

 

http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/lib/MultipleOutputs.html
 

The class comments says below that I want to implement in my project:

 

*  

 * Case one: writing to additional outputs other than the job default output.

 *

 * Each additional output, or named output, may be configured with its own

 * OutputFormat, with its own key class and with its own value

 * class.

 * 

 

Anyway, thanks! Rahul. Your link leads me to the class MultipleOutputs :).

 

 

 

Thanks,

Francis.Hu.

 

 

发件人: Rahul Bhattacharjee [mailto:rahul.rec@gmail.com] 
发送时间: Monday, September 02, 2013 17:12
收件人: user@hadoop.apache.org
主题: Re: Is there any way to set Reducer to output to multi-places?

 

This might help

http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/lib/MultipleOutputFormat.html

Thanks,
Rahul

 

On Mon, Sep 2, 2013 at 2:38 PM, Francis.Hu  wrote:

hi, All

 

Is there any way to set Reducer to output to multi-places ?  For example: a 
reducer's result can be output to HDFS and Database concurrently.

 

Thanks,

Francis.Hu

 



Is there any way to set Reducer to output to multi-places?

2013-09-02 Thread Francis . Hu
hi, All

 

Is there any way to set Reducer to output to multi-places ?  For example: a
reducer's result can be output to HDFS and Database concurrently.

 

Thanks,

Francis.Hu



答复: Hadoop HA error "JOURNAL is not supported in state standby"

2013-08-30 Thread Francis . Hu
Did you start up your ZKFC service on both of your name nodes ?

Thanks,
Francis.Hu

-邮件原件-
发件人: orahad bigdata [mailto:oracle...@gmail.com] 
发送时间: Friday, August 30, 2013 4:09
收件人: user
主题: Hadoop HA error "JOURNAL is not supported in state standby"

Hi,

I'm facing an error while starting Hadoop in HA(2.0.5) cluster , both
the NameNode started in standby mode and not changing the state.

When I tried to do health check through  "hdfs haadmin -checkhealth
 " it's giving me below error.

Failed on local exception:
com.google.protobuf.InvalidProtocolBufferException: Message missing
required fields: callId, status; Host Details : local host is:
"clone2/XX.XX.XX.XX"; destination host is: "clone1":8020;

 I checked the logs at NN side.

2013-08-30 00:49:16,074 ERROR
org.apache.hadoop.security.UserGroupInformation:
PriviledgedActionException as:hadoop (auth:SIMPLE)
cause:org.apache.hadoop.ipc.StandbyException: Operation category
JOURNAL is not supported in state standby
2013-08-30 00:49:16,074 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 1 on 8020, call
org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol.rollEditLog
from 192.168.126.31:48266: error:
org.apache.hadoop.ipc.StandbyException: Operation category JOURNAL is
not supported in state standby
2013-08-30 00:49:32,391 INFO
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Triggering
log roll on remote NameNode clone2:8020
2013-08-30 00:49:32,403 WARN
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Unable to
trigger a roll of the active NN
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException
):
Operation category JOURNAL is not supported in state standby
at
org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(Standb
yState.java:87)
at
org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOpera
tion(NameNode.java:1411)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesys
tem.java:859)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.rollEditLog(FSNamesystem
.java:4445)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.rollEditLog(NameNod
eRpcServer.java:766)
at
org.apache.hadoop.hdfs.protocolPB.NamenodeProtocolServerSideTranslatorPB.rol
lEditLog(NamenodeProtocolServerSideTranslatorPB.java:139)
at
org.apache.hadoop.hdfs.protocol.proto.NamenodeProtocolProtos$NamenodeProtoco
lService$2.callBlockingMethod(NamenodeProtocolProtos.java:8758)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(Proto
bufRpcEngine.java:454)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1014)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1741)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1737)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.ja
va:1478)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1735)

at org.apache.hadoop.ipc.Client.call(Client.java:1235)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.jav
a:202)
at $Proxy11.rollEditLog(Unknown Source)
at
org.apache.hadoop.hdfs.protocolPB.NamenodeProtocolTranslatorPB.rollEditLog(N
amenodeProtocolTranslatorPB.java:139)
at
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer.triggerActiveLogRoll
(EditLogTailer.java:268)
at
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer.access$600(EditLogTa
iler.java:61)
at
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.
doWork(EditLogTailer.java:310)
at
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.
access$200(EditLogTailer.java:279)
at
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread$
1.run(EditLogTailer.java:296)
at
org.apache.hadoop.security.SecurityUtil.doAsLoginUserOrFatal(SecurityUtil.ja
va:456)
at
org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer$EditLogTailerThread.
run(EditLogTailer.java:292)

Did I missed something?

Thanks



答复: about append

2013-08-20 Thread Francis . Hu
it is due to that it cannot handle the situation well about the concurrent 
access to same file.

 

发件人: gsw204 [mailto:gsw...@163.com] 
发送时间: Tuesday, August 20, 2013 18:28
收件人: user
主题: about append 

 

hi,

   I want to know why append does not support in hadoop-1.1.x?

 

  _  

gsw

 



Cannot kill a job with correct job ID.

2013-08-20 Thread Francis . Hu
Hi, All

 

In my resources manager web page, I see the job
application_1376968101508_0009 is running.

I want to kill the job with the command:  bin/mapred job -kill
application_1376968101508_0009

It throws below error at all the time :

 

Exception in thread "main" java.lang.IllegalArgumentException: JobId string
: application_1376968101508_0009 is not properly formed

at org.apache.hadoop.mapreduce.JobID.forName(JobID.java:156)

at org.apache.hadoop.mapreduce.tools.CLI.run(CLI.java:276)

at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)

at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)

at org.apache.hadoop.mapred.JobClient.main(JobClient.java:1199)

 

And then looked at the source code of JobID class, found below method :
JobID.forName(String). 

It means we need to kill a job with name "job_1376968101508_0009", not
"application_1376968101508_0009".

That is really strange to kill a job with special name and without the
actual job name showing in resources manager UI.

Is this an existing issue ?

 

protected static final String JOB = "job";

public static JobID forName(String str) throws IllegalArgumentException {

if(str == null)

  return null;

try {

  String[] parts = str.split("_");

  if(parts.length == 3) {

if(parts[0].equals(JOB)) {

  return new org.apache.hadoop.mapred.JobID(parts[1], 

 
Integer.parseInt(parts[2]));

}

  }

}catch (Exception ex) {//fall below

}

throw new IllegalArgumentException("JobId string : " + str 

+ " is not properly formed");

  }

 

 

Thanks,

Francis.Hu



答复: Passing an object in mapper

2013-08-14 Thread Francis . Hu
hi, 

 

what I did when data need to be saved into DB :

 

//configure DB connection info.

DBConfiguration.configureDB(config,

HadoopJobConfigParser.getProperty(DB_DRIVER),

HadoopJobConfigParser.getProperty(DB_URL),

HadoopJobConfigParser.getProperty(DB_USER),

HadoopJobConfigParser.getProperty(DB_PWD));

 

//write a OutputKeyClass like below:

public class DisplayRequestRecord implements DBWritable,
WritableComparable{

//TODO implement write() and readFiels() methods

}

 

//set the job's output

job.setOutputKeyClass(DisplayRequestRecord.class);

job.setOutputValueClass(DisplayRequestRecord.class);

DBOutputFormat.setOutput(job, tableName(),Fields());

 

//define reducer class like below:

public class DBDisplayRequestReducerNew extends Reducer{

public void reduce(Text key, Iterable values,

Context context) throws IOException, InterruptedException {

DisplayRequestRecord dr = new DisplayRequestRecord();

context.write(dr,dr);

}

}

 

 

Hope it helps!

 

Thanks,

Francis.Hu

 

发件人: Sivaram RL [mailto:sivaram...@gmail.com] 
发送时间: Thursday, August 15, 2013 11:26
收件人: user@hadoop.apache.org
主题: Re: Passing an object in mapper

 

Hi,

 

In the Configuration object in your driver class you can set the properties
as key value pair. This configuration object will be set in the Job Object.
The same properties can be accessed in the mapper/reducer using the Context
Object -> getConfiguration() -> get(propertyName).

 

Hope this helps.

 

Regards,

Sivaram R L

 

On Thu, Aug 15, 2013 at 8:39 AM, jamal sasha  wrote:

Hi,

   I am initializing an object in driver code.

For sake of argument let say I want to save data to some database..

say:

Connection con = new Connection(host, db);

Now, in reducer I want to do something like

con.write(key,value) 

 

So, how do i pass this object from driver to mapper / reducer/?
Any clue? suggestions?

Thanks

 



答复: cannot submit a job via java client in hadoop- 2.0.5-alpha

2013-07-10 Thread Francis . Hu
Hi, Devaraj k  and Azuryy Yu

 

Thanks both of you.

I just get it resolved.  The problem is that below highlighted jar is not
included in my java client side so that when the Job is initializing, it can
not find the class YarnClientProtocolProvider to do further initialization.
Then it causes the exception in Cluster.initialize() is thrown out.

hadoop-mapreduce-client-jobclient-2.0.5-alpha.jar

 

Thanks,

Francis.Hu

 

 

发件人: Devaraj k [mailto:devara...@huawei.com] 
发送时间: Wednesday, July 10, 2013 19:11
收件人: user@hadoop.apache.org
主题: RE: cannot submit a job via java client in hadoop- 2.0.5-alpha

 

Hi Francis,

 

Could you check whether those configuration files are
getting loaded or not, There could be a chance that these configuration
files are not getting loaded into configuration object due to some invalid
path reason.

 

conf1.addResource("resources/mapred-site.xml");

   // Can you try printing the properties
‘yarn.resourcemanager.address’ & ‘mapreduce.framework.name’ values and
check they are   coming as per the config files

JobConf conf = new JobConf(conf1);

 



 

Thanks

Devaraj k

 

From: Azuryy Yu [mailto:azury...@gmail.com] 
Sent: 10 July 2013 16:22
To: user@hadoop.apache.org
Subject: Re: cannot submit a job via java client in hadoop- 2.0.5-alpha

 

you didn't set yarn.nodemanager.address in your yarn-site.xml

 

 

 

On Wed, Jul 10, 2013 at 4:33 PM, Francis.Hu 
wrote:

Hi,All

 

I have a hadoop- 2.0.5-alpha cluster with 3 data nodes . I have Resource
Manager and all data nodes started and can access web ui of Resource
Manager. 

I wrote a java client to submit a job as TestJob class below. But the job is
never submitted successfully. It throws out exception all the time. 

My configurations are attached.  Can anyone help me? Thanks.

 

-my-java client

public class TestJob {



public void execute() {

 

Configuration conf1 = new Configuration();

conf1.addResource("resources/core-site.xml");

conf1.addResource("resources/hdfs-site.xml");

conf1.addResource("resources/yarn-site.xml");

conf1.addResource("resources/mapred-site.xml");

JobConf conf = new JobConf(conf1);



conf.setJar("/home/francis/hadoop-jobs/MapReduceJob.jar");

conf.setJobName("Test");

 

conf.setInputFormat(TextInputFormat.class);

conf.setOutputFormat(TextOutputFormat.class);

 

conf.setOutputKeyClass(Text.class);

conf.setOutputValueClass(IntWritable.class);

 

conf.setMapperClass(DisplayRequestMapper.class);

conf.setReducerClass(DisplayRequestReducer.class);

 

FileInputFormat.setInputPaths(conf,new
Path("/home/francis/hadoop-jobs/2013070907.FNODE.2.txt"));

FileOutputFormat.setOutputPath(conf, new
Path("/home/francis/hadoop-jobs/result/"));

 

try {

JobClient client = new JobClient(conf);

RunningJob job = client.submitJob(conf);

job.waitForCompletion();

} catch (IOException e) {

e.printStackTrace();

}

}

}

 

--Exception 

 

jvm 1| java.io.IOException: Cannot initialize Cluster. Please check your
configuration for mapreduce.framework.name and the correspond server
addresses.

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:119)

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.(Cluster.java:81)

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.(Cluster.java:74)

jvm 1|  at
org.apache.hadoop.mapred.JobClient.init(JobClient.java:482)

jvm 1|  at
org.apache.hadoop.mapred.JobClient.(JobClient.java:461)

jvm 1|  at
com.rh.elastic.hadoop.job.TestJob.execute(TestJob.java:59)

 

 

Thanks,

Francis.Hu

 

 



答复: cannot submit a job via java client in hadoop- 2.0.5-alpha

2013-07-10 Thread Francis . Hu
 

Actually ,I have mapreduce.framework.name configured in mapred-site.xml, see
below:

 



mapreduce.framework.name

yarn

Execution framework set to Hadoop YARN.



 

 

发件人: hadoop hive [mailto:hadooph...@gmail.com] 
发送时间: Wednesday, July 10, 2013 18:39
收件人: user@hadoop.apache.org
主题: Re: cannot submit a job via java client in hadoop- 2.0.5-alpha

 

Here its showing like you are not using mapreduce.framework.name as yarn,
please resend it we are unable to see the configuration

 

On Wed, Jul 10, 2013 at 1:33 AM, Francis.Hu 
wrote:

Hi,All

 

I have a hadoop- 2.0.5-alpha cluster with 3 data nodes . I have Resource
Manager and all data nodes started and can access web ui of Resource
Manager. 

I wrote a java client to submit a job as TestJob class below. But the job is
never submitted successfully. It throws out exception all the time. 

My configurations are attached.  Can anyone help me? Thanks.

 

-my-java client

public class TestJob {



public void execute() {

 

Configuration conf1 = new Configuration();

conf1.addResource("resources/core-site.xml");

conf1.addResource("resources/hdfs-site.xml");

conf1.addResource("resources/yarn-site.xml");

conf1.addResource("resources/mapred-site.xml");

JobConf conf = new JobConf(conf1);



conf.setJar("/home/francis/hadoop-jobs/MapReduceJob.jar");

conf.setJobName("Test");

 

conf.setInputFormat(TextInputFormat.class);

conf.setOutputFormat(TextOutputFormat.class);

 

conf.setOutputKeyClass(Text.class);

conf.setOutputValueClass(IntWritable.class);

 

conf.setMapperClass(DisplayRequestMapper.class);

conf.setReducerClass(DisplayRequestReducer.class);

 

FileInputFormat.setInputPaths(conf,new
Path("/home/francis/hadoop-jobs/2013070907.FNODE.2.txt"));

FileOutputFormat.setOutputPath(conf, new
Path("/home/francis/hadoop-jobs/result/"));

 

try {

JobClient client = new JobClient(conf);

RunningJob job = client.submitJob(conf);

job.waitForCompletion();

} catch (IOException e) {

e.printStackTrace();

}

}

}

 

--Exception 

 

jvm 1| java.io.IOException: Cannot initialize Cluster. Please check your
configuration for mapreduce.framework.name and the correspond server
addresses.

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:119)

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.(Cluster.java:81)

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.(Cluster.java:74)

jvm 1|  at
org.apache.hadoop.mapred.JobClient.init(JobClient.java:482)

jvm 1|  at
org.apache.hadoop.mapred.JobClient.(JobClient.java:461)

jvm 1|  at
com.rh.elastic.hadoop.job.TestJob.execute(TestJob.java:59)

 

 

Thanks,

Francis.Hu

 

 



cannot submit a job via java client in hadoop- 2.0.5-alpha

2013-07-10 Thread Francis . Hu
Hi,All

 

I have a hadoop- 2.0.5-alpha cluster with 3 data nodes . I have Resource
Manager and all data nodes started and can access web ui of Resource
Manager. 

I wrote a java client to submit a job as TestJob class below. But the job is
never submitted successfully. It throws out exception all the time. 

My configurations are attached.  Can anyone help me? Thanks.

 

-my-java client

public class TestJob {



public void execute() {

 

Configuration conf1 = new Configuration();

conf1.addResource("resources/core-site.xml");

conf1.addResource("resources/hdfs-site.xml");

conf1.addResource("resources/yarn-site.xml");

conf1.addResource("resources/mapred-site.xml");

JobConf conf = new JobConf(conf1);



conf.setJar("/home/francis/hadoop-jobs/MapReduceJob.jar");

conf.setJobName("Test");

 

conf.setInputFormat(TextInputFormat.class);

conf.setOutputFormat(TextOutputFormat.class);

 

conf.setOutputKeyClass(Text.class);

conf.setOutputValueClass(IntWritable.class);

 

conf.setMapperClass(DisplayRequestMapper.class);

conf.setReducerClass(DisplayRequestReducer.class);

 

FileInputFormat.setInputPaths(conf,new
Path("/home/francis/hadoop-jobs/2013070907.FNODE.2.txt"));

FileOutputFormat.setOutputPath(conf, new
Path("/home/francis/hadoop-jobs/result/"));

 

try {

JobClient client = new JobClient(conf);

RunningJob job = client.submitJob(conf);

job.waitForCompletion();

} catch (IOException e) {

e.printStackTrace();

}

}

}

 

--Exception 

 

jvm 1| java.io.IOException: Cannot initialize Cluster. Please check your
configuration for mapreduce.framework.name and the correspond server
addresses.

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:119)

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.(Cluster.java:81)

jvm 1|  at
org.apache.hadoop.mapreduce.Cluster.(Cluster.java:74)

jvm 1|  at
org.apache.hadoop.mapred.JobClient.init(JobClient.java:482)

jvm 1|  at
org.apache.hadoop.mapred.JobClient.(JobClient.java:461)

jvm 1|  at
com.rh.elastic.hadoop.job.TestJob.execute(TestJob.java:59)

 

 

Thanks,

Francis.Hu

 









	yarn.acl.enable
	false
	Enable ACLs? Defaults to false.






	yarn.resourcemanager.address
	192.168.219.129:9001
	ResourceManager host:port for clients to submit jobs.


	yarn.resourcemanager.scheduler.address
	192.168.219.129:8030
	ResourceManager host:port for ApplicationMasters to talk to Scheduler to obtain resources.


	yarn.resourcemanager.resource-tracker.address
	192.168.219.129:8031
	ResourceManager host:port for NodeManagers.


	yarn.resourcemanager.admin.address
	192.168.219.129:8033
	ResourceManager host:port for administrative commands.


	yarn.resourcemanager.webapp.address
	192.168.219.129:8088
	ResourceManager web-ui host:port.


	yarn.resourcemanager.scheduler.class
	org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler
	ResourceManager Scheduler class.


	yarn.scheduler.minimum-allocation-mb
	1024
	Minimum limit of memory to allocate to each container request at the Resource Manager.


	yarn.scheduler.maximum-allocation-mb
	8192
	Maximum limit of memory to allocate to each container request at the Resource Manager.





	yarn.nodemanager.resource.memory-mb
	8192
	Resource i.e. available physical memory, in MB, for given NodeManager.Defines total available resources on the NodeManager to be made available to running containers


	yarn.nodemanager.vmem-pmem-ratio
	2.1
	Maximum ratio by which virtual memory usage of tasks may exceed physical memory


	yarn.nodemanager.local-dirs
	/home/francis/hadoop2-hdfs/yarn
	Comma-separated list of paths on the local filesystem where intermediate data is written.Multiple paths help spread disk i/o.


	yarn.nodemanager.log-dirs
	/home/francis/hadoop2-hdfs/yarn-log
	Comma-separated list of paths on the local filesystem where logs are written.Multiple paths help spread disk i/o.


	yarn.nodemanager.log.retain-seconds
	10800
	Default time (in seconds) to retain log files on the NodeManager Only applicable if log-aggregation is disabled.


	yarn.nodemanager.remote-app-log-dir
	/logs
	HDFS directory where the application logs are moved on application completion. Need to set appropriate permissions. Only applicable if log-aggregation is enabled.


	yarn.nodemanager.remote-app-log-dir-suffix
	logs
	Suffix appended to the remote log dir. Logs will be aggregated to ${yarn.nodemanager.remote-app-log-dir}/${user}/${thisParam} Only applicable if log-aggregation is enabled.


	yarn.nodemanager.aux-services
	mapreduce.shuffle
	Shuffle service that needs to be set for Map Reduce applications.




	yarn.log-aggregation.retain-seconds
	-1
	How long to keep aggregation logs before deleting them. -1 disables. Be c

答复: Decomssion datanode - no response

2013-07-04 Thread Francis . Hu
I know the default value is 10 minutes and 30 seconds for switching
datanodes from live to dead. 

 

发件人: Azuryy Yu [mailto:azury...@gmail.com] 
发送时间: Friday, July 05, 2013 10:42
收件人: user@hadoop.apache.org
主题: Decomssion datanode - no response

 

Hi,

I am using hadoop-2.0.5-alpha, and I added 5 datanodes into dfs_exclude, 

 

hdfs-site.xml:

  
dfs.hosts.exclude
/usr/local/hadoop/conf/dfs_exclude
  

 

then:

hdfs dfsadmin -refreshNodes

 

but there is no decomssion nodes showed on the webUI. and not any releated
logs in the datanode log. what's wrong?



答复: interval from Live to Dead about DataNode

2013-07-04 Thread Francis . Hu
Nicolas,

 

Thanks for your help. had a look at HDFS-3703.

 

So I need to turn on dfs.namenode.check.stale.datanode 

and set a shorter time for dfs.namenode.stale.datanode.interval

 

Thanks,

Francis.Hu

 

发件人: Nicolas Liochon [mailto:nkey...@gmail.com] 
发送时间: Thursday, July 04, 2013 16:28
收件人: user@hadoop.apache.org
主题: Re: interval from Live to Dead about DataNode

 

Hi,


It's 10 minutes and 30s.

See the stale mode described in HDFS-3703 if you need something shorter.

Cheers,


Nicolas

 

On Thu, Jul 4, 2013 at 10:05 AM, Francis.Hu 
wrote:

Hi,All

 

Does anyone know when the status of datanode switch from live to dead at the
internal of namenode ?

 

the scenario:

When i stopped a datanode with command, the status of that datanode in the
web UI of namenode displays 'live' and 'In Service' for almost 5 minutes.  

I know the default value of datanode dfs.hreadbeat.interval is set as 3
seconds and dfs.namenode.decommission.interval is set 30 seconds.

 

Thanks,

Francis.Hu

 



interval from Live to Dead about DataNode

2013-07-04 Thread Francis . Hu
Hi,All

 

Does anyone know when the status of datanode switch from live to dead at the
internal of namenode ?

 

the scenario:

When i stopped a datanode with command, the status of that datanode in the
web UI of namenode displays 'live' and 'In Service' for almost 5 minutes.  

I know the default value of datanode dfs.hreadbeat.interval is set as 3
seconds and dfs.namenode.decommission.interval is set 30 seconds.

 

Thanks,

Francis.Hu



reply: a question about dfs.replication

2013-07-01 Thread Francis . Hu
YouPeng Yang,

 

you said that may be the answer. Thank you. 

 

发件人: YouPeng Yang [mailto:yypvsxf19870...@gmail.com] 
发送时间: Tuesday, July 02, 2013 12:52
收件人: user@hadoop.apache.org
主题: Re: reply: a question about dfs.replication

 

HI HU and Yu

 

Aggree with dfs.replication is a client side configuration, not server 
side. It make the point in my last mail  sense.

   

And the cmd:hdfs dfs -setrep -R -w 2 /  solve the problem that I can not 
change the existed file's replication value.

 

 

 

2013/7/2 Azuryy Yu 

It's not HDFS issue. 

dfs.replication is a client side configuration, not server side. so you need to 
set it to '2' on your client side( your application running on). THEN execute 
command such as : hdfs dfs -put  or call HDFS API in java application.



 

On Tue, Jul 2, 2013 at 12:25 PM, Francis.Hu  
wrote:

Thanks all of you, I just get the problem fixed through the command: 

hdfs dfs -setrep -R -w 2 /

 

Is that an issue of HDFS ? Why do i need to execute manually a command to tell 
the hadoop the replication factor even it is set in hdfs-site.xml ?

 

Thanks,

Francis.Hu

 

发件人: Francis.Hu [mailto:francis...@reachjunction.com] 
发送时间: Tuesday, July 02, 2013 11:30
收件人: user@hadoop.apache.org
主题: 答复: 答复: a question about dfs.replication

 

Yes , it returns 2 correctly after "hdfs getconf -confkey dfs.replication"

 



 

but in web page ,it is 3 as below:



 

发件人: yypvsxf19870706 [mailto:yypvsxf19870...@gmail.com] 
发送时间: Monday, July 01, 2013 23:24
收件人: user@hadoop.apache.org
主题: Re: 答复: a question about dfs.replication

 

Hi 

 

Could you please get the property value by using : hdfs getconf -confkey 
dfs.replication.


鍙戣嚜鎴戠殑 iPhone


鍦?2013-7-1锛?5:51锛孎rancis.Hu  鍐欓亾锛?br>

 

Actually, My java client is running with the same configuration as the hadoop's 
. The dfs.replication is already set as 2 in my hadoop's configuration.

So i think the dfs.replication is already overrided by my configuration in 
hdfs-site.xml. but seems it doesn't work even i overrided the parameter 
evidently.

 

 

鍙戜欢浜?span lang="EN-US">: 袝屑械谢褜褟薪芯胁 袘芯褉懈褋 [mailto:emelya...@post.km.ru] 
鍙戦€佹椂闂?span lang="EN-US">: Monday, July 01, 2013 15:18
鏀朵欢浜?span lang="EN-US">: user@hadoop.apache.org
涓婚: Re: a question about dfs.replication

 

On 01.07.2013 10:19, Francis.Hu wrote:

Hi, All

 

I am installing a cluster with Hadoop 2.0.5-alpha. I have one namenode and two 
datanodes. The dfs.replication is set as 2 in hdfs-site.xml. After all 
configuration work is done, I started all nodes. Then I saved a file into HDFS 
through java client. nOW I can access hdfs web page: x.x.x.x:50070,and also see 
the file is already listed in the hdfs list.

My question is:  The replication column in HDFS web page is showing as 3, not 
2.  Does anyone know What the problem is?

 

---Actual setting of hdfs-site.xml



dfs.replication

2



 

After that, I typed dfsamdin command to check the file:

hdfs fsck /test3/

The result of above command:

/test3/hello005.txt:  Under replicated 
BP-609310498-192.168.219.129-1372323727200:blk_-1069303317294683372_1006. 
Target Replicas is 3 but found 2 replica(s).

Status: HEALTHY

 Total size:35 B

 Total dirs:1

 Total files:   1

 Total blocks (validated):  1 (avg. block size 35 B)

 Minimally replicated blocks:   1 (100.0 %)

 Over-replicated blocks:0 (0.0 %)

 Under-replicated blocks:   1 (100.0 %)

 Mis-replicated blocks: 0 (0.0 %)

 Default replication factor:2

 Average block replication: 2.0

 Corrupt blocks:0

 Missing replicas:  1 (33.32 %)

 Number of data-nodes:  3

 Number of racks:   1

FSCK ended at Sat Jun 29 16:51:37 CST 2013 in 6 milliseconds

 

 

Thanks,

Francis Hu

 

If I'm not mistaking "dfs.replication" parameter in config sets only default 
replication factor, which can be overrided when putting file to hdfs.

 

 

<><>

reply: a question about dfs.replication

2013-07-01 Thread Francis . Hu
actually, my client side is already set to "2". 

 

发件人: Azuryy Yu [mailto:azury...@gmail.com] 
发送时间: Tuesday, July 02, 2013 12:40
收件人: user@hadoop.apache.org
主题: Re: reply: a question about dfs.replication

 

It's not HDFS issue. 

dfs.replication is a client side configuration, not server side. so you need to 
set it to '2' on your client side( your application running on). THEN execute 
command such as : hdfs dfs -put  or call HDFS API in java application.



 

On Tue, Jul 2, 2013 at 12:25 PM, Francis.Hu  
wrote:

Thanks all of you, I just get the problem fixed through the command: 

hdfs dfs -setrep -R -w 2 /

 

Is that an issue of HDFS ? Why do i need to execute manually a command to tell 
the hadoop the replication factor even it is set in hdfs-site.xml ?

 

Thanks,

Francis.Hu

 

发件人: Francis.Hu [mailto:francis...@reachjunction.com] 
发送时间: Tuesday, July 02, 2013 11:30
收件人: user@hadoop.apache.org
主题: 答复: 答复: a question about dfs.replication

 

Yes , it returns 2 correctly after "hdfs getconf -confkey dfs.replication"

 



 

but in web page ,it is 3 as below:



 

发件人: yypvsxf19870706 [mailto:yypvsxf19870...@gmail.com] 
发送时间: Monday, July 01, 2013 23:24
收件人: user@hadoop.apache.org
主题: Re: 答复: a question about dfs.replication

 

Hi 

 

Could you please get the property value by using : hdfs getconf -confkey 
dfs.replication.


鍙戣嚜鎴戠殑 iPhone


鍦?2013-7-1锛?5:51锛孎rancis.Hu  鍐欓亾锛?br>

 

Actually, My java client is running with the same configuration as the hadoop's 
. The dfs.replication is already set as 2 in my hadoop's configuration.

So i think the dfs.replication is already overrided by my configuration in 
hdfs-site.xml. but seems it doesn't work even i overrided the parameter 
evidently.

 

 

鍙戜欢浜?span lang="EN-US">: 袝屑械谢褜褟薪芯胁 袘芯褉懈褋 [mailto:emelya...@post.km.ru] 
鍙戦€佹椂闂?span lang="EN-US">: Monday, July 01, 2013 15:18
鏀朵欢浜?span lang="EN-US">: user@hadoop.apache.org
涓婚: Re: a question about dfs.replication

 

On 01.07.2013 10:19, Francis.Hu wrote:

Hi, All

 

I am installing a cluster with Hadoop 2.0.5-alpha. I have one namenode and two 
datanodes. The dfs.replication is set as 2 in hdfs-site.xml. After all 
configuration work is done, I started all nodes. Then I saved a file into HDFS 
through java client. nOW I can access hdfs web page: x.x.x.x:50070,and also see 
the file is already listed in the hdfs list.

My question is:  The replication column in HDFS web page is showing as 3, not 
2.  Does anyone know What the problem is?

 

---Actual setting of hdfs-site.xml



dfs.replication

2



 

After that, I typed dfsamdin command to check the file:

hdfs fsck /test3/

The result of above command:

/test3/hello005.txt:  Under replicated 
BP-609310498-192.168.219.129-1372323727200:blk_-1069303317294683372_1006. 
Target Replicas is 3 but found 2 replica(s).

Status: HEALTHY

 Total size:35 B

 Total dirs:1

 Total files:   1

 Total blocks (validated):  1 (avg. block size 35 B)

 Minimally replicated blocks:   1 (100.0 %)

 Over-replicated blocks:0 (0.0 %)

 Under-replicated blocks:   1 (100.0 %)

 Mis-replicated blocks: 0 (0.0 %)

 Default replication factor:2

 Average block replication: 2.0

 Corrupt blocks:0

 Missing replicas:  1 (33.32 %)

 Number of data-nodes:  3

 Number of racks:           1

FSCK ended at Sat Jun 29 16:51:37 CST 2013 in 6 milliseconds

 

 

Thanks,

Francis Hu

 

If I'm not mistaking "dfs.replication" parameter in config sets only default 
replication factor, which can be overrided when putting file to hdfs.

 

<><>

reply: a question about dfs.replication

2013-07-01 Thread Francis . Hu
Thanks all of you, I just get the problem fixed through the command: 

hdfs dfs -setrep -R -w 2 /

 

Is that an issue of HDFS ? Why do i need to execute manually a command to tell 
the hadoop the replication factor even it is set in hdfs-site.xml ?

 

Thanks,

Francis.Hu

 

发件人: Francis.Hu [mailto:francis...@reachjunction.com] 
发送时间: Tuesday, July 02, 2013 11:30
收件人: user@hadoop.apache.org
主题: 答复: 答复: a question about dfs.replication

 

Yes , it returns 2 correctly after "hdfs getconf -confkey dfs.replication"

 



 

but in web page ,it is 3 as below:



 

发件人: yypvsxf19870706 [mailto:yypvsxf19870...@gmail.com] 
发送时间: Monday, July 01, 2013 23:24
收件人: user@hadoop.apache.org
主题: Re: 答复: a question about dfs.replication

 

Hi 

 

Could you please get the property value by using : hdfs getconf -confkey 
dfs.replication.


鍙戣嚜鎴戠殑 iPhone


鍦?2013-7-1锛?5:51锛孎rancis.Hu  鍐欓亾锛?br>

 

Actually, My java client is running with the same configuration as the hadoop's 
. The dfs.replication is already set as 2 in my hadoop's configuration.

So i think the dfs.replication is already overrided by my configuration in 
hdfs-site.xml. but seems it doesn't work even i overrided the parameter 
evidently.

 

 

鍙戜欢浜?span lang="EN-US">: 袝屑械谢褜褟薪芯胁 袘芯褉懈褋 [mailto:emelya...@post.km.ru] 
鍙戦€佹椂闂?span lang="EN-US">: Monday, July 01, 2013 15:18
鏀朵欢浜?span lang="EN-US">: user@hadoop.apache.org
涓婚: Re: a question about dfs.replication

 

On 01.07.2013 10:19, Francis.Hu wrote:

Hi, All

 

I am installing a cluster with Hadoop 2.0.5-alpha. I have one namenode and two 
datanodes. The dfs.replication is set as 2 in hdfs-site.xml. After all 
configuration work is done, I started all nodes. Then I saved a file into HDFS 
through java client. nOW I can access hdfs web page: x.x.x.x:50070,and also see 
the file is already listed in the hdfs list.

My question is:  The replication column in HDFS web page is showing as 3, not 
2.  Does anyone know What the problem is?

 

---Actual setting of hdfs-site.xml



dfs.replication

2



 

After that, I typed dfsamdin command to check the file:

hdfs fsck /test3/

The result of above command:

/test3/hello005.txt:  Under replicated 
BP-609310498-192.168.219.129-1372323727200:blk_-1069303317294683372_1006. 
Target Replicas is 3 but found 2 replica(s).

Status: HEALTHY

 Total size:35 B

 Total dirs:1

 Total files:   1

 Total blocks (validated):  1 (avg. block size 35 B)

 Minimally replicated blocks:   1 (100.0 %)

 Over-replicated blocks:0 (0.0 %)

 Under-replicated blocks:   1 (100.0 %)

 Mis-replicated blocks: 0 (0.0 %)

 Default replication factor:2

 Average block replication: 2.0

 Corrupt blocks:0

 Missing replicas:  1 (33.32 %)

 Number of data-nodes:  3

 Number of racks:   1

FSCK ended at Sat Jun 29 16:51:37 CST 2013 in 6 milliseconds

 

 

Thanks,

Francis Hu

 

If I'm not mistaking "dfs.replication" parameter in config sets only default 
replication factor, which can be overrided when putting file to hdfs.

<><>

答复: 答复: a question about dfs.replication

2013-07-01 Thread Francis . Hu
Yes , it returns 2 correctly after "hdfs getconf -confkey dfs.replication"

 



 

but in web page ,it is 3 as below:



 

发件人: yypvsxf19870706 [mailto:yypvsxf19870...@gmail.com] 
发送时间: Monday, July 01, 2013 23:24
收件人: user@hadoop.apache.org
主题: Re: 答复: a question about dfs.replication

 

Hi 

 

Could you please get the property value by using : hdfs getconf -confkey 
dfs.replication.


鍙戣嚜鎴戠殑 iPhone


鍦?2013-7-1锛?5:51锛孎rancis.Hu  鍐欓亾锛?br>

 

Actually, My java client is running with the same configuration as the hadoop's 
. The dfs.replication is already set as 2 in my hadoop's configuration.

So i think the dfs.replication is already overrided by my configuration in 
hdfs-site.xml. but seems it doesn't work even i overrided the parameter 
evidently.

 

 

鍙戜欢浜?span lang="EN-US">: 袝屑械谢褜褟薪芯胁 袘芯褉懈褋 [mailto:emelya...@post.km.ru] 
鍙戦€佹椂闂?span lang="EN-US">: Monday, July 01, 2013 15:18
鏀朵欢浜?span lang="EN-US">: user@hadoop.apache.org
涓婚: Re: a question about dfs.replication

 

On 01.07.2013 10:19, Francis.Hu wrote:

Hi, All

 

I am installing a cluster with Hadoop 2.0.5-alpha. I have one namenode and two 
datanodes. The dfs.replication is set as 2 in hdfs-site.xml. After all 
configuration work is done, I started all nodes. Then I saved a file into HDFS 
through java client. nOW I can access hdfs web page: x.x.x.x:50070,and also see 
the file is already listed in the hdfs list.

My question is:  The replication column in HDFS web page is showing as 3, not 
2.  Does anyone know What the problem is?

 

---Actual setting of hdfs-site.xml



dfs.replication

2



 

After that, I typed dfsamdin command to check the file:

hdfs fsck /test3/

The result of above command:

/test3/hello005.txt:  Under replicated 
BP-609310498-192.168.219.129-1372323727200:blk_-1069303317294683372_1006. 
Target Replicas is 3 but found 2 replica(s).

Status: HEALTHY

 Total size:35 B

 Total dirs:1

 Total files:   1

 Total blocks (validated):  1 (avg. block size 35 B)

 Minimally replicated blocks:   1 (100.0 %)

 Over-replicated blocks:0 (0.0 %)

 Under-replicated blocks:   1 (100.0 %)

 Mis-replicated blocks: 0 (0.0 %)

 Default replication factor:2

 Average block replication: 2.0

 Corrupt blocks:0

 Missing replicas:  1 (33.32 %)

 Number of data-nodes:  3

 Number of racks:   1

FSCK ended at Sat Jun 29 16:51:37 CST 2013 in 6 milliseconds

 

 

Thanks,

Francis Hu

 

If I'm not mistaking "dfs.replication" parameter in config sets only default 
replication factor, which can be overrided when putting file to hdfs.

<><>

答复: a question about dfs.replication

2013-07-01 Thread Francis . Hu
 

Actually, My java client is running with the same configuration as the hadoop's 
. The dfs.replication is already set as 2 in my hadoop's configuration.

So i think the dfs.replication is already overrided by my configuration in 
hdfs-site.xml. but seems it doesn't work even i overrided the parameter 
evidently.

 

 

发件人: Емельянов Борис [mailto:emelya...@post.km.ru] 
发送时间: Monday, July 01, 2013 15:18
收件人: user@hadoop.apache.org
主题: Re: a question about dfs.replication

 

On 01.07.2013 10:19, Francis.Hu wrote:

Hi, All

 

I am installing a cluster with Hadoop 2.0.5-alpha. I have one namenode and two 
datanodes. The dfs.replication is set as 2 in hdfs-site.xml. After all 
configuration work is done, I started all nodes. Then I saved a file into HDFS 
through java client. nOW I can access hdfs web page: x.x.x.x:50070,and also see 
the file is already listed in the hdfs list.

My question is:  The replication column in HDFS web page is showing as 3, not 
2.  Does anyone know What the problem is?

 

---Actual setting of hdfs-site.xml



dfs.replication

2



 

After that, I typed dfsamdin command to check the file:

hdfs fsck /test3/

The result of above command:

/test3/hello005.txt:  Under replicated 
BP-609310498-192.168.219.129-1372323727200:blk_-1069303317294683372_1006. 
Target Replicas is 3 but found 2 replica(s).

Status: HEALTHY

 Total size:35 B

 Total dirs:1

 Total files:   1

 Total blocks (validated):  1 (avg. block size 35 B)

 Minimally replicated blocks:   1 (100.0 %)

 Over-replicated blocks:0 (0.0 %)

 Under-replicated blocks:   1 (100.0 %)

 Mis-replicated blocks: 0 (0.0 %)

 Default replication factor:2

 Average block replication: 2.0

 Corrupt blocks:0

 Missing replicas:  1 (33.32 %)

 Number of data-nodes:  3

 Number of racks:   1

FSCK ended at Sat Jun 29 16:51:37 CST 2013 in 6 milliseconds

 

 

Thanks,

Francis Hu

 

If I'm not mistaking "dfs.replication" parameter in config sets only default 
replication factor, which can be overrided when putting file to hdfs.



a question about dfs.replication

2013-06-30 Thread Francis . Hu
Hi, All

 

I am installing a cluster with Hadoop 2.0.5-alpha. I have one namenode and
two datanodes. The dfs.replication is set as 2 in hdfs-site.xml. After all
configuration work is done, I started all nodes. Then I saved a file into
HDFS through java client. nOW I can access hdfs web page: x.x.x.x:50070,and
also see the file is already listed in the hdfs list.

My question is:  The replication column in HDFS web page is showing as 3,
not 2.  Does anyone know What the problem is?

 

---Actual setting of hdfs-site.xml



dfs.replication

2



 

After that, I typed dfsamdin command to check the file:

hdfs fsck /test3/

The result of above command:

/test3/hello005.txt:  Under replicated
BP-609310498-192.168.219.129-1372323727200:blk_-1069303317294683372_1006.
Target Replicas is 3 but found 2 replica(s).

Status: HEALTHY

 Total size:35 B

 Total dirs:1

 Total files:   1

 Total blocks (validated):  1 (avg. block size 35 B)

 Minimally replicated blocks:   1 (100.0 %)

 Over-replicated blocks:0 (0.0 %)

 Under-replicated blocks:   1 (100.0 %)

 Mis-replicated blocks: 0 (0.0 %)

 Default replication factor:2

 Average block replication: 2.0

 Corrupt blocks:0

 Missing replicas:  1 (33.32 %)

 Number of data-nodes:  3

 Number of racks:   1

FSCK ended at Sat Jun 29 16:51:37 CST 2013 in 6 milliseconds

 

 

Thanks,

Francis Hu