Hadoop-Fuse issue

2012-04-06 Thread Manu S
Dear All,

I have installed Hadoop-fuse to mount the HDFS filesystem locally for more
flexibility. I could mount the HDFS without any issues.But I am not able to
do any file operations like *delete, copy, move* orations directly. The
directory ownership automatically changed to nobody:nobody while mounting.

[root@namenode ~]# ls -ld /hdfs_mount/
drwxr-xr-x 2 *root root* 4096 Apr  3 16:22 /hdfs_mount/

[root@namenode ~]#  *hadoop-fuse-dfs dfs://namenode:8020 /hdfs_mount/*
INFO fuse_options.c:162 Adding FUSE arg /hdfs_mount/

[root@namenode ~]# ls -ld /hdfs_mount/
drwxrwxr-x 10* nobody nobody* 4096 Apr  5 13:22 /hdfs_mount/


I tried the same with pseudo-distributed node, its working fine. I can do
any normal file operations after mounting the HDFS.

Appreciate your help on the same.

 --
Thanks  Regards

*Manu S*
SI Engineer - OpenSource  HPC
Wipro Infotech
Mob: +91 8861302855Skype: manuspkd
www.opensourcetalk.co.in


Re: map task execution time

2012-04-06 Thread Luca Pireddu

On 04/04/2012 11:20 PM, bikash sharma wrote:

Hi,
Is it possible to get the execution time of the constituent map/reduce
tasks of a MapReduce job (say sort) at the end of a job run?
Preferably, can we obtain this programatically?

Thanks,
Bikash



You could parse the logs inside the job's output directory (output 
dir/_logs)


You may also find this tool to be interesting:
http://www.cs.duke.edu/starfish/

It can extract all that task info for you, after doing a profiling run.

--
Luca Pireddu
CRS4 - Distributed Computing Group
Loc. Pixina Manna Edificio 1
09010 Pula (CA), Italy
Tel: +39 0709250452


Data Node is not Started

2012-04-06 Thread Sujit Dhamale
Hi all,
my DataNode is not started .

even after deleting hadoop*.pid file from /tmp , But still Data node is not
started ,


Hadoop Version: hadoop-1.0.1.tar.gz
Java version : java version 1.6.0_26
Operating System : Ubuntu 11.10


i did below procedure


*hduser@sujit:~/Desktop/hadoop/bin$ jps*
11455 Jps


*hduser@sujit:~/Desktop/hadoop/bin$ start-all.sh*
Warning: $HADOOP_HOME is deprecated.

starting namenode, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-namenode-sujit.out
localhost: starting datanode, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-datanode-sujit.out
localhost: starting secondarynamenode, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-secondarynamenode-sujit.out
starting jobtracker, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-jobtracker-sujit.out
localhost: starting tasktracker, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-tasktracker-sujit.out

*hduser@sujit:~/Desktop/hadoop/bin$ jps*
11528 NameNode
12019 SecondaryNameNode
12355 TaskTracker
12115 JobTracker
12437 Jps


*hduser@sujit:~/Desktop/hadoop/bin$ stop-all.sh*
Warning: $HADOOP_HOME is deprecated.

stopping jobtracker
localhost: stopping tasktracker
stopping namenode
localhost: no datanode to stop
localhost: stopping secondarynamenode


*hduser@sujit:~/Desktop/hadoop/bin$ jps*
13127 Jps


*hduser@sujit:~/Desktop/hadoop/bin$ ls /tmp*
hadoop-hduser-datanode.pid
hsperfdata_hduserkeyring-meecr7
ssh-JXYCAJsX1324
hadoop-hduser-jobtracker.pid
hsperfdata_sujit plugtmp
unity_support_test.0
hadoop-hduser-namenode.pid
Jetty_0_0_0_0_50030_jobyn7qmkpulse-2L9K88eMlGn7
virtual-hduser.Q8j5nJ
hadoop-hduser-secondarynamenode.pid
Jetty_0_0_0_0_50070_hdfsw2cu08   pulse-Ob9vyJcXyHZz
hadoop-hduser-tasktracker.pid
Jetty_0_0_0_0_50090_secondaryy6aanv  pulse-PKdhtXMmr18n

*Deleted *.pid file :)

hduser@sujit:~$ ls /tmp*
hsperfdata_hduserpulse-2L9K88eMlGn7
hsperfdata_sujit pulse-Ob9vyJcXyHZz
Jetty_0_0_0_0_50030_jobyn7qmkpulse-PKdhtXMmr18n
Jetty_0_0_0_0_50070_hdfsw2cu08   ssh-JXYCAJsX1324
Jetty_0_0_0_0_50090_secondaryy6aanv  unity_support_test.0
keyring-meecr7   virtual-hduser.Q8j5nJ
plugtmp





*hduser@sujit:~/Desktop/hadoop$ bin/hadoop namenode -format*
Warning: $HADOOP_HOME is deprecated.

12/04/06 23:23:22 INFO namenode.NameNode: STARTUP_MSG:
/
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = sujit.(null)/127.0.1.1
STARTUP_MSG:   args = [-format]
STARTUP_MSG:   version = 1.0.1
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
/
Re-format filesystem in /app/hadoop/tmp/dfs/name ? (Y or N) Y
12/04/06 23:23:25 INFO util.GSet: VM type   = 32-bit
12/04/06 23:23:25 INFO util.GSet: 2% max memory = 17.77875 MB
12/04/06 23:23:25 INFO util.GSet: capacity  = 2^22 = 4194304 entries
12/04/06 23:23:25 INFO util.GSet: recommended=4194304, actual=4194304
12/04/06 23:23:25 INFO namenode.FSNamesystem: fsOwner=hduser
12/04/06 23:23:25 INFO namenode.FSNamesystem: supergroup=supergroup
12/04/06 23:23:25 INFO namenode.FSNamesystem: isPermissionEnabled=true
12/04/06 23:23:25 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=100
12/04/06 23:23:25 INFO namenode.FSNamesystem: isAccessTokenEnabled=false
accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
12/04/06 23:23:25 INFO namenode.NameNode: Caching file names occuring more
than 10 times
12/04/06 23:23:26 INFO common.Storage: Image file of size 112 saved in 0
seconds.
12/04/06 23:23:26 INFO common.Storage: Storage directory
/app/hadoop/tmp/dfs/name has been successfully formatted.
12/04/06 23:23:26 INFO namenode.NameNode: SHUTDOWN_MSG:
/
SHUTDOWN_MSG: Shutting down NameNode at sujit.(null)/127.0.1.1
/
hduser@sujit:~/Desktop/hadoop$ bin/start-all.sh
Warning: $HADOOP_HOME is deprecated.

starting namenode, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-namenode-sujit.out
localhost: starting datanode, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-datanode-sujit.out
localhost: starting secondarynamenode, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-secondarynamenode-sujit.out
starting jobtracker, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-jobtracker-sujit.out
localhost: starting tasktracker, logging to
/home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-tasktracker-sujit.out


*hduser@sujit:~/Desktop/hadoop$ jps*
14157 JobTracker
14492 Jps
14397 TaskTracker
14063 SecondaryNameNode

Re: Data Node is not Started

2012-04-06 Thread Prashant Kommireddi
Can you check the datanode logs? May  its an incompatible namespace issue.

On Apr 6, 2012, at 11:13 AM, Sujit Dhamale sujitdhamal...@gmail.com wrote:

 Hi all,
 my DataNode is not started .

 even after deleting hadoop*.pid file from /tmp , But still Data node is not
 started ,


 Hadoop Version: hadoop-1.0.1.tar.gz
 Java version : java version 1.6.0_26
 Operating System : Ubuntu 11.10


 i did below procedure


 *hduser@sujit:~/Desktop/hadoop/bin$ jps*
 11455 Jps


 *hduser@sujit:~/Desktop/hadoop/bin$ start-all.sh*
 Warning: $HADOOP_HOME is deprecated.

 starting namenode, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-namenode-sujit.out
 localhost: starting datanode, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-datanode-sujit.out
 localhost: starting secondarynamenode, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-secondarynamenode-sujit.out
 starting jobtracker, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-jobtracker-sujit.out
 localhost: starting tasktracker, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-tasktracker-sujit.out

 *hduser@sujit:~/Desktop/hadoop/bin$ jps*
 11528 NameNode
 12019 SecondaryNameNode
 12355 TaskTracker
 12115 JobTracker
 12437 Jps


 *hduser@sujit:~/Desktop/hadoop/bin$ stop-all.sh*
 Warning: $HADOOP_HOME is deprecated.

 stopping jobtracker
 localhost: stopping tasktracker
 stopping namenode
 localhost: no datanode to stop
 localhost: stopping secondarynamenode


 *hduser@sujit:~/Desktop/hadoop/bin$ jps*
 13127 Jps


 *hduser@sujit:~/Desktop/hadoop/bin$ ls /tmp*
 hadoop-hduser-datanode.pid
 hsperfdata_hduserkeyring-meecr7
 ssh-JXYCAJsX1324
 hadoop-hduser-jobtracker.pid
 hsperfdata_sujit plugtmp
 unity_support_test.0
 hadoop-hduser-namenode.pid
 Jetty_0_0_0_0_50030_jobyn7qmkpulse-2L9K88eMlGn7
 virtual-hduser.Q8j5nJ
 hadoop-hduser-secondarynamenode.pid
 Jetty_0_0_0_0_50070_hdfsw2cu08   pulse-Ob9vyJcXyHZz
 hadoop-hduser-tasktracker.pid
 Jetty_0_0_0_0_50090_secondaryy6aanv  pulse-PKdhtXMmr18n

 *Deleted *.pid file :)

 hduser@sujit:~$ ls /tmp*
 hsperfdata_hduserpulse-2L9K88eMlGn7
 hsperfdata_sujit pulse-Ob9vyJcXyHZz
 Jetty_0_0_0_0_50030_jobyn7qmkpulse-PKdhtXMmr18n
 Jetty_0_0_0_0_50070_hdfsw2cu08   ssh-JXYCAJsX1324
 Jetty_0_0_0_0_50090_secondaryy6aanv  unity_support_test.0
 keyring-meecr7   virtual-hduser.Q8j5nJ
 plugtmp





 *hduser@sujit:~/Desktop/hadoop$ bin/hadoop namenode -format*
 Warning: $HADOOP_HOME is deprecated.

 12/04/06 23:23:22 INFO namenode.NameNode: STARTUP_MSG:
 /
 STARTUP_MSG: Starting NameNode
 STARTUP_MSG:   host = sujit.(null)/127.0.1.1
 STARTUP_MSG:   args = [-format]
 STARTUP_MSG:   version = 1.0.1
 STARTUP_MSG:   build =
 https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
 /
 Re-format filesystem in /app/hadoop/tmp/dfs/name ? (Y or N) Y
 12/04/06 23:23:25 INFO util.GSet: VM type   = 32-bit
 12/04/06 23:23:25 INFO util.GSet: 2% max memory = 17.77875 MB
 12/04/06 23:23:25 INFO util.GSet: capacity  = 2^22 = 4194304 entries
 12/04/06 23:23:25 INFO util.GSet: recommended=4194304, actual=4194304
 12/04/06 23:23:25 INFO namenode.FSNamesystem: fsOwner=hduser
 12/04/06 23:23:25 INFO namenode.FSNamesystem: supergroup=supergroup
 12/04/06 23:23:25 INFO namenode.FSNamesystem: isPermissionEnabled=true
 12/04/06 23:23:25 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=100
 12/04/06 23:23:25 INFO namenode.FSNamesystem: isAccessTokenEnabled=false
 accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
 12/04/06 23:23:25 INFO namenode.NameNode: Caching file names occuring more
 than 10 times
 12/04/06 23:23:26 INFO common.Storage: Image file of size 112 saved in 0
 seconds.
 12/04/06 23:23:26 INFO common.Storage: Storage directory
 /app/hadoop/tmp/dfs/name has been successfully formatted.
 12/04/06 23:23:26 INFO namenode.NameNode: SHUTDOWN_MSG:
 /
 SHUTDOWN_MSG: Shutting down NameNode at sujit.(null)/127.0.1.1
 /
 hduser@sujit:~/Desktop/hadoop$ bin/start-all.sh
 Warning: $HADOOP_HOME is deprecated.

 starting namenode, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-namenode-sujit.out
 localhost: starting datanode, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-datanode-sujit.out
 localhost: starting secondarynamenode, logging to
 /home/hduser/Desktop/hadoop/libexec/../logs/hadoop-hduser-secondarynamenode-sujit.out
 starting jobtracker, logging to
 

Re: Data Node is not Started

2012-04-06 Thread Sujit Dhamale
Below are DataNode logs









hduser@sujit:~/Desktop/hadoop/logs$ cat hadoop-hduser-datanode-sujit.log
2012-04-06 22:11:34,566 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = sujit.(null)/127.0.1.1
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 1.0.1
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
/
2012-04-06 22:11:34,749 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
loaded properties from hadoop-metrics2.properties
2012-04-06 22:11:34,768 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
MetricsSystem,sub=Stats registered.
2012-04-06 22:11:34,769 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
period at 10 second(s).
2012-04-06 22:11:34,769 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
started
2012-04-06 22:11:34,950 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
registered.
2012-04-06 22:11:34,956 WARN
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
exists!
2012-04-06 22:11:36,149 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: localhost/127.0.0.1:54310. Already tried 0 time(s).
2012-04-06 22:11:41,923 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
Incompatible namespaceIDs in /app/hadoop/tmp/dfs/data: namenode namespaceID
= 387652554; datanode namespaceID = 1269725409
at
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:232)
at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:147)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:385)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:299)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1539)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)

2012-04-06 22:11:41,924 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/
SHUTDOWN_MSG: Shutting down DataNode at sujit.(null)/127.0.1.1
/
2012-04-06 22:30:04,713 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = sujit.(null)/127.0.1.1
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 1.0.1
STARTUP_MSG:   build =
https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
/
2012-04-06 22:30:04,870 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
loaded properties from hadoop-metrics2.properties
2012-04-06 22:30:04,883 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
MetricsSystem,sub=Stats registered.
2012-04-06 22:30:04,884 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
period at 10 second(s).
2012-04-06 22:30:04,884 INFO
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
started
2012-04-06 22:30:05,046 INFO
org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
registered.
2012-04-06 22:30:05,051 WARN
org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
exists!
2012-04-06 22:30:06,273 INFO org.apache.hadoop.ipc.Client: Retrying connect
to server: localhost/127.0.0.1:54310. Already tried 0 time(s).
2012-04-06 22:30:11,073 ERROR
org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
Incompatible namespaceIDs in /app/hadoop/tmp/dfs/data: namenode namespaceID
= 361514980; datanode namespaceID = 1269725409
at
org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:232)
at
org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:147)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:385)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:299)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
at
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
at

Re: Data Node is not Started

2012-04-06 Thread Arpit Gupta
according to the logs the namespace id in the datanode data directories is 
incompatible.

Since you formatted the namenode these id's do not match. Clean up the contents 
of the data dir (/app/hadoop/tmp/dfs/data) and then start the datanode.

--
Arpit Gupta
Hortonworks Inc.
http://hortonworks.com/

On Apr 6, 2012, at 11:27 AM, Sujit Dhamale wrote:

 Below are DataNode logs
 
 
 
 
 
 
 
 
 
 hduser@sujit:~/Desktop/hadoop/logs$ cat hadoop-hduser-datanode-sujit.log
 2012-04-06 22:11:34,566 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
 /
 STARTUP_MSG: Starting DataNode
 STARTUP_MSG:   host = sujit.(null)/127.0.1.1
 STARTUP_MSG:   args = []
 STARTUP_MSG:   version = 1.0.1
 STARTUP_MSG:   build =
 https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
 /
 2012-04-06 22:11:34,749 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
 loaded properties from hadoop-metrics2.properties
 2012-04-06 22:11:34,768 INFO
 org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
 MetricsSystem,sub=Stats registered.
 2012-04-06 22:11:34,769 INFO
 org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
 period at 10 second(s).
 2012-04-06 22:11:34,769 INFO
 org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
 started
 2012-04-06 22:11:34,950 INFO
 org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
 registered.
 2012-04-06 22:11:34,956 WARN
 org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
 exists!
 2012-04-06 22:11:36,149 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: localhost/127.0.0.1:54310. Already tried 0 time(s).
 2012-04-06 22:11:41,923 ERROR
 org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
 Incompatible namespaceIDs in /app/hadoop/tmp/dfs/data: namenode namespaceID
 = 387652554; datanode namespaceID = 1269725409
at
 org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:232)
at
 org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:147)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:385)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:299)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1582)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1539)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
at
 org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)
 
 2012-04-06 22:11:41,924 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
 /
 SHUTDOWN_MSG: Shutting down DataNode at sujit.(null)/127.0.1.1
 /
 2012-04-06 22:30:04,713 INFO
 org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
 /
 STARTUP_MSG: Starting DataNode
 STARTUP_MSG:   host = sujit.(null)/127.0.1.1
 STARTUP_MSG:   args = []
 STARTUP_MSG:   version = 1.0.1
 STARTUP_MSG:   build =
 https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0 -r
 1243785; compiled by 'hortonfo' on Tue Feb 14 08:15:38 UTC 2012
 /
 2012-04-06 22:30:04,870 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
 loaded properties from hadoop-metrics2.properties
 2012-04-06 22:30:04,883 INFO
 org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
 MetricsSystem,sub=Stats registered.
 2012-04-06 22:30:04,884 INFO
 org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
 period at 10 second(s).
 2012-04-06 22:30:04,884 INFO
 org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system
 started
 2012-04-06 22:30:05,046 INFO
 org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
 registered.
 2012-04-06 22:30:05,051 WARN
 org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already
 exists!
 2012-04-06 22:30:06,273 INFO org.apache.hadoop.ipc.Client: Retrying connect
 to server: localhost/127.0.0.1:54310. Already tried 0 time(s).
 2012-04-06 22:30:11,073 ERROR
 org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException:
 Incompatible namespaceIDs in /app/hadoop/tmp/dfs/data: namenode namespaceID
 = 361514980; datanode namespaceID = 1269725409
at
 org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:232)
at
 

Namespace logs : a common issue?

2012-04-06 Thread JAX
Hi guys : I'm noticing that namespace conflicts or differences are a common 
theme in hadoop both in my experience and now on this list Serv.

Does anyone have any thoughts on why this is such a common issue and how it 
will be dealt with in new releases?

Jay Vyas 
MMSB
UCHC

Re: Namespace logs : a common issue?

2012-04-06 Thread Harsh J
If you are talking about the Incompatible namespace ID issues, then
look at https://issues.apache.org/jira/browse/HDFS-107 for a
discussion on the topic.

It is a common issue with newcomers as they end up formatting at the
first sight of trouble, or skip reading most of the tutorial that
otherwise helps get configurations right.

On Sat, Apr 7, 2012 at 12:31 AM, JAX jayunit...@gmail.com wrote:
 Hi guys : I'm noticing that namespace conflicts or differences are a common 
 theme in hadoop both in my experience and now on this list Serv.

 Does anyone have any thoughts on why this is such a common issue and how it 
 will be dealt with in new releases?

 Jay Vyas
 MMSB
 UCHC



-- 
Harsh J


Reading properties file from command line + passing parameters from main to mapper/reducer?

2012-04-06 Thread Sky USC




Dear friends I am new to hadoop on aws. I am using AWS Elastic mapreduce. I am 
trying to convert an old java program to elastic mapreduce. I would appreciate 
your help. My question is about:
1. How do I pass a Properties file which i used to pass via command line 
paramters - to elastic mapreduce?
2. how do I pass parameters from this properties file from main() to the Mapper 
and Reducer? Details:
1. properties file: 
My program used to read a properties file as follows:
org.apache.commons.configuration.Configuration config = new 
PropertiesConfiguration(my_app.properties);
How do I read this in Amazon EMR?
I tried launching the program with :
jar file: s3n://my.bucket.name/myjar-job.jar
args: s3n://my.bucket.name/my_app.properties I get an stderr of:
org.apache.commons.configuration.ConfigurationException: Cannot locate 
configuration source s3n://my.bucket.name/my_app.properties In order to debug, 
I tried to do the following:
String c = FileUtils.readFileToString(new File(remainingArgs[0]));
I got an exception:
Exception in thread main java.io.FileNotFoundException: File 
's3n:/my.bucket.name/my_app.propertiess' does not exist
Notice that the file name - even though I entered s3n:// is shown as s3n:/ - 
could the // be getting dropped?  2. Passing the properties file from my main() 
to Mapper and Reducer. How do I do that? Or pass other generic parameters?
-Sky  

Reading properties file from command line + passing parameters from main to mapper/reducer?

2012-04-06 Thread Sky USC





Dear friends
 
I am new to hadoop on aws. I am using AWS Elastic mapreduce. I am trying to 
convert an old java program to elastic mapreduce. I would appreciate your help.
 
My question is about:
1. How do I pass a Properties file which i used to pass via command line 
paramters - to elastic mapreduce?
2. how do I pass parameters from this properties file from main() to the Mapper 
and Reducer?
 
Details:
1. properties file: 
My program used to read a properties file as follows:
org.apache.commons.configuration.Configuration config = new 
PropertiesConfiguration(my_app.properties);
How do I read this in Amazon EMR?
I tried launching the program with :
jar file: s3n://my.bucket.name/myjar-job.jar
args: s3n://my.bucket.name/my_app.properties
 
I get an stderr of:
org.apache.commons.configuration.ConfigurationException: Cannot locate 
configuration source s3n://my.bucket.name/my_app.properties
 
In order to debug, I tried to do the following:
String c = FileUtils.readFileToString(new File(remainingArgs[0]));
I got an exception:
Exception in thread main java.io.FileNotFoundException: File 
's3n:/my.bucket.name/my_app.propertiess' does not exist
Notice that the file name - even though I entered s3n:// is shown as s3n:/ - 
could the // be getting dropped?
 
 
2. Passing the properties file from my main() to Mapper and Reducer. How do I 
do that? Or pass other generic parameters?

-Sky

  

Reading properties file from command line + passing parameters from main to mapper/reducer?

2012-04-06 Thread Sky USC

Dear friends
 
I am new to hadoop on aws. I am using AWS Elastic mapreduce. I am trying to 
convert an old java program to elastic mapreduce. I would appreciate your help.
 
My question is about:
1. How do I pass a Properties file which i used to pass via command line 
paramters - to elastic mapreduce?
2. how do I pass parameters from this properties file from main() to the Mapper 
and Reducer?
 
Details:
1. properties file: 
My program used to read a properties file as follows:
org.apache.commons.configuration.Configuration config = new 
PropertiesConfiguration(my_app.properties);
How do I read this in Amazon EMR?
I tried launching the program with :
jar file: s3n://my.bucket.name/myjar-job.jar
args: s3n://my.bucket.name/my_app.properties
 
I get an stderr of:
org.apache.commons.configuration.ConfigurationException: Cannot locate 
configuration source s3n://my.bucket.name/my_app.properties
 
In order to debug, I tried to do the following:
String c = FileUtils.readFileToString(new File(remainingArgs[0]));
I got an exception:
Exception in thread main java.io.FileNotFoundException: File 
's3n:/my.bucket.name/my_app.propertiess' does not exist
Notice that the file name - even though I entered s3n:// is shown as s3n:/ - 
could the // be getting dropped?
 
 
2. Passing the properties file from my main() to Mapper and Reducer. How do I 
do that? Or pass other generic parameters?

-Sky