What is the o/p of command

hdfs dfsadmin -upgradeProgress status

If it says upgrade is complete then you can do some sanity check by hdfs fsck.

Stop the servers by stop-dfs.sh and then rollback by command
start-dfs.sh -rollback

On 9/18/14, sam liu <samliuhad...@gmail.com> wrote:
> Thanks for your comment!
>
> I can upgrade from 2.2.0 to 2.4.1 using command 'start-dfs.sh -upgrade',
> however failed to rollback from 2.4.1 to 2.2.0 using command 'start-dfs.sh
> -rollback': the namenode always stays on safe mode(awaiting reported blocks
> (0/315)).
>
> Why?
>
> 2014-09-18 1:51 GMT-07:00 Susheel Kumar Gadalay <skgada...@gmail.com>:
>
>> You have to upgrade both name node and data node.
>>
>> Better issue start-dfs.sh -upgrade.
>>
>> Check whether current and previous directories are present in both
>> dfs.namenode.name.dir and dfs.datanode.data.dir directory.
>>
>> On 9/18/14, sam liu <samliuhad...@gmail.com> wrote:
>> > Hi Expert,
>> >
>> > Below are my steps and is it a hadoop bug or did I miss any thing?
>> Thanks!
>> >
>> > Step:
>> > [A] Upgrade
>> > 1. Install Hadoop 2.2.0 cluster
>> > 2. Stop Hadoop services
>> > 3. Replace 2.2.0 binaries with 2.4.1 binaries
>> > 4. Start datanodes: $HADOOP_HOME/sbin/hadoop-daemon.sh start datanode
>> > 5. Start namenode with option upgrade:
>> > $HADOOP_HOME/sbin/hadoop-daemon.sh
>> > start namenode -upgrade
>> > 6. Start secondary namenode, tasktracker and jobtracker
>> >
>> > Result:
>> >
>> >     Whole upgrade process could be completed successfully.
>> >
>> > [B] Rollback
>> > 1. Stop all hadoop services
>> > 2. Replace 2.4.1 binaries with 2.2.0 binaries
>> > 3. Start datanodes: $HADOOP_HOME/sbin/hadoop-daemon.sh start datanode
>> > 4. Start namenode with option upgrade:
>> > $HADOOP_HOME/sbin/hadoop-daemon.sh
>> > start namenode -rollback
>> >
>> > Result:
>> >
>> >     Namenode service could be started
>> >     Datanodes failed with exception:
>> >     Issue: DataNode failed with following exception
>> >     2014-09-17 11:04:51,416 INFO
>> > org.apache.hadoop.hdfs.server.common.Storage: Lock on
>> > /hadoop/hdfs/data/in_use.lock acquired by nodename
>> > 817443@shihc071-public
>> >     2014-09-17 11:04:51,418 FATAL
>> > org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed
>> for
>> > block pool Block pool BP-977402492-9.181.64.185-1410497086460 (storage
>> id )
>> > service to hostname/ip:9000
>> >     org.apache.hadoop.hdfs.server.common.IncorrectVersionException:
>> > Unexpected version of storage directory /hadoop/hdfs/data. Reported:
>> > -55.
>> > Expecting = -47.
>> >     at
>> >
>> org.apache.hadoop.hdfs.server.common.Storage.setLayoutVersion(Storage.java:1082)
>> >     at
>> >
>> org.apache.hadoop.hdfs.server.datanode.DataStorage.setFieldsFromProperties(DataStorage.java:302)
>> >     at
>> >
>> org.apache.hadoop.hdfs.server.common.Storage.readProperties(Storage.java:921)
>> >
>>
>

Reply via email to