Not sure if anyone else answered... 1. You need to run hadoop dfsadmin -finalizeUpgrade. Be careful, because you can't go back once you do this.
http://wiki.apache.org/hadoop/Hadoop_Upgrade I don't know about 2. -Michael On 12/3/08 5:49 PM, "Songting Chen" <[EMAIL PROTECTED]> wrote: 1. The namenode webpage shows: Upgrades: Upgrade for version -18 has been completed. Upgrade is not finalized. 2. SequenceFile.Writer failed when trying to creating a new file with the following error: (we have two HaDoop clusters, both have issue 1; one has issue 2, but the other is fine on issue 2). Any idea what's going on? Thanks, -Songting java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:198) at org.apache.hadoop.hdfs.DFSClient.access$600(DFSClient.java:65) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.closeInternal(DFSClient.java:3084) at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.close(DFSClient.java:3053) at org.apache.hadoop.hdfs.DFSClient$LeaseChecker.close(DFSClient.java:942) at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:210) at org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:243) at org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:1413) at org.apache.hadoop.fs.FileSystem.closeAll(FileSystem.java:236) at org.apache.hadoop.fs.FileSystem$ClientFinalizer.run(FileSystem.java:221)