Daniel, I see that in previous hdfs report, you had: hosta.subdom1.tld1, but now you have feynman.cids.ca. What is the content of your /etc/hosts file, and output of $hostname command?
2013/12/3 Daniel Savard <daniel.sav...@gmail.com> > I did that more than once, I just retry it from the beginning. I zapped > the directories and recreated them with hdfs namenode -format and restarted > HDFS and I am still getting the very same error. > > I have posted previously the report. Is there anything in this report that > indicates I am not having enough free space somewhere? That's the only > thing I can see may cause this problem after everything I read on the > subject. I am new to Hadoop and I just want to setup a standalone node for > starting to experiment a while with it before going ahead with a complete > cluster. > > I repost the report for convenience: > > Configured Capacity: 2939899904 (2.74 GB) > Present Capacity: 534421504 (509.66 MB) > DFS Remaining: 534417408 (509.66 MB) > > DFS Used: 4096 (4 KB) > DFS Used%: 0.00% > Under replicated blocks: 0 > Blocks with corrupt replicas: 0 > Missing blocks: 0 > > ------------------------------------------------- > Datanodes available: 1 (1 total, 0 dead) > > Live datanodes: > Name: 127.0.0.1:50010 (feynman.cids.ca) > Hostname: feynman.cids.ca > Decommission Status : Normal > Configured Capacity: 2939899904 (2.74 GB) > > DFS Used: 4096 (4 KB) > Non DFS Used: 2405478400 (2.24 GB) > DFS Remaining: 534417408 (509.66 MB) > DFS Used%: 0.00% > DFS Remaining%: 18.18% > Last contact: Tue Dec 03 13:37:02 EST 2013 > > > ----------------- > Daniel Savard > > > 2013/12/3 Adam Kawa <kawa.a...@gmail.com> > >> Daniel, >> >> It looks that you can only communicate with NameNode to do >> "metadata-only" operations (e.g. listing, creating a dir, empty file)... >> >> Did you format the NameNode correctly? >> A quite similar issue is described here: >> http://www.manning-sandbox.com/thread.jspa?messageID=126741. The last >> reply says: "The most common is that you have reformatted the namenode >> leaving it in an inconsistent state. The most common solution is to stop >> dfs, remove the contents of the dfs directories on all the machines, run >> “hadoop namenode -format” on the controller, then restart dfs. That >> consistently fixes the problem for me. This may be serious overkill but it >> works." >> >> >> 2013/12/3 Daniel Savard <daniel.sav...@gmail.com> >> >>> Thanks Arun, >>> >>> I already read and did everything recommended at the referred URL. There >>> isn't any error message in the logfiles. The only error message appears >>> when I try to put a non-zero file on the HDFS as posted above. Beside that, >>> absolutely nothing in the logs is telling me something is wrong with the >>> configuration so far. >>> >>> Is there some sort of diagnostic tool that can query/ping each server to >>> make sure it responds properly to requests? When trying to put my file, in >>> the datanode log I see nothing, the message appears in the namenode log. Is >>> this the expected behavior or should I see at least some kind of request >>> message in the datanode logfile? >>> >>> >>> ----------------- >>> Daniel Savard >>> >>> >>> 2013/12/2 Arun C Murthy <a...@hortonworks.com> >>> >>>> Daniel, >>>> >>>> Apologies if you had a bad experience. If you can point them out to >>>> us, we'd be more than happy to fix it - alternately, we'd *love* it if you >>>> could help us improve docs too. >>>> >>>> Now, for the problem at hand: >>>> http://wiki.apache.org/hadoop/CouldOnlyBeReplicatedTo is one place to >>>> look. Basically NN cannot find any datanodes. Anything in your NN logs to >>>> indicate trouble? >>>> >>>> Also, pls feel free to open liras with issues you find and we'll help. >>>> >>>> thanks, >>>> Arun >>>> >>>> On Dec 2, 2013, at 8:44 AM, Daniel Savard <daniel.sav...@gmail.com> >>>> wrote: >>>> >>>> André, >>>> >>>> good for you that greedy instructions on the reference page were enough >>>> to setup your cluster. However, read them again and see how many >>>> assumptions are made into them about what you are supposed to already know >>>> and should come without saying more about it. >>>> >>>> I did try the single node setup, it is worst than the cluster setup >>>> regarding the instructions. You are supposed to already have a near working >>>> system as far as I understand the instructions. It is assumed the HDFS is >>>> already setup and working properly. Try to find the instructions to setup >>>> HDFS for version 2.2.0 and you will end up with a lot of inappropriate >>>> instructions about previous version (some properties were renamed). >>>> >>>> It may appear hard at people to say this is toxic, but it is. The first >>>> place a newcomer will go is setup a single node. This will be his starting >>>> point and he will be left with a bunch of a priori and no clue. >>>> >>>> To go back to my very problem at this point: >>>> >>>> 13/12/02 11:34:07 WARN hdfs.DFSClient: DataStreamer Exception >>>> org.apache.hadoop.ipc.RemoteException(java.io.IOException): File >>>> /test._COPYING_ could only be replicated to 0 nodes instead of >>>> minReplication (=1). There are 1 datanode(s) running and no node(s) are >>>> excluded in this operation. >>>> at >>>> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2477) >>>> at >>>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555) >>>> at >>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387) >>>> at >>>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582) >>>> at >>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2048) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2044) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>> at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2042) >>>> >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1347) >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1300) >>>> at >>>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206) >>>> at com.sun.proxy.$Proxy9.addBlock(Unknown Source) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>> at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>> at >>>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186) >>>> at >>>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) >>>> at com.sun.proxy.$Proxy9.addBlock(Unknown Source) >>>> at >>>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330) >>>> at >>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1226) >>>> at >>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078) >>>> at >>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514) >>>> >>>> I can copy an empty file, but as soon as its content is non-zero I am >>>> getting this message. Searching on the message is of no help so far. >>>> >>>> And I skimmed through the cluster instructions and found nothing there >>>> that could help in any way neither. >>>> >>>> >>>> ----------------- >>>> Daniel Savard >>>> >>>> >>>> 2013/12/2 Andre Kelpe <ake...@concurrentinc.com> >>>> >>>>> Hi Daniel, >>>>> >>>>> first of all, before posting to a mailing list, take a deep breath and >>>>> let your frustrations out. Then write the email. Using words like >>>>> "crappy", "toxicware", "nightmare" are not going to help you getting >>>>> useful responses. >>>>> >>>>> While I agree that the docs can be confusing, we should try to stay >>>>> constructive. You haven't mentioned which documentation you are >>>>> using. I found the cluster tutorial sufficient to get me started: >>>>> >>>>> http://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-common/ClusterSetup.html >>>>> >>>>> If you are looking for an easy way to spin up a small cluster with >>>>> hadoop 2.2, try the hadoop2 branch of this vagrant setup: >>>>> >>>>> https://github.com/fs111/vagrant-hadoop-cluster/tree/hadoop2 >>>>> >>>>> - André >>>>> >>>>> On Mon, Dec 2, 2013 at 5:34 AM, Daniel Savard <daniel.sav...@gmail.com> >>>>> wrote: >>>>> > I am trying to configure hadoop 2.2.0 from source code and I found >>>>> the >>>>> > instructions really crappy and incomplete. It is like they were >>>>> written to >>>>> > avoid someone can do the job himself and must contract someone else >>>>> to do it >>>>> > or buy a packaged version. >>>>> > >>>>> > It is about three days I am struggling with this stuff with partial >>>>> success. >>>>> > The documentation is less than clear and most of the stuff out there >>>>> apply >>>>> > to earlier version and they haven't been updated for version 2.2.0. >>>>> > >>>>> > I was able to setup HDFS, however I am still unable to use it. I am >>>>> doing a >>>>> > single node installation and the instruction page doesn't explain >>>>> anything >>>>> > beside telling you to do this and that without documenting what each >>>>> thing >>>>> > is doing and what choices are available and what guidelines you >>>>> should >>>>> > follow. There is even environment variables you are told to set, but >>>>> nothing >>>>> > is said about what they mean and to which value they should be set. >>>>> It seems >>>>> > it assumes prior knowledge of everything about hadoop. >>>>> > >>>>> > Anyone knows a site with proper documentation about hadoop or it's >>>>> hopeless >>>>> > and this whole thing is just a piece of toxicware? >>>>> > >>>>> > I am already looking for alternate solutions to hadoop which for >>>>> sure will >>>>> > be a nightmare to manage and install each time a new version, >>>>> release will >>>>> > become available. >>>>> > >>>>> > TIA >>>>> > ----------------- >>>>> > Daniel Savard >>>>> >>>>> >>>>> >>>>> -- >>>>> André Kelpe >>>>> an...@concurrentinc.com >>>>> http://concurrentinc.com >>>>> >>>> >>>> >>>> -- >>>> Arun C. Murthy >>>> Hortonworks Inc. >>>> http://hortonworks.com/ >>>> >>>> >>>> >>>> CONFIDENTIALITY NOTICE >>>> NOTICE: This message is intended for the use of the individual or >>>> entity to which it is addressed and may contain information that is >>>> confidential, privileged and exempt from disclosure under applicable law. >>>> If the reader of this message is not the intended recipient, you are hereby >>>> notified that any printing, copying, dissemination, distribution, >>>> disclosure or forwarding of this communication is strictly prohibited. If >>>> you have received this communication in error, please contact the sender >>>> immediately and delete it from your system. Thank You. >>> >>> >>> >> >