Hi everyone,
I did a some deeper investigation this morning and found out the following:
One of my crawlers created/opened a file using the create method and the namenode allocates appropriate blocks for this file. Then another crawler opens a file with the same file name (no exception is thrown). The namenode allocates again appropriate blocks for this file and seems to invalidate the ones from the first create. In the meantime the first crawler puts same data into the stream without any exceptions, except when the crawler tries to close the stream - then the exceptions below/previous posts are thrown.
Can anybody verify that behavior?
Thx and have a great weekend.

Cu on the 'net,
                     Bye - bye,

                                <<<<< André <<<< >>>> èrbnA >>>>>

Chris K Wensel wrote:
I think I've seen that in EC2 where the group wasn't authorized to connect to itself. but it was obvious things were wrong as no nodes where showing up as they booted.

since then, I may have seen something similar intermittently in EC2 on 0.16.0, but many other issues prevented me from pursuing it.

bottom line it looks like a network issue.

On Feb 29, 2008, at 2:48 PM, André Martin wrote:

Also, before the "Could not complete write to file" exception shows up - I see the following exception in my logs:
java.io.IOException: Could not get block locations. Aborting...
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.processDatanodeError(DFSClient.java:1824) at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.access$1100(DFSClient.java:1479) at org.apache.hadoop.dfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:1571)
André Martin wrote:
Hi everyone,
I'm seeing the above exception on my DFS clients:
org.apache.hadoop.ipc.RemoteException: java.io.IOException: Could not complete write to file /seDNS/mapred-out/17A83EC8CC5DD15549657DF36CA9F3236EC121DB/acengineering-kz-20080229232339788-756.dns by DFSClient_1073032245
   at org.apache.hadoop.dfs.NameNode.complete(NameNode.java:341)
   at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
   at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
   at java.lang.reflect.Method.invoke(Unknown Source)
   at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409)
   at org.apache.hadoop.ipc.Server$Handler.run(Server.java:910)
   at org.apache.hadoop.ipc.Client.call(Client.java:512)
   at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
   at org.apache.hadoop.dfs.$Proxy0.complete(Unknown Source)
   at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
   at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
   at java.lang.reflect.Method.invoke(Unknown Source)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
   at org.apache.hadoop.dfs.$Proxy0.complete(Unknown Source)
at org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:2288) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:51) at org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:67)
Any idea why this exception is thrown?
Thx in advance.

Cu on the 'net,
                      Bye - bye,

                                 <<<<< André <<<< >>>> èrbnA >>>>>



Chris K Wensel
[EMAIL PROTECTED]
http://chris.wensel.net/





Reply via email to