[
https://issues.apache.org/jira/browse/HADOOP-2569?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12557563#action_12557563
]
enis edited comment on HADOOP-2569 at 1/10/08 2:00 AM:
----------------------------------------------------------------
Well it seems to have stuck in lines below because the log after calling
initFiles() is not printed.
{code}
LOG.info("Creating " + numFiles + " file(s) in " + multiFileDir);
for(int i=0; i<numFiles ;i++) {
Path path = new Path(multiFileDir, "file_" + i);
FSDataOutputStream out = fs.create(path);
if (numBytes == -1) {
numBytes = rand.nextInt(MAX_BYTES);
}
for(int j=0; j< numBytes; j++) {
out.write(rand.nextInt());
}
out.close();
if(LOG.isDebugEnabled()) {
LOG.debug("Created file " + path + " with length " + numBytes);
}
lengths.put(path.getName(), new Long(numBytes));
}
{code}
I assume there is some space related problem in the build server. The test log
reports many other problems such as :
{noformat}
...
[junit] 2008-01-09 11:35:47,543 WARN dfs.DataNode
(DataNode.java:copyBlock(1144)) - Got exception while serving
blk_-4990242691406600933 to 127.0.0.1:35838: java.io.IOException: Block
blk_-4990242691406600933 is not valid.
[junit] at
org.apache.hadoop.dfs.FSDataset.getBlockFile(FSDataset.java:549)
[junit] at
org.apache.hadoop.dfs.FSDataset.getMetaFile(FSDataset.java:466)
[junit] at
org.apache.hadoop.dfs.FSDataset.getMetaDataInputStream(FSDataset.java:480)
[junit] at
org.apache.hadoop.dfs.DataNode$BlockSender.<init>(DataNode.java:1298)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.copyBlock(DataNode.java:1114)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:877)
[junit] at java.lang.Thread.run(Thread.java:595)
...
[junit] 2008-01-09 11:35:56,574 ERROR dfs.DataNode (DataNode.java:run(1738)) -
Exception: java.lang.reflect.UndeclaredThrowableException
[junit] at org.apache.hadoop.dfs.$Proxy1.blockReport(Unknown Source)
[junit] at
org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:616)
[junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1736)
[junit] at java.lang.Thread.run(Thread.java:595)
[junit] Caused by: java.lang.InterruptedException
[junit] at java.lang.Object.wait(Native Method)
[junit] at org.apache.hadoop.ipc.Client.call(Client.java:504)
[junit] at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
[junit] ... 4 more
...
[junit] 2008-01-09 11:37:31,788 WARN dfs.DataNode
(DataNode.java:sendChunk(1401)) - Could not read checksum for data at offset
1536 for block blk_-5451683627314743712 got : java.io.EOFException
[junit] at java.io.DataInputStream.readFully(DataInputStream.java:178)
[junit] at
org.apache.hadoop.dfs.DataNode$BlockSender.sendChunk(DataNode.java:1399)
[junit] at
org.apache.hadoop.dfs.DataNode$BlockSender.sendBlock(DataNode.java:1449)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:920)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:865)
[junit] at java.lang.Thread.run(Thread.java:595)
...
[junit] 2008-01-09 12:36:51,570 WARN fs.AllocatorPerContext
(LocalDirAllocator.java:createPath(256)) -
org.apache.hadoop.util.DiskChecker$DiskErrorException: directory is not
writable: build/test/temp/tmp4
[junit] at
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:85)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createPath(LocalDirAllocator.java:253)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:301)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForWrite(LocalDirAllocator.java:326)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocator.java:155)
[junit] at
org.apache.hadoop.fs.TestLocalDirAllocator.createTempFile(TestLocalDirAllocator.java:77)
...
{noformat}
can you check the server whether there is a space problem or the directories
are writable (I don't have a hudson account).
was (Author: enis):
Well it seems to have stuck in lines below because the log after calling
initFiles() is not printed.
{code}
LOG.info("Creating " + numFiles + " file(s) in " + multiFileDir);
for(int i=0; i<numFiles ;i++) {
Path path = new Path(multiFileDir, "file_" + i);
FSDataOutputStream out = fs.create(path);
if (numBytes == -1) {
numBytes = rand.nextInt(MAX_BYTES);
}
for(int j=0; j< numBytes; j++) {
out.write(rand.nextInt());
}
out.close();
if(LOG.isDebugEnabled()) {
LOG.debug("Created file " + path + " with length " + numBytes);
}
lengths.put(path.getName(), new Long(numBytes));
}
{code}
I assume there is some space related problem in the build server. The test log
reports many other problems such as :
{noformat}
...
[junit] 2008-01-09 11:35:47,543 WARN dfs.DataNode
(DataNode.java:copyBlock(1144)) - Got exception while serving
blk_-4990242691406600933 to 127.0.0.1:35838: java.io.IOException: Block
blk_-4990242691406600933 is not valid.
[junit] at
org.apache.hadoop.dfs.FSDataset.getBlockFile(FSDataset.java:549)
[junit] at
org.apache.hadoop.dfs.FSDataset.getMetaFile(FSDataset.java:466)
[junit] at
org.apache.hadoop.dfs.FSDataset.getMetaDataInputStream(FSDataset.java:480)
[junit] at
org.apache.hadoop.dfs.DataNode$BlockSender.<init>(DataNode.java:1298)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.copyBlock(DataNode.java:1114)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:877)
[junit] at java.lang.Thread.run(Thread.java:595)
...
[junit] 2008-01-09 11:35:56,574 ERROR dfs.DataNode (DataNode.java:run(1738)) -
Exception: java.lang.reflect.UndeclaredThrowableException
[junit] at org.apache.hadoop.dfs.$Proxy1.blockReport(Unknown Source)
[junit] at
org.apache.hadoop.dfs.DataNode.offerService(DataNode.java:616)
[junit] at org.apache.hadoop.dfs.DataNode.run(DataNode.java:1736)
[junit] at java.lang.Thread.run(Thread.java:595)
[junit] Caused by: java.lang.InterruptedException
[junit] at java.lang.Object.wait(Native Method)
[junit] at org.apache.hadoop.ipc.Client.call(Client.java:504)
[junit] at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
[junit] ... 4 more
...
[junit] 2008-01-09 11:37:31,788 WARN dfs.DataNode
(DataNode.java:sendChunk(1401)) - Could not read checksum for data at offset
1536 for block blk_-5451683627314743712 got : java.io.EOFException
[junit] at java.io.DataInputStream.readFully(DataInputStream.java:178)
[junit] at
org.apache.hadoop.dfs.DataNode$BlockSender.sendChunk(DataNode.java:1399)
[junit] at
org.apache.hadoop.dfs.DataNode$BlockSender.sendBlock(DataNode.java:1449)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.readBlock(DataNode.java:920)
[junit] at
org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:865)
[junit] at java.lang.Thread.run(Thread.java:595)
...
[junit] 2008-01-09 12:36:51,570 WARN fs.AllocatorPerContext
(LocalDirAllocator.java:createPath(256)) -
org.apache.hadoop.util.DiskChecker$DiskErrorException: directory is not
writable: build/test/temp/tmp4
[junit] at
org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:85)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createPath(LocalDirAllocator.java:253)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:301)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForWrite(LocalDirAllocator.java:326)
[junit] at
org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocator.java:155)
[junit] at
org.apache.hadoop.fs.TestLocalDirAllocator.createTempFile(TestLocalDirAllocator.java:77)
...
{noformat}
can you check for the server whether there is a space problem or the
directories are writable (I don't have a hudson account).
> Unit test times out on Solaris nightly build: mapred.TestMultiFileInputFormat
> -----------------------------------------------------------------------------
>
> Key: HADOOP-2569
> URL: https://issues.apache.org/jira/browse/HADOOP-2569
> Project: Hadoop
> Issue Type: Bug
> Components: mapred
> Affects Versions: 0.16.0
> Environment: solaris
> Reporter: Mukund Madhugiri
> Assignee: Enis Soztutar
> Priority: Blocker
> Fix For: 0.16.0
>
>
> Unit test failed in the nightly build:
> org.apache.hadoop.mapred.TestMultiFileInputFormat.unknown
> junit.framework.AssertionFailedError: Timeout occurred
> Logs are at:
> http://lucene.zones.apache.org:8080/hudson/job/Hadoop-Nightly/361/
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.