[jira] [Updated] (HADOOP-9103) UTF8 class does not properly decode Unicode characters outside the basic multilingual plane

2012-11-28 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HADOOP-9103:


Attachment: hadoop-9103.txt

Attached patch should fix this issue. I also amended the javadoc for UTF8 to 
indicate that it encodes CESU-8

> UTF8 class does not properly decode Unicode characters outside the basic 
> multilingual plane
> ---
>
> Key: HADOOP-9103
> URL: https://issues.apache.org/jira/browse/HADOOP-9103
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: io
>Affects Versions: 0.20.1
> Environment: SUSE LINUX
>Reporter: yixiaohua
>Assignee: Todd Lipcon
> Attachments: FSImage.java, hadoop-9103.txt, ProblemString.txt, 
> TestUTF8AndStringGetBytes.java, TestUTF8AndStringGetBytes.java
>
>   Original Estimate: 12h
>  Remaining Estimate: 12h
>
> this the log information  of the  exception  from the SecondaryNameNode: 
> 2012-03-28 00:48:42,553 ERROR 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: 
> java.io.IOException: Found lease for
>  non-existent file 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/@???
> ??tor.qzone.qq.com/keypart-00174
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFilesUnderConstruction(FSImage.java:1211)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:959)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.doMerge(SecondaryNameNode.java:589)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.access$000(SecondaryNameNode.java:473)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doMerge(SecondaryNameNode.java:350)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(SecondaryNameNode.java:314)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNameNode.java:225)
> at java.lang.Thread.run(Thread.java:619)
> this is the log information  about the file from namenode:
> 2012-03-28 00:32:26,528 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: ugi=boss,boss 
> ip=/10.131.16.34cmd=create  
> src=/user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 dst=null
> perm=boss:boss:rw-r--r--
> 2012-03-28 00:37:42,387 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> NameSystem.allocateBlock: 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174. 
> blk_2751836614265659170_184668759
> 2012-03-28 00:37:42,696 INFO org.apache.hadoop.hdfs.StateChange: DIR* 
> NameSystem.completeFile: file 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 is closed by 
> DFSClient_attempt_201203271849_0016_r_000174_0
> 2012-03-28 00:37:50,315 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: ugi=boss,boss 
> ip=/10.131.16.34cmd=rename  
> src=/user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 
> dst=/user/boss/pgv/fission/task16/split/  @?
> tor.qzone.qq.com/keypart-00174  perm=boss:boss:rw-r--r--
> after check the code that save FSImage,I found there are a problem that maybe 
> a bug of HDFS Code,I past below:
> -this is the saveFSImage method  in  FSImage.java, I make some 
> mark at the problem code
> /**
>* Save the contents of the FS image to the file.
>*/
>   void saveFSImage(File newFile) throws IOException {
> FSNamesystem fsNamesys = FSNamesystem.getFSNamesystem();
> FSDirectory fsDir = fsNamesys.dir;
> long startTime = FSNamesystem.now();
> //
> // Write out data
> //
> DataOutputStream out = new DataOutputStream(
> new BufferedOutputStream(
>  new 
> FileOutputStream(newFile)));
> try {
>   .
> 
>   // save the rest of the nodes
>   saveImage(strbuf, 0, fsDir.rootDir, out);--problem
>   fsNamesys.saveFilesUnderConstruction(out);--problem  
> detail is below
>   strbuf = null;
> } finally {
>   out.close();
> }
> LOG.info("Image file of size " + newFile.length() + " saved in " 
> + (FSNamesystem.now() - startTime)/1000 + " seconds.");
>   }
>  /**
>* Save file tree image starting from the given root.
>

[jira] [Updated] (HADOOP-9103) UTF8 class does not properly decode Unicode characters outside the basic multilingual plane

2012-11-28 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HADOOP-9103:


Summary: UTF8 class does not properly decode Unicode characters outside the 
basic multilingual plane  (was: when save FSImage  ,HDFS( or  SecondaryNameNode 
or FSImage)can't handle some file whose file name has some special messy 
code(乱码))

> UTF8 class does not properly decode Unicode characters outside the basic 
> multilingual plane
> ---
>
> Key: HADOOP-9103
> URL: https://issues.apache.org/jira/browse/HADOOP-9103
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: io
>Affects Versions: 0.20.1
> Environment: SUSE LINUX
>Reporter: yixiaohua
>Assignee: Todd Lipcon
> Attachments: FSImage.java, hadoop-9103.txt, ProblemString.txt, 
> TestUTF8AndStringGetBytes.java, TestUTF8AndStringGetBytes.java
>
>   Original Estimate: 12h
>  Remaining Estimate: 12h
>
> this the log information  of the  exception  from the SecondaryNameNode: 
> 2012-03-28 00:48:42,553 ERROR 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: 
> java.io.IOException: Found lease for
>  non-existent file 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/@???
> ??tor.qzone.qq.com/keypart-00174
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFilesUnderConstruction(FSImage.java:1211)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:959)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.doMerge(SecondaryNameNode.java:589)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.access$000(SecondaryNameNode.java:473)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doMerge(SecondaryNameNode.java:350)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(SecondaryNameNode.java:314)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNameNode.java:225)
> at java.lang.Thread.run(Thread.java:619)
> this is the log information  about the file from namenode:
> 2012-03-28 00:32:26,528 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: ugi=boss,boss 
> ip=/10.131.16.34cmd=create  
> src=/user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 dst=null
> perm=boss:boss:rw-r--r--
> 2012-03-28 00:37:42,387 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> NameSystem.allocateBlock: 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174. 
> blk_2751836614265659170_184668759
> 2012-03-28 00:37:42,696 INFO org.apache.hadoop.hdfs.StateChange: DIR* 
> NameSystem.completeFile: file 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 is closed by 
> DFSClient_attempt_201203271849_0016_r_000174_0
> 2012-03-28 00:37:50,315 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: ugi=boss,boss 
> ip=/10.131.16.34cmd=rename  
> src=/user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 
> dst=/user/boss/pgv/fission/task16/split/  @?
> tor.qzone.qq.com/keypart-00174  perm=boss:boss:rw-r--r--
> after check the code that save FSImage,I found there are a problem that maybe 
> a bug of HDFS Code,I past below:
> -this is the saveFSImage method  in  FSImage.java, I make some 
> mark at the problem code
> /**
>* Save the contents of the FS image to the file.
>*/
>   void saveFSImage(File newFile) throws IOException {
> FSNamesystem fsNamesys = FSNamesystem.getFSNamesystem();
> FSDirectory fsDir = fsNamesys.dir;
> long startTime = FSNamesystem.now();
> //
> // Write out data
> //
> DataOutputStream out = new DataOutputStream(
> new BufferedOutputStream(
>  new 
> FileOutputStream(newFile)));
> try {
>   .
> 
>   // save the rest of the nodes
>   saveImage(strbuf, 0, fsDir.rootDir, out);--problem
>   fsNamesys.saveFilesUnderConstruction(out);--problem  
> detail is below
>   strbuf = null;
> } finally {
>   out.close();
> }
> LOG.info("Image file of size " + newFile.length() + " saved in " 
> + (FSNamesystem.now() - startTime)/10

[jira] [Moved] (HADOOP-9103) when save FSImage ,HDFS( or SecondaryNameNode or FSImage)can't handle some file whose file name has some special messy code(乱码)

2012-11-28 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon moved HDFS-3307 to HADOOP-9103:
---

  Component/s: (was: namenode)
   io
Affects Version/s: (was: 0.20.1)
   0.20.1
  Key: HADOOP-9103  (was: HDFS-3307)
  Project: Hadoop Common  (was: Hadoop HDFS)

> when save FSImage  ,HDFS( or  SecondaryNameNode or FSImage)can't handle some 
> file whose file name has some special messy code(乱码)
> -
>
> Key: HADOOP-9103
> URL: https://issues.apache.org/jira/browse/HADOOP-9103
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: io
>Affects Versions: 0.20.1
> Environment: SUSE LINUX
>Reporter: yixiaohua
>Assignee: Todd Lipcon
> Attachments: FSImage.java, hadoop-9103.txt, ProblemString.txt, 
> TestUTF8AndStringGetBytes.java, TestUTF8AndStringGetBytes.java
>
>   Original Estimate: 12h
>  Remaining Estimate: 12h
>
> this the log information  of the  exception  from the SecondaryNameNode: 
> 2012-03-28 00:48:42,553 ERROR 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: 
> java.io.IOException: Found lease for
>  non-existent file 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/@???
> ??tor.qzone.qq.com/keypart-00174
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFilesUnderConstruction(FSImage.java:1211)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:959)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.doMerge(SecondaryNameNode.java:589)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode$CheckpointStorage.access$000(SecondaryNameNode.java:473)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doMerge(SecondaryNameNode.java:350)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.doCheckpoint(SecondaryNameNode.java:314)
> at 
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.run(SecondaryNameNode.java:225)
> at java.lang.Thread.run(Thread.java:619)
> this is the log information  about the file from namenode:
> 2012-03-28 00:32:26,528 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: ugi=boss,boss 
> ip=/10.131.16.34cmd=create  
> src=/user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 dst=null
> perm=boss:boss:rw-r--r--
> 2012-03-28 00:37:42,387 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
> NameSystem.allocateBlock: 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174. 
> blk_2751836614265659170_184668759
> 2012-03-28 00:37:42,696 INFO org.apache.hadoop.hdfs.StateChange: DIR* 
> NameSystem.completeFile: file 
> /user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 is closed by 
> DFSClient_attempt_201203271849_0016_r_000174_0
> 2012-03-28 00:37:50,315 INFO 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: ugi=boss,boss 
> ip=/10.131.16.34cmd=rename  
> src=/user/boss/pgv/fission/task16/split/_temporary/_attempt_201203271849_0016_r_000174_0/
>   @?tor.qzone.qq.com/keypart-00174 
> dst=/user/boss/pgv/fission/task16/split/  @?
> tor.qzone.qq.com/keypart-00174  perm=boss:boss:rw-r--r--
> after check the code that save FSImage,I found there are a problem that maybe 
> a bug of HDFS Code,I past below:
> -this is the saveFSImage method  in  FSImage.java, I make some 
> mark at the problem code
> /**
>* Save the contents of the FS image to the file.
>*/
>   void saveFSImage(File newFile) throws IOException {
> FSNamesystem fsNamesys = FSNamesystem.getFSNamesystem();
> FSDirectory fsDir = fsNamesys.dir;
> long startTime = FSNamesystem.now();
> //
> // Write out data
> //
> DataOutputStream out = new DataOutputStream(
> new BufferedOutputStream(
>  new 
> FileOutputStream(newFile)));
> try {
>   .
> 
>   // save the rest of the nodes
>   saveImage(strbuf, 0, fsDir.rootDir, out);--problem
>   fsNamesys.saveFilesUnderConstruction(out);--problem  
> detail is below
>   strbuf = null;
> } finally {
>   out.close();
> }
> LOG.info("I

[jira] [Updated] (HADOOP-8981) TestMetricsSystemImpl fails on Windows

2012-11-28 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated HADOOP-8981:
--

Attachment: HADOOP-8981-branch-trunk-win.2.patch

Format the code

> TestMetricsSystemImpl fails on Windows
> --
>
> Key: HADOOP-8981
> URL: https://issues.apache.org/jira/browse/HADOOP-8981
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: metrics
>Affects Versions: trunk-win
>Reporter: Chris Nauroth
>Assignee: Arpit Agarwal
> Attachments: HADOOP-8981-branch-trunk-win.1.patch, 
> HADOOP-8981-branch-trunk-win.2.patch
>
>
> The test is failing on an expected mock interaction.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9090) Refactor MetricsSystemImpl to allow for an on-demand publish system

2012-11-28 Thread Mostafa Elhemali (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Elhemali updated HADOOP-9090:
-

Attachment: HADOOP-9090.justEnhanceDefaultImpl.4.patch

(I highly appreciate the responsive feedback I'm getting from you Luke.)

Thanks - you raise good points. My fear was that do we want the immediate sinks 
to be queued up or not, and do we want to leave the queue thread hanging if the 
sink times out or not. I think though I was swayed to your way of thinking by 
two things: a) the flush() problem you point out and b) that my previous patch 
would've left sinks possible called by multiple threads instead of just the 
queue thread which can be bad.

I've implemented the spirit of what you said. I opted though to have the 
putMetricsImmediate() wait for its specific MetricsBuffer since I figured the 
way you had it may end up in race conditions if multiple threads are calling 
publishMetricsNow() at the same time. Let me know what you think - thanks!

> Refactor MetricsSystemImpl to allow for an on-demand publish system
> ---
>
> Key: HADOOP-9090
> URL: https://issues.apache.org/jira/browse/HADOOP-9090
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: metrics
>Reporter: Mostafa Elhemali
>Priority: Minor
> Attachments: HADOOP-9090.2.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.2.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.3.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.4.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.patch, HADOOP-9090.patch
>
>
> We have a need to publish metrics out of some short-living processes, which 
> is not really well-suited to the current metrics system implementation which 
> periodically publishes metrics asynchronously (a behavior that works great 
> for long-living processes). Of course I could write my own metrics system, 
> but it seems like such a waste to rewrite all the awesome code currently in 
> the MetricsSystemImpl and supporting classes.
> The way I'm proposing to solve this is to:
> 1. Refactor the MetricsSystemImpl class into an abstract base 
> MetricsSystemImpl class (common configuration and other code) and a concrete 
> PeriodicPublishMetricsSystemImpl class (timer thread).
> 2. Refactor the MetricsSinkAdapter class into an abstract base 
> MetricsSinkAdapter class (common configuration and other code) and a concrete 
> AsyncMetricsSinkAdapter class (asynchronous publishing using the SinkQueue).
> 3. Derive a new simple class OnDemandPublishMetricsSystemImpl from 
> MetricsSystemImpl, that just exposes a synchronous publish() method to do all 
> the work.
> 4. Derive a SyncMetricsSinkAdapter class from MetricsSinkAdapter to just 
> synchronously push metrics to the underlying sink.
> Does that sound reasonable? I'll attach the patch with all this coded up and 
> simple tests (could use some polish I guess, but wanted to get everyone's 
> opinion first). Notice that this is somewhat of a breaking change since 
> MetricsSystemImpl is public (although it's marked with 
> InterfaceAudience.Private); if the breaking change is a problem I could just 
> rename the refactored classes so that PeriodicPublishMetricsSystemImpl is 
> still called MetricsSystemImpl (and MetricsSystemImpl -> 
> BaseMetricsSystemImpl).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9057) TestMetricsSystemImpl.testInitFirst fails intermittently

2012-11-28 Thread Luke Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9057?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13506103#comment-13506103
 ] 

Luke Lu commented on HADOOP-9057:
-

Mostafa: the timeout in mockito verify is a good idea. If you can provide trunk 
patch and branch-1 patches, I'll commit them.

> TestMetricsSystemImpl.testInitFirst fails intermittently
> 
>
> Key: HADOOP-9057
> URL: https://issues.apache.org/jira/browse/HADOOP-9057
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 1-win
>Reporter: Ivan Mitic
>Assignee: Ivan Mitic
> Attachments: HADOOP-9057.branch-1-win.metricsrace.patch
>
>
> Error Message
>  Wanted but not invoked: metricsSink.putMetrics(); -> at 
> org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl.testInitFirst(TestMetricsSystemImpl.java:80)
>  Actually, there were zero interactions with this mock. 
> Stacktrace
> Wanted but not invoked:
> metricsSink.putMetrics();
> -> at 
> org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl.testInitFirst(TestMetricsSystemImpl.java:80)
> Actually, there were zero interactions with this mock.
>   at 
> org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl.testInitFirst(TestMetricsSystemImpl.java:80)
>   at 
> org.mockito.internal.runners.JUnit45AndHigherRunnerImpl.run(JUnit45AndHigherRunnerImpl.java:37)
>   at 
> org.mockito.runners.MockitoJUnitRunner.run(MockitoJUnitRunner.java:62)
> Standard Output
> 2012-10-04 11:43:55,641 INFO  impl.MetricsConfig 
> (MetricsConfig.java:loadFirst(99)) - loaded properties from 
> hadoop-metrics2-test.properties

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HADOOP-9102) winutils task isAlive does not return a non-zero exit code if the requested task is not alive

2012-11-28 Thread Chris Nauroth (JIRA)
Chris Nauroth created HADOOP-9102:
-

 Summary: winutils task isAlive does not return a non-zero exit 
code if the requested task is not alive
 Key: HADOOP-9102
 URL: https://issues.apache.org/jira/browse/HADOOP-9102
 Project: Hadoop Common
  Issue Type: Bug
  Components: util
Affects Versions: 1-win
Reporter: Chris Nauroth
Assignee: Chris Nauroth


Work on YARN-233 noted that winutils task isAlive does not return a non-zero 
exit code if the requested task is not alive.  This is inconsistent with the 
equivalent check on Unix, kill -0, which uses a non-zero exit code to 
communicate status.  By changing winutils task isAlive to return a non-zero 
exit code, we can make the error handling code on the Java side consistent, 
avoiding the need for additional if (WINDOWS) checks.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8981) TestMetricsSystemImpl fails on Windows

2012-11-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8981?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13506049#comment-13506049
 ] 

Arpit Agarwal commented on HADOOP-8981:
---

Yes, it is the same. Your fix is very similar to what Chuan has submitted i.e. 
adding a short sleep before the verification.

> TestMetricsSystemImpl fails on Windows
> --
>
> Key: HADOOP-8981
> URL: https://issues.apache.org/jira/browse/HADOOP-8981
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: metrics
>Affects Versions: trunk-win
>Reporter: Chris Nauroth
>Assignee: Arpit Agarwal
> Attachments: HADOOP-8981-branch-trunk-win.1.patch
>
>
> The test is failing on an expected mock interaction.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9090) Refactor MetricsSystemImpl to allow for an on-demand publish system

2012-11-28 Thread Luke Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13506045#comment-13506045
 ] 

Luke Lu commented on HADOOP-9090:
-

I don't think you can control what happens to external systems, which _should_ 
handle arbitrary connection errors etc by unexpected client (even 
router/firewall) shutdown. Another problem of the new patch is that you're 
creating a latch and a semaphore per record (vs per MetricsBuffer) and there 
can be hundreds (up to a few thousands) of records per put. If the sink hangs, 
you'll be recreating new thread/latch/semaphore per record and the user 
perceived timeout would be configured timeout * number of records. Another 
issue is that hanging can happen in sink.flush as well.

Why not do the simple notification in the existing code like the following 
(untested sketch)?:
{code}
boolean oobPut;

// illustration only, should be in the ctor after retry* variables are defined
final long OOB_PUT_TIMEOUT = retryDelay * Math.pow(retryBackoff, retryCount) * 
1000;

synchronized void putMetricsImmediate(MetricsBuffer mb) {
  if (!oobPut) {
oobPut = true;
if (queue.enqueue(buffer)) {
  wait(OOB_PUT_TIMEOUT);
  oobPut = false;
} // otherwise queue is full due to sink issues anyway.
  } else { // another oobPut in progress
wait(OOB_PUT_TIMEOUT);
oobPut = false; // just in case
  }
}

// after queue.consumeAll(this); in publishMetricsFromQueue (needs to be 
synchronized now)
if (oobPut) {
  notifyAll();
}
{code}

Now you get all the retry/timeout logic for free :)


> Refactor MetricsSystemImpl to allow for an on-demand publish system
> ---
>
> Key: HADOOP-9090
> URL: https://issues.apache.org/jira/browse/HADOOP-9090
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: metrics
>Reporter: Mostafa Elhemali
>Priority: Minor
> Attachments: HADOOP-9090.2.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.2.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.3.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.patch, HADOOP-9090.patch
>
>
> We have a need to publish metrics out of some short-living processes, which 
> is not really well-suited to the current metrics system implementation which 
> periodically publishes metrics asynchronously (a behavior that works great 
> for long-living processes). Of course I could write my own metrics system, 
> but it seems like such a waste to rewrite all the awesome code currently in 
> the MetricsSystemImpl and supporting classes.
> The way I'm proposing to solve this is to:
> 1. Refactor the MetricsSystemImpl class into an abstract base 
> MetricsSystemImpl class (common configuration and other code) and a concrete 
> PeriodicPublishMetricsSystemImpl class (timer thread).
> 2. Refactor the MetricsSinkAdapter class into an abstract base 
> MetricsSinkAdapter class (common configuration and other code) and a concrete 
> AsyncMetricsSinkAdapter class (asynchronous publishing using the SinkQueue).
> 3. Derive a new simple class OnDemandPublishMetricsSystemImpl from 
> MetricsSystemImpl, that just exposes a synchronous publish() method to do all 
> the work.
> 4. Derive a SyncMetricsSinkAdapter class from MetricsSinkAdapter to just 
> synchronously push metrics to the underlying sink.
> Does that sound reasonable? I'll attach the patch with all this coded up and 
> simple tests (could use some polish I guess, but wanted to get everyone's 
> opinion first). Notice that this is somewhat of a breaking change since 
> MetricsSystemImpl is public (although it's marked with 
> InterfaceAudience.Private); if the breaking change is a problem I could just 
> rename the refactored classes so that PeriodicPublishMetricsSystemImpl is 
> still called MetricsSystemImpl (and MetricsSystemImpl -> 
> BaseMetricsSystemImpl).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8981) TestMetricsSystemImpl fails on Windows

2012-11-28 Thread Ivan Mitic (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8981?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13506040#comment-13506040
 ] 

Ivan Mitic commented on HADOOP-8981:


Is this the same issue as reported by HADOOP-9057?

> TestMetricsSystemImpl fails on Windows
> --
>
> Key: HADOOP-8981
> URL: https://issues.apache.org/jira/browse/HADOOP-8981
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: metrics
>Affects Versions: trunk-win
>Reporter: Chris Nauroth
>Assignee: Arpit Agarwal
> Attachments: HADOOP-8981-branch-trunk-win.1.patch
>
>
> The test is failing on an expected mock interaction.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8981) TestMetricsSystemImpl fails on Windows

2012-11-28 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8981?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13506031#comment-13506031
 ] 

Arpit Agarwal commented on HADOOP-8981:
---

Tabs should be replaced with two spaces. Looks fine otherwise.

> TestMetricsSystemImpl fails on Windows
> --
>
> Key: HADOOP-8981
> URL: https://issues.apache.org/jira/browse/HADOOP-8981
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: metrics
>Affects Versions: trunk-win
>Reporter: Chris Nauroth
>Assignee: Arpit Agarwal
> Attachments: HADOOP-8981-branch-trunk-win.1.patch
>
>
> The test is failing on an expected mock interaction.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8989) hadoop dfs -find feature

2012-11-28 Thread Jonathan Allen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8989?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13506017#comment-13506017
 ] 

Jonathan Allen commented on HADOOP-8989:


Allen - I haven't done any specific studies of the NN impact but I would expect 
it to be the same as any of the existing recursive command (e.g. ls -R).  The 
code uses the existing recursion code and applies the expressions to the 
returned FileStatus on the client side.  The -exec option obviously impacts the 
NN but no more than if you ran the command directly (though you don't have the 
JVM start-up overhead so the rate of NN messages will be higher).

Wolfgang - I'll look at adding those option expressions in (I got to the point 
where it looked like I'd be adding expressions forever so I decided to stop and 
see what people asked for).

Daryn - if you do get a chance to review this then the main code is pretty much 
finished (though I need to sort out the formatting before submitting properly) 
but I've reworked the unit tests since uploading the last patch (I'll upload a 
new at the weekend).  I still need to update the documentation web page and add 
some stuff in the CLI test suites.

> hadoop dfs -find feature
> 
>
> Key: HADOOP-8989
> URL: https://issues.apache.org/jira/browse/HADOOP-8989
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Marco Nicosia
>Assignee: Jonathan Allen
> Attachments: HADOOP-8989.patch, HADOOP-8989.patch, HADOOP-8989.patch
>
>
> Both sysadmins and users make frequent use of the unix 'find' command, but 
> Hadoop has no correlate. Without this, users are writing scripts which make 
> heavy use of hadoop dfs -lsr, and implementing find one-offs. I think hdfs 
> -lsr is somewhat taxing on the NameNode, and a really slow experience on the 
> client side. Possibly an in-NameNode find operation would be only a bit more 
> taxing on the NameNode, but significantly faster from the client's point of 
> view?
> The minimum set of options I can think of which would make a Hadoop find 
> command generally useful is (in priority order):
> * -type (file or directory, for now)
> * -atime/-ctime-mtime (... and -creationtime?) (both + and - arguments)
> * -print0 (for piping to xargs -0)
> * -depth
> * -owner/-group (and -nouser/-nogroup)
> * -name (allowing for shell pattern, or even regex?)
> * -perm
> * -size
> One possible special case, but could possibly be really cool if it ran from 
> within the NameNode:
> * -delete
> The "hadoop dfs -lsr | hadoop dfs -rm" cycle is really, really slow.
> Lower priority, some people do use operators, mostly to execute -or searches 
> such as:
> * find / \(-nouser -or -nogroup\)
> Finally, I thought I'd include a link to the [Posix spec for 
> find|http://www.opengroup.org/onlinepubs/009695399/utilities/find.html]

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9090) Refactor MetricsSystemImpl to allow for an on-demand publish system

2012-11-28 Thread Mostafa Elhemali (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9090?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mostafa Elhemali updated HADOOP-9090:
-

Attachment: HADOOP-9090.justEnhanceDefaultImpl.3.patch

Thanks Luke. It's a fundamentally hard problem: if we assume that putMetrics() 
from arbitrary sinks can hang, then we'll have to interrupt the thread that 
doing the work and hope that unsticks it (that's how thread pools cancel work 
for example). Having said that, I got uneasy about using stop(), and I agree 
with your comment that I was creating too many threads willy-nilly. The new 
patch thus does two things:
1. Creates a single thread (if needed) for on-demand publishing per sink, and 
reuses that for subsequent publishing. This should cut down considerably on how 
many threads are created.
2. If it sees a sink timeout, it interrupts the worker thread but doesn't 
stop() it (joins it and hangs with it if needed).

I personally think that's the most reliable thing we can do. One alternative as 
you say is to mark the worker thread as a daemon, so we don't have to join it 
and hang with it if it hangs, but just abandon it. The problem there is that 
sinks may be doing things to external systems that will leave them in 
inconsistent states if the process just shuts down while they're still working 
with no warning. Let me know what you think.

> Refactor MetricsSystemImpl to allow for an on-demand publish system
> ---
>
> Key: HADOOP-9090
> URL: https://issues.apache.org/jira/browse/HADOOP-9090
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: metrics
>Reporter: Mostafa Elhemali
>Priority: Minor
> Attachments: HADOOP-9090.2.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.2.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.3.patch, 
> HADOOP-9090.justEnhanceDefaultImpl.patch, HADOOP-9090.patch
>
>
> We have a need to publish metrics out of some short-living processes, which 
> is not really well-suited to the current metrics system implementation which 
> periodically publishes metrics asynchronously (a behavior that works great 
> for long-living processes). Of course I could write my own metrics system, 
> but it seems like such a waste to rewrite all the awesome code currently in 
> the MetricsSystemImpl and supporting classes.
> The way I'm proposing to solve this is to:
> 1. Refactor the MetricsSystemImpl class into an abstract base 
> MetricsSystemImpl class (common configuration and other code) and a concrete 
> PeriodicPublishMetricsSystemImpl class (timer thread).
> 2. Refactor the MetricsSinkAdapter class into an abstract base 
> MetricsSinkAdapter class (common configuration and other code) and a concrete 
> AsyncMetricsSinkAdapter class (asynchronous publishing using the SinkQueue).
> 3. Derive a new simple class OnDemandPublishMetricsSystemImpl from 
> MetricsSystemImpl, that just exposes a synchronous publish() method to do all 
> the work.
> 4. Derive a SyncMetricsSinkAdapter class from MetricsSinkAdapter to just 
> synchronously push metrics to the underlying sink.
> Does that sound reasonable? I'll attach the patch with all this coded up and 
> simple tests (could use some polish I guess, but wanted to get everyone's 
> opinion first). Notice that this is somewhat of a breaking change since 
> MetricsSystemImpl is public (although it's marked with 
> InterfaceAudience.Private); if the breaking change is a problem I could just 
> rename the refactored classes so that PeriodicPublishMetricsSystemImpl is 
> still called MetricsSystemImpl (and MetricsSystemImpl -> 
> BaseMetricsSystemImpl).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9100) Fix TODO items in FsShell

2012-11-28 Thread Eli Collins (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9100?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505846#comment-13505846
 ] 

Eli Collins commented on HADOOP-9100:
-

See HADOOP-8690 for one of them. 

> Fix TODO items in FsShell
> -
>
> Key: HADOOP-9100
> URL: https://issues.apache.org/jira/browse/HADOOP-9100
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Suresh Srinivas
>Assignee: Suresh Srinivas
>Priority: Minor
>
> Bunch of TODO items in FsShell needs to be fixed.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8981) TestMetricsSystemImpl fails on Windows

2012-11-28 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated HADOOP-8981:
--

Attachment: HADOOP-8981-branch-trunk-win.1.patch

Attached the patch that addresses the test fails on windows. Add a test case to 
verify the callback correctly, abd test case to verify the stop call immediately

> TestMetricsSystemImpl fails on Windows
> --
>
> Key: HADOOP-8981
> URL: https://issues.apache.org/jira/browse/HADOOP-8981
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: metrics
>Affects Versions: trunk-win
>Reporter: Chris Nauroth
>Assignee: Arpit Agarwal
> Attachments: HADOOP-8981-branch-trunk-win.1.patch
>
>
> The test is failing on an expected mock interaction.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9078) enhance unit-test coverage of class org.apache.hadoop.fs.FileContext

2012-11-28 Thread Ivan A. Veselovsky (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan A. Veselovsky updated HADOOP-9078:
---

Affects Version/s: 0.23.6
   2.0.3-alpha
   3.0.0
   Status: Patch Available  (was: Open)

> enhance unit-test coverage of class org.apache.hadoop.fs.FileContext
> 
>
> Key: HADOOP-9078
> URL: https://issues.apache.org/jira/browse/HADOOP-9078
> Project: Hadoop Common
>  Issue Type: Test
>Affects Versions: 3.0.0, 2.0.3-alpha, 0.23.6
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
> Attachments: HADOOP-9078-branch-0.23.patch, 
> HADOOP-9078-branch-2.patch, HADOOP-9078.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9078) enhance unit-test coverage of class org.apache.hadoop.fs.FileContext

2012-11-28 Thread Ivan A. Veselovsky (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan A. Veselovsky updated HADOOP-9078:
---

Attachment: HADOOP-9078.patch

"HADOOP-9078" is for trunk, other patches are for the branches respectively.

> enhance unit-test coverage of class org.apache.hadoop.fs.FileContext
> 
>
> Key: HADOOP-9078
> URL: https://issues.apache.org/jira/browse/HADOOP-9078
> Project: Hadoop Common
>  Issue Type: Test
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
> Attachments: HADOOP-9078-branch-0.23.patch, 
> HADOOP-9078-branch-2.patch, HADOOP-9078.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9078) enhance unit-test coverage of class org.apache.hadoop.fs.FileContext

2012-11-28 Thread Ivan A. Veselovsky (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9078?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan A. Veselovsky updated HADOOP-9078:
---

Attachment: HADOOP-9078-branch-2.patch
HADOOP-9078-branch-0.23.patch

> enhance unit-test coverage of class org.apache.hadoop.fs.FileContext
> 
>
> Key: HADOOP-9078
> URL: https://issues.apache.org/jira/browse/HADOOP-9078
> Project: Hadoop Common
>  Issue Type: Test
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
> Attachments: HADOOP-9078-branch-0.23.patch, 
> HADOOP-9078-branch-2.patch, HADOOP-9078.patch
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HADOOP-9101) make s3n NativeFileSystemStore interface public instead of package-private

2012-11-28 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-9101:
--

 Summary: make s3n NativeFileSystemStore interface public instead 
of package-private
 Key: HADOOP-9101
 URL: https://issues.apache.org/jira/browse/HADOOP-9101
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs/s3
Affects Versions: 3.0.0
Reporter: Steve Loughran
Priority: Trivial


It would be easier to implement new blockstore filesystems if the the 
{{NativeFileSystemStore} and dependent classes in the 
{{org.apache.hadoop.fs.s3native}} package were public -currently you need to 
put them into the s3 directory.

They could be made public with the appropriate scope attribute. Internal?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9046) provide unit-test coverage of class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction

2012-11-28 Thread Robert Joseph Evans (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505661#comment-13505661
 ] 

Robert Joseph Evans commented on HADOOP-9046:
-

It is looking good, and I like how much faster the tests run.

I have a few more comments.

# {code}TODO: should we reject the addition if an action for this 'fs' is 
already present in the queue?{code} If this is an issue then we should file a 
JIRA to fix it, otherwise please remove the TODO. I am leaning more towards 
filing a separate JIRA.
# I don't really like the name lock0 and available0.  I would rather have them 
be named something more descriptive.  This is very minor
# in removeRenewAction we are no longer canceling the token when it is removed. 
I am not sure if that is intentional or not, but it definitely changes the 
functionality of the method. 
# I am not sure that renewerThreadStarted.compairAndSet is that much better 
then isAlive().  isAlive has the problem that if the thread stopped, calling 
add again would result in an IllegalThreadStateException, but compairAndSet has 
the problem that if a new thread cannot be created (Which I have seen happen) 
then only the first call to add will get an exception all other calls to add 
will look like they succeeded but no renewal will ever happen. I would rather 
see errors happen on other calls to add then to have them silently fail.


> provide unit-test coverage of class 
> org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction
> --
>
> Key: HADOOP-9046
> URL: https://issues.apache.org/jira/browse/HADOOP-9046
> Project: Hadoop Common
>  Issue Type: Test
>Affects Versions: 3.0.0, 2.0.3-alpha, 0.23.6
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
>Priority: Minor
> Attachments: HADOOP-9046-branch-0.23--c.patch, 
> HADOOP-9046-branch-0.23-over-9049.patch, HADOOP-9046-branch-0.23.patch, 
> HADOOP-9046--c.patch, HADOOP-9046-over-9049.patch, HADOOP-9046.patch
>
>
> The class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction has zero 
> coverage in entire cumulative test run. Provide test(s) to cover this class.
> Note: the request submitted to HDFS project because the class likely to be 
> tested by tests in that project.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9046) provide unit-test coverage of class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction

2012-11-28 Thread Ivan A. Veselovsky (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9046?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan A. Veselovsky updated HADOOP-9046:
---

Affects Version/s: 0.23.6
   2.0.3-alpha
   3.0.0

> provide unit-test coverage of class 
> org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction
> --
>
> Key: HADOOP-9046
> URL: https://issues.apache.org/jira/browse/HADOOP-9046
> Project: Hadoop Common
>  Issue Type: Test
>Affects Versions: 3.0.0, 2.0.3-alpha, 0.23.6
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
>Priority: Minor
> Attachments: HADOOP-9046-branch-0.23--c.patch, 
> HADOOP-9046-branch-0.23-over-9049.patch, HADOOP-9046-branch-0.23.patch, 
> HADOOP-9046--c.patch, HADOOP-9046-over-9049.patch, HADOOP-9046.patch
>
>
> The class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction has zero 
> coverage in entire cumulative test run. Provide test(s) to cover this class.
> Note: the request submitted to HDFS project because the class likely to be 
> tested by tests in that project.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9046) provide unit-test coverage of class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction

2012-11-28 Thread Ivan A. Veselovsky (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9046?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan A. Veselovsky updated HADOOP-9046:
---

Attachment: HADOOP-9046-branch-0.23--c.patch

Patch "HADOOP-9046-branch-0.23--c.patch" ports the changes to "branch-0.23".

The trunk patch "HADOOP-9046--c.patch" is applicable to "branch-2" as well.

> provide unit-test coverage of class 
> org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction
> --
>
> Key: HADOOP-9046
> URL: https://issues.apache.org/jira/browse/HADOOP-9046
> Project: Hadoop Common
>  Issue Type: Test
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
>Priority: Minor
> Attachments: HADOOP-9046-branch-0.23--c.patch, 
> HADOOP-9046-branch-0.23-over-9049.patch, HADOOP-9046-branch-0.23.patch, 
> HADOOP-9046--c.patch, HADOOP-9046-over-9049.patch, HADOOP-9046.patch
>
>
> The class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction has zero 
> coverage in entire cumulative test run. Provide test(s) to cover this class.
> Note: the request submitted to HDFS project because the class likely to be 
> tested by tests in that project.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9044) add FindClass main class to provide classpath checking of installations

2012-11-28 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-9044:
---

Affects Version/s: 2.0.3-alpha
   Status: Patch Available  (was: In Progress)

> add FindClass main class to provide classpath checking of installations
> ---
>
> Key: HADOOP-9044
> URL: https://issues.apache.org/jira/browse/HADOOP-9044
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 1.1.0, 2.0.3-alpha
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
> Attachments: HADOOP-9044.patch, HADOOP-9044.patch
>
>
> It's useful in postflight checking of a hadoop installation to verify that 
> classes load, especially codes with external JARs and native codecs. 
> An entry point designed to load a named class and create an instance of that 
> class can do this -and be invoked from any shell script or tool that does the 
> installation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9044) add FindClass main class to provide classpath checking of installations

2012-11-28 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-9044:
---

Attachment: HADOOP-9044.patch

patch against trunk

> add FindClass main class to provide classpath checking of installations
> ---
>
> Key: HADOOP-9044
> URL: https://issues.apache.org/jira/browse/HADOOP-9044
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 1.1.0, 2.0.3-alpha
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
> Attachments: HADOOP-9044.patch, HADOOP-9044.patch
>
>
> It's useful in postflight checking of a hadoop installation to verify that 
> classes load, especially codes with external JARs and native codecs. 
> An entry point designed to load a named class and create an instance of that 
> class can do this -and be invoked from any shell script or tool that does the 
> installation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9046) provide unit-test coverage of class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction

2012-11-28 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9046?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505561#comment-13505561
 ] 

Hadoop QA commented on HADOOP-9046:
---

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12555168/HADOOP-9046--c.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-common-project/hadoop-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/1834//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/1834//console

This message is automatically generated.

> provide unit-test coverage of class 
> org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction
> --
>
> Key: HADOOP-9046
> URL: https://issues.apache.org/jira/browse/HADOOP-9046
> Project: Hadoop Common
>  Issue Type: Test
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
>Priority: Minor
> Attachments: HADOOP-9046-branch-0.23-over-9049.patch, 
> HADOOP-9046-branch-0.23.patch, HADOOP-9046--c.patch, 
> HADOOP-9046-over-9049.patch, HADOOP-9046.patch
>
>
> The class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction has zero 
> coverage in entire cumulative test run. Provide test(s) to cover this class.
> Note: the request submitted to HDFS project because the class likely to be 
> tested by tests in that project.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9046) provide unit-test coverage of class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction

2012-11-28 Thread Ivan A. Veselovsky (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9046?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ivan A. Veselovsky updated HADOOP-9046:
---

Attachment: HADOOP-9046--c.patch

In the new version of patch "HADOOP-9046--c.patch" the following is done:
1) fixed the sync problem found by Robert: now the method 
org.apache.hadoop.fs.DelegationTokenRenewer#removeRenewAction(T) will not block 
for long time.
2) in the test all the time periods made scalable relative one constant.
3) the test made fater: now it takes ~7 seconds.

> provide unit-test coverage of class 
> org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction
> --
>
> Key: HADOOP-9046
> URL: https://issues.apache.org/jira/browse/HADOOP-9046
> Project: Hadoop Common
>  Issue Type: Test
>Reporter: Ivan A. Veselovsky
>Assignee: Ivan A. Veselovsky
>Priority: Minor
> Attachments: HADOOP-9046-branch-0.23-over-9049.patch, 
> HADOOP-9046-branch-0.23.patch, HADOOP-9046--c.patch, 
> HADOOP-9046-over-9049.patch, HADOOP-9046.patch
>
>
> The class org.apache.hadoop.fs.DelegationTokenRenewer.RenewAction has zero 
> coverage in entire cumulative test run. Provide test(s) to cover this class.
> Note: the request submitted to HDFS project because the class likely to be 
> tested by tests in that project.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8989) hadoop dfs -find feature

2012-11-28 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8989?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505509#comment-13505509
 ] 

Daryn Sharp commented on HADOOP-8989:
-

Sorry, this fell off my radar.  I've been swamped.

bq. Has anyone studied the impact on the NN yet?
W/o having reviewed the patch, I initially don't think this could be any worse 
than ls -R.

> hadoop dfs -find feature
> 
>
> Key: HADOOP-8989
> URL: https://issues.apache.org/jira/browse/HADOOP-8989
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Marco Nicosia
>Assignee: Jonathan Allen
> Attachments: HADOOP-8989.patch, HADOOP-8989.patch, HADOOP-8989.patch
>
>
> Both sysadmins and users make frequent use of the unix 'find' command, but 
> Hadoop has no correlate. Without this, users are writing scripts which make 
> heavy use of hadoop dfs -lsr, and implementing find one-offs. I think hdfs 
> -lsr is somewhat taxing on the NameNode, and a really slow experience on the 
> client side. Possibly an in-NameNode find operation would be only a bit more 
> taxing on the NameNode, but significantly faster from the client's point of 
> view?
> The minimum set of options I can think of which would make a Hadoop find 
> command generally useful is (in priority order):
> * -type (file or directory, for now)
> * -atime/-ctime-mtime (... and -creationtime?) (both + and - arguments)
> * -print0 (for piping to xargs -0)
> * -depth
> * -owner/-group (and -nouser/-nogroup)
> * -name (allowing for shell pattern, or even regex?)
> * -perm
> * -size
> One possible special case, but could possibly be really cool if it ran from 
> within the NameNode:
> * -delete
> The "hadoop dfs -lsr | hadoop dfs -rm" cycle is really, really slow.
> Lower priority, some people do use operators, mostly to execute -or searches 
> such as:
> * find / \(-nouser -or -nogroup\)
> Finally, I thought I'd include a link to the [Posix spec for 
> find|http://www.opengroup.org/onlinepubs/009695399/utilities/find.html]

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-7868) Hadoop native fails to compile when default linker option is -Wl,--as-needed

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7868?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505422#comment-13505422
 ] 

Hudson commented on HADOOP-7868:


Integrated in Hadoop-Hdfs-0.23-Build #449 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/449/])
HADOOP-7868. Hadoop native fails to compile when default linker option is 
-Wl,--as-needed. (Trevor Robinson via tgraves) (Revision 1414370)

 Result = SUCCESS
tgraves : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1414370
Files : 
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/main/native/acinclude.m4
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/src/main/native/configure.ac


> Hadoop native fails to compile when default linker option is -Wl,--as-needed
> 
>
> Key: HADOOP-7868
> URL: https://issues.apache.org/jira/browse/HADOOP-7868
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: native
>Affects Versions: 1.0.0, 2.0.0-alpha
> Environment: Ubuntu Precise, Ubuntu Oneiric, Debian Unstable
>Reporter: James Page
>Assignee: Trevor Robinson
> Fix For: 1.2.0, 2.0.2-alpha, 0.23.6
>
> Attachments: hadoop-7868-b1.txt, HADOOP-7868.patch, 
> HADOOP-7868-portable.patch
>
>
> Recent releases of Ubuntu and Debian have switched to using --as-needed as 
> default when linking binaries.
> As a result the AC_COMPUTE_NEEDED_DSO fails to find the required DSO names 
> during execution of configure resulting in a build failure.
> Explicitly using "-Wl,--no-as-needed" in this macro when required resolves 
> this issue.
> See http://wiki.debian.org/ToolChain/DSOLinking for a few more details

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9084) TotalOrderPartitioner fails on hadoop running on top of gpfs (or any parallel or distributed filesystem)

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9084?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505360#comment-13505360
 ] 

Hudson commented on HADOOP-9084:


Integrated in Hadoop-Yarn-trunk #50 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/50/])
HADOOP-9064. Ameding CHANGES.txt with correct JIRA ID, wrongfully used 
HADOOP-9084 before (tucu) (Revision 1414347)
HADOOP-9084. Augment DelegationTokenRenewer API to cancel the tokens on calls 
to removeRenewAction. (kkambatl via tucu) (Revision 1414326)

 Result = SUCCESS
tucu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1414347
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt

tucu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1414326
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/dev-support/findbugsExcludeFile.xml
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/DelegationTokenRenewer.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestDelegationTokenRenewer.java


> TotalOrderPartitioner fails on hadoop running on top of gpfs (or any parallel 
> or distributed filesystem)
> 
>
> Key: HADOOP-9084
> URL: https://issues.apache.org/jira/browse/HADOOP-9084
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: filecache, fs, native
>Affects Versions: 1.0.3, 1.0.4
>Reporter: giovanni delussu
>Assignee: giovanni delussu
>Priority: Critical
> Fix For: 1.0.3, 1.0.4
>
> Attachments: PATCH-HADOOP-9084-origin-branch-1.0.patch
>
>
> When running a job who uses TotalOrderPartitioner (like TeraSort or 
> BulkImport of HBase) on hadoop running on top of gpfs (instead of hdfs) the 
> program fails to find the file _partition.lst because is looking for it in 
> the wrong directory. The confusion is between local fs meaning not hdfs and 
> local meaning distributed fs. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9064) Augment DelegationTokenRenewer API to cancel the tokens on calls to removeRenewAction

2012-11-28 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13505361#comment-13505361
 ] 

Hudson commented on HADOOP-9064:


Integrated in Hadoop-Yarn-trunk #50 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/50/])
HADOOP-9064. Ameding CHANGES.txt with correct JIRA ID, wrongfully used 
HADOOP-9084 before (tucu) (Revision 1414347)

 Result = SUCCESS
tucu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1414347
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Augment DelegationTokenRenewer API to cancel the tokens on calls to 
> removeRenewAction
> -
>
> Key: HADOOP-9064
> URL: https://issues.apache.org/jira/browse/HADOOP-9064
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.0.2-alpha
>Reporter: Karthik Kambatla
>Assignee: Karthik Kambatla
> Fix For: 2.0.3-alpha
>
> Attachments: hadoop-9064.patch, hadoop-9064.patch, hadoop-9064.patch, 
> hadoop-9064.patch
>
>
> Post HADOOP-9049, FileSystems register with DelegationTokenRenewer (a 
> singleton), to renew tokens. 
> To avoid a bunch of defunct tokens clog the NN, we should augment the API to 
> {{#removeRenewAction(boolean cancel)}} and cancel the token appropriately.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira