[jira] [Created] (HADOOP-9620) Stabilize branch-2 on Windows

2013-06-04 Thread Chuan Liu (JIRA)
Chuan Liu created HADOOP-9620:
-

 Summary: Stabilize branch-2 on Windows
 Key: HADOOP-9620
 URL: https://issues.apache.org/jira/browse/HADOOP-9620
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.1.0-beta
Reporter: Chuan Liu


After merging trunk change into branch-2, we are seeing a new set of unit test 
failures in branch-2 on Windows. This an umbrella jira to track the remaining 
issues in branch-2 on Windows with the goal to stabilize branch-2 on Windows 
and achieve 100% unit test pass rate.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9619) Mark stability of .proto files

2013-06-04 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675526#comment-13675526
 ] 

Sanjay Radia commented on HADOOP-9619:
--

Here is the proposed rules that will be put in a file.

When a .proto file is marked as stable it means that changes should be made in 
a compatible fashion. Compatibility can be broken  at a major release, though 
breaking compatibility even at major releases has grave consequences.

The following changes are *never* allowed. 
* Change a field id 
* reuse an old field that was previously deleted.
Field numbers are cheap and changing and reusing is not a good idea.


The following changes *cannot* be made to a *stable* .proto except at a major 
release since they are considered *incompatible*.
* Modify a field type in an *incompatible* way (as defined recursively)
* Add or delete a required field
* Delete an optional field  - Q. I am not sure if all our current optional 
fields have reasonable defaults to allow deletions

The following changes are allowed at any time
* Add an optional field, with the expectation that the code deals with the 
field missing due to communication with an older version of the code.
* Rename a field 
* Rename a .proto file
* Change .proto annotations that effect code generation (e.g. name of java 
package)



> Mark stability of .proto files
> --
>
> Key: HADOOP-9619
> URL: https://issues.apache.org/jira/browse/HADOOP-9619
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9422) HADOOP_HOME should not be required to be set to be able to launch commands using hadoop.util.Shell

2013-06-04 Thread Enis Soztutar (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9422?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675522#comment-13675522
 ] 

Enis Soztutar commented on HADOOP-9422:
---

The requirement as far as I can tell, was that windows support needs to find a 
native library winutils, and it looks under hadoop home. +1 on making it a soft 
dependency on non-windows playforms, and in a further issue change the winutils 
to be under hadoop native lib dir. 

> HADOOP_HOME should not be required to be set to be able to launch commands 
> using hadoop.util.Shell
> --
>
> Key: HADOOP-9422
> URL: https://issues.apache.org/jira/browse/HADOOP-9422
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Hitesh Shah
>
> Not sure why this is an enforced requirement especially in cases where a 
> deployment is done using multiple tar-balls ( one each for 
> common/hdfs/mapreduce/yarn ). 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9619) Mark stability of .proto files

2013-06-04 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675473#comment-13675473
 ] 

Sanjay Radia commented on HADOOP-9619:
--

We simultaneously added our comments.

bq.   I don't think the .proto files themselves should be stable
Todd, stable means evolve in a compatible way - hence you can change things 
like the name of fields but not their ids, add new optional fields etc. Yes we 
can rename a file. Let me document that in a separate file and have the .proto 
files refer to it.

bq. If a downstream project ...
agreed.


> Mark stability of .proto files
> --
>
> Key: HADOOP-9619
> URL: https://issues.apache.org/jira/browse/HADOOP-9619
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675471#comment-13675471
 ] 

Todd Lipcon commented on HADOOP-9618:
-

bq. I kind of wish we could use the JVM's Xloggc:logfile to get this 
information, since theoretically it should be more trustworthy than trying to 
guess. Is that too much hassle to configure by default?

The problem is that the GC logs don't roll, plus it's difficult to correlate 
that into the log4j stream, since the timestamps in the GC logs are different 
format than log4j, etc -- plus they won't rollup through alternate log4j 
appenders to centralized monitoring.

bq. I suppose the thread method detects machine pauses which are not the result 
of GCs, so you could say that it gives more information (although perhaps more 
questionable information).

Yep - I've seen cases where the kernel locks up for multiple seconds due to 
some bug, and that's interesting. Also there's JVM "safepoint pauses" which are 
nasty and aren't in the gc logs unless you use -XX:+PrintSafepointStatistics, 
which is super verbose.

bq. I'm a little gun-shy of the 1 second timeout. It wasn't too long ago that 
the Linux scheduler quantum was 100 milliseconds. So if you had ten threads 
hogging the CPU, you'd already have no time left to run your watchdog thread. I 
think the timeout either needs to be longer, or the thread needs to be a 
high-priority thread, possibly even realtime priority.

If one of your important Hadoop daemons is so overloaded, I think that would be 
interesting as well. This only logs if the 1-second pause takes 3 seconds, so 
things like scheduling jitter won't cause log messages unless the "jitter" is 
multiple seconds long. At that point, I'd want to know about it regardless of 
whether it's GC, a kernel issue, contention for machine resources, swap, etc. 
Do you disagree?

> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9619) Mark stability of .proto files

2013-06-04 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675469#comment-13675469
 ] 

Sanjay Radia commented on HADOOP-9619:
--

Proposing to mark the .proto files in Hadoop common and HDFS as 
*private-stable*. 
This is for both client facing protocols and for internal protocols.
We are promising wire compatibility and hence the label stable.
One we decide to make the rpc and the data transfer protocols public-stable we 
can then mark the client facing .protos as public.

Since there are no annotations for .proto files I am planning to simply add a 
comment plus a short description of what changes can be made (add optional 
fields, ...) 

> Mark stability of .proto files
> --
>
> Key: HADOOP-9619
> URL: https://issues.apache.org/jira/browse/HADOOP-9619
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9619) Mark stability of .proto files

2013-06-04 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675463#comment-13675463
 ] 

Todd Lipcon commented on HADOOP-9619:
-

One thing I think we should clarify: I don't think the .proto files themselves 
should be stable. Rather, the field IDs and types are stable -- ie we should 
not make any changes to the proto files which cause wire incompatibility. But 
we should feel free to rename the .proto files themselves, or change the 
annotations that affect code generation.

If a downstream project wants to use the Hadoop wire protocol, they should be 
making a _copy_ of the proto files for their use case, not assuming that we'll 
never rename a protobuf field (which is wire-compatible).

> Mark stability of .proto files
> --
>
> Key: HADOOP-9619
> URL: https://issues.apache.org/jira/browse/HADOOP-9619
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: documentation
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HADOOP-9619) Mark stability of .proto files

2013-06-04 Thread Sanjay Radia (JIRA)
Sanjay Radia created HADOOP-9619:


 Summary: Mark stability of .proto files
 Key: HADOOP-9619
 URL: https://issues.apache.org/jira/browse/HADOOP-9619
 Project: Hadoop Common
  Issue Type: Sub-task
Reporter: Sanjay Radia
Assignee: Sanjay Radia




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9517) Document Hadoop Compatibility

2013-06-04 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675461#comment-13675461
 ] 

Sanjay Radia commented on HADOOP-9517:
--

bq.  ... Optional fields can be added any time .. Fields can be renamed any 
time ...

This is what stable means. Hence i suggest that we add a comment to the .proto 
files to say that the .protos are private-stable, and we could add the comment 
on kinds of changes allowed. Will file a jira for this. Given that are no 
annotations for .proto, a comment is the best that can be done.

bq. Don't you think the proto files for client interfaces should be public? I 
was chatting with Todd about this, and it seems to us they should.
I would still mark it as private till we make the rpc and data transfer 
protocol itself public (ie the protos being public is useless without the rpc 
proto being public. 
Todd and I occasionally disagree ;-)




> Document Hadoop Compatibility
> -
>
> Key: HADOOP-9517
> URL: https://issues.apache.org/jira/browse/HADOOP-9517
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Reporter: Arun C Murthy
>Assignee: Karthik Kambatla
> Attachments: hadoop-9517.patch, hadoop-9517.patch, hadoop-9517.patch, 
> hadoop-9517.patch, hadoop-9517-proposal-v1.patch, 
> hadoop-9517-proposal-v1.patch, hadoop-9517-v2.patch
>
>
> As we get ready to call hadoop-2 stable we need to better define 'Hadoop 
> Compatibility'.
> http://wiki.apache.org/hadoop/Compatibility is a start, let's document 
> requirements clearly and completely.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675439#comment-13675439
 ] 

Colin Patrick McCabe commented on HADOOP-9618:
--

er, that should read "10 milliseconds" / 100 CPU-bound threads

> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675436#comment-13675436
 ] 

Colin Patrick McCabe commented on HADOOP-9618:
--

I kind of wish we could use the JVM's {{Xloggc:logfile}} to get this 
information, since theoretically it should be more trustworthy than trying to 
guess.  Is that too much hassle to configure by default?

I suppose the thread method detects machine pauses which are *not* the result 
of GCs, so you could say that it gives more information (although perhaps more 
questionable information).

I'm a little gun-shy of the 1 second timeout.  It wasn't too long ago that the 
Linux scheduler quantum was 100 milliseconds.  So if you had ten threads 
hogging the CPU, you'd already have no time left to run your watchdog thread.  
I think the timeout either needs to be longer, or the thread needs to be a 
high-priority thread, possibly even realtime priority.

Have you tried running this with a gnarly MapReduce job going on?

> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9614) smart-test-patch.sh hangs for new version of patch (2.7.1)

2013-06-04 Thread Konstantin Boudnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Konstantin Boudnik updated HADOOP-9614:
---

Fix Version/s: (was: 2.0.5-alpha)

> smart-test-patch.sh hangs for new version of patch (2.7.1)
> --
>
> Key: HADOOP-9614
> URL: https://issues.apache.org/jira/browse/HADOOP-9614
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Ravi Prakash
>Assignee: Ravi Prakash
> Fix For: 3.0.0, 0.23.8
>
> Attachments: HADOOP-9614.patch, HADOOP-9614.patch
>
>
> patch -p0 -E --dry-run prints "checking file " for the new version of 
> patch(2.7.1) rather than "patching file" as it did for older versions. This 
> causes TMP2 to become empty, which causes the script to hang on this command 
> forever:
> PREFIX_DIRS_AND_FILES=$(cut -d '/' -f 1 | sort | uniq)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8717) JAVA_HOME detected in hadoop-config.sh under OS X does not work

2013-06-04 Thread Josh Rosen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675384#comment-13675384
 ] 

Josh Rosen commented on HADOOP-8717:


The attached patch fixed things for me, but I just thought I'd mention that 
this affects that Hadoop version, too.

> JAVA_HOME detected in hadoop-config.sh under OS X does not work
> ---
>
> Key: HADOOP-8717
> URL: https://issues.apache.org/jira/browse/HADOOP-8717
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
> Environment: OS: Darwin 11.4.0 Darwin Kernel Version 11.4.0: Mon Apr  
> 9 19:32:15 PDT 2012; root:xnu-1699.26.8~1/RELEASE_X86_64 x86_64
> java version "1.6.0_33"
> Java(TM) SE Runtime Environment (build 1.6.0_33-b03-424-11M3720)
> Java HotSpot(TM) 64-Bit Server VM (build 20.8-b03-424, mixed mode)
>Reporter: Jianbin Wei
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HADOOP-8717.patch, HADOOP-8717.patch, HADOOP-8717.patch, 
> HADOOP-8717.patch
>
>
> After setting up a single node hadoop on mac, copy some text file to it and 
> run
> $ hadoop jar 
> ./share/hadoop/mapreduce/hadoop-mapreduce-examples-3.0.0-SNAPSHOT.jar  
> wordcount /file.txt output
> It reports
> 12/08/21 15:32:18 INFO Job.java:mapreduce.Job:1265: Running job: 
> job_1345588312126_0001
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1286: Job 
> job_1345588312126_0001 running in uber mode : false
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1293:  map 0% reduce 0%
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1306: Job 
> job_1345588312126_0001 failed with state FAILED due to: Application 
> application_1345588312126_0001 failed 1 times due to AM Container for 
> appattempt_1345588312126_0001_01 exited with  exitCode: 127 due to: 
> .Failing this attempt.. Failing the application.
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1311: Counters: 0
> $ cat 
> /tmp/logs/application_1345588312126_0001/container_1345588312126_0001_01_01/stderr
> /bin/bash: /bin/java: No such file or directory
> The detected JAVA_HOME is not used somehow.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8717) JAVA_HOME detected in hadoop-config.sh under OS X does not work

2013-06-04 Thread Josh Rosen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675382#comment-13675382
 ] 

Josh Rosen commented on HADOOP-8717:


I just ran into this same issue with Hadoop 2.0.4-alpha on OS X 10.8.3.

> JAVA_HOME detected in hadoop-config.sh under OS X does not work
> ---
>
> Key: HADOOP-8717
> URL: https://issues.apache.org/jira/browse/HADOOP-8717
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
> Environment: OS: Darwin 11.4.0 Darwin Kernel Version 11.4.0: Mon Apr  
> 9 19:32:15 PDT 2012; root:xnu-1699.26.8~1/RELEASE_X86_64 x86_64
> java version "1.6.0_33"
> Java(TM) SE Runtime Environment (build 1.6.0_33-b03-424-11M3720)
> Java HotSpot(TM) 64-Bit Server VM (build 20.8-b03-424, mixed mode)
>Reporter: Jianbin Wei
>Priority: Minor
> Fix For: 3.0.0
>
> Attachments: HADOOP-8717.patch, HADOOP-8717.patch, HADOOP-8717.patch, 
> HADOOP-8717.patch
>
>
> After setting up a single node hadoop on mac, copy some text file to it and 
> run
> $ hadoop jar 
> ./share/hadoop/mapreduce/hadoop-mapreduce-examples-3.0.0-SNAPSHOT.jar  
> wordcount /file.txt output
> It reports
> 12/08/21 15:32:18 INFO Job.java:mapreduce.Job:1265: Running job: 
> job_1345588312126_0001
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1286: Job 
> job_1345588312126_0001 running in uber mode : false
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1293:  map 0% reduce 0%
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1306: Job 
> job_1345588312126_0001 failed with state FAILED due to: Application 
> application_1345588312126_0001 failed 1 times due to AM Container for 
> appattempt_1345588312126_0001_01 exited with  exitCode: 127 due to: 
> .Failing this attempt.. Failing the application.
> 12/08/21 15:32:22 INFO Job.java:mapreduce.Job:1311: Counters: 0
> $ cat 
> /tmp/logs/application_1345588312126_0001/container_1345588312126_0001_01_01/stderr
> /bin/bash: /bin/java: No such file or directory
> The detected JAVA_HOME is not used somehow.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9517) Document Hadoop Compatibility

2013-06-04 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated HADOOP-9517:
-

Attachment: hadoop-9517-v2.patch

Updated the patch to:
# differentiate between the kinds of file formats
# list the kinds of HDFS metadata upgrades
# include the policy proposal for proto files

The policies being proposed (from comments in the JIRA) are explicitly labelled 
(Proposal)

At this point, it would be great to make sure the document
# captures all the items that affect compatibility
# the policies for the not-newly-proposed ones are accurate
# the newly proposed policies are reasonable
# improve the presentation, if need be

Once done with the above, it might be a good idea to get this in and address 
items with no policies in subsequent JIRAs or sub-tasks to make sure we discuss 
them in isolation and detail. Thoughts? 

> Document Hadoop Compatibility
> -
>
> Key: HADOOP-9517
> URL: https://issues.apache.org/jira/browse/HADOOP-9517
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Reporter: Arun C Murthy
>Assignee: Karthik Kambatla
> Attachments: hadoop-9517.patch, hadoop-9517.patch, hadoop-9517.patch, 
> hadoop-9517.patch, hadoop-9517-proposal-v1.patch, 
> hadoop-9517-proposal-v1.patch, hadoop-9517-v2.patch
>
>
> As we get ready to call hadoop-2 stable we need to better define 'Hadoop 
> Compatibility'.
> http://wiki.apache.org/hadoop/Compatibility is a start, let's document 
> requirements clearly and completely.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9582) Non-existent file to "hadoop fs -conf" doesn't throw error

2013-06-04 Thread Ashwin Shankar (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9582?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675372#comment-13675372
 ] 

Ashwin Shankar commented on HADOOP-9582:


Hi Jason,
Thanks for your comments. Yes this patch doesn't address the problem in other 
places apart from FsShell. But could we check this in so that we are good 
atleast from the FsShell standpoint. Maybe we could create another JIRA to 
investigate whether we break any use-case if we make the change directly in 
GenericOptionsParser.   

> Non-existent file to "hadoop fs -conf" doesn't throw error
> --
>
> Key: HADOOP-9582
> URL: https://issues.apache.org/jira/browse/HADOOP-9582
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: conf
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Ashwin Shankar
> Attachments: HADOOP-9582.txt, HADOOP-9582.txt, HADOOP-9582.txt
>
>
> When we run :
> hadoop fs -conf BAD_FILE -ls /
> we expect hadoop to throw an error,but it doesn't.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9487) Deprecation warnings in Configuration should go to their own log or otherwise be suppressible

2013-06-04 Thread Chu Tong (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675334#comment-13675334
 ] 

Chu Tong commented on HADOOP-9487:
--

As I am still kind of new here, please correct me if I am wrong. I am proposing 
to add a new boolean field in mapred-site.xml to suppress the warnings if user 
wants? any suggestions?

> Deprecation warnings in Configuration should go to their own log or otherwise 
> be suppressible
> -
>
> Key: HADOOP-9487
> URL: https://issues.apache.org/jira/browse/HADOOP-9487
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: conf
>Affects Versions: 3.0.0
>Reporter: Steve Loughran
>
> Running local pig jobs triggers large quantities of warnings about deprecated 
> properties -something I don't care about as I'm not in a position to fix 
> without delving into Pig. 
> I can suppress them by changing the log level, but that can hide other 
> warnings that may actually matter.
> If there was a special Configuration.deprecated log for all deprecation 
> messages, this log could be suppressed by people who don't want noisy logs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9392) Token based authentication and Single Sign On

2013-06-04 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675264#comment-13675264
 ] 

Sanjay Radia commented on HADOOP-9392:
--

Thanks for the Jira and the slides on what you are proposing.
{quote}
There is also no consistent delegation model. HDFS has a simple delegation 
capability, and only Oozie can take limited advantage of it. We will implement 
a common token based authentication framework to decouple internal user and 
service authentication from external mechanisms used to support it (like 
Kerberos)
{quote}
I am puzzled by the above statement. Hadoop has delegation tokens and a trust 
model. For the largest part we use delegation tokens (e.g. MR job client gets 
HDFS delegation tokens etc) so that the job can run as the user that submitted 
the job. Further in some cases we use trusted proxies like Ozzie (but this can 
be any trusted service, not just Oozie), to access system services as specific 
users. The delegation tokens and the trusted proxies are two independent 
mechanisms. So I feel the statements in the quoted block are not correct or 
perhaps you are using the term "delegation" in a different sense. Details of 
the Hadoop delegation tokens are in the following very detailed paper on Hadoop 
security (see 
http://hortonworks.com/wp-content/uploads/2011/10/security-design_withCover-1.pdf).

You also state  "We will implement a common token based authentication 
framework to decouple internal user and service authentication from external 
mechanisms used to support it (like Kerberos)" Note the at internal Hadoop 
tokens are separate from the Kerberos tokens - indeed they are nicely decoupled 
- the problem, IMHO,  is not the decoupling but other issues. Another issue is  
the authentication implementation which has a burnt-in notion of supporting 
UGI, Keberos or the delegation tokens for authentication. As Daryn pointed out 
this implementation needs to change so that the authentication mechanism is 
pluggable. This jira, I believe, is proposing much much more than making this 
implementation pluggable. 


Don't get me wrong, I am not criticizing the  Jira but merely trying to 
understand some of the statements in the description and the slides you have 
posted.  I do agree that we need to allow other forms of authentication besides 
Kerberos/ActiveDir. I also agree that attributes like group membership should 
have been part of the Hadoop delegation tokens to avoid back calls which are 
not feasible in cloud environments. Do you have a more detailed design besides 
the slides that you have uploaded to this jira?  I would like to get the next 
level of details. Your comments in this jira do give some more details but it 
would be good to put them in a design document. Further, I suspect you are 
trying to replace the Hadoop delegation tokens - i don't disagree with that but 
would like to understand the why and how   from your perspective.

Would this be an accurate description of this Jira: "Single sign on for 
non-kerberos environments using tokens". Hadoop does support "single sign on" 
for kerberos/activeDir environments; of course that is not good enough since 
many customers do not have Kerberos/ActiveDir. 

> Token based authentication and Single Sign On
> -
>
> Key: HADOOP-9392
> URL: https://issues.apache.org/jira/browse/HADOOP-9392
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Kai Zheng
>Assignee: Kai Zheng
> Fix For: 3.0.0
>
> Attachments: token-based-authn-plus-sso.pdf
>
>
> This is an umbrella entry for one of project Rhino’s topic, for details of 
> project Rhino, please refer to 
> https://github.com/intel-hadoop/project-rhino/. The major goal for this entry 
> as described in project Rhino was 
>  
> “Core, HDFS, ZooKeeper, and HBase currently support Kerberos authentication 
> at the RPC layer, via SASL. However this does not provide valuable attributes 
> such as group membership, classification level, organizational identity, or 
> support for user defined attributes. Hadoop components must interrogate 
> external resources for discovering these attributes and at scale this is 
> problematic. There is also no consistent delegation model. HDFS has a simple 
> delegation capability, and only Oozie can take limited advantage of it. We 
> will implement a common token based authentication framework to decouple 
> internal user and service authentication from external mechanisms used to 
> support it (like Kerberos)”
>  
> We’d like to start our work from Hadoop-Common and try to provide common 
> facilities by extending existing authentication framework which support:
> 1.Pluggable token provider interface 
> 2.Pluggable token verification

[jira] [Commented] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675235#comment-13675235
 ] 

Hadoop QA commented on HADOOP-9618:
---

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12586143/hadoop-9618.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/2599//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/2599//console

This message is automatically generated.

> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9585) unit test failure :org.apache.hadoop.fs.TestFsShellReturnCode.testChgrp

2013-06-04 Thread Sanjay Radia (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9585?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675201#comment-13675201
 ] 

Sanjay Radia commented on HADOOP-9585:
--

The error message for non existing files is normal. Appears that the only 
failed test is the chgrp. Is the user id jenkins  a member of the group admin? 

> unit test failure :org.apache.hadoop.fs.TestFsShellReturnCode.testChgrp
> ---
>
> Key: HADOOP-9585
> URL: https://issues.apache.org/jira/browse/HADOOP-9585
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: test
>Affects Versions: 1.3.0
> Environment: 
> https://builds.apache.org/job/Hadoop-branch1/lastCompletedBuild/testReport/org.apache.hadoop.fs/TestFsShellReturnCode/testChgrp/
>Reporter: Giridharan Kesavan
>
> Standard Error
> chmod: could not get status for 
> '/home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChmod/fileDoesNotExist':
>  File 
> /home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChmod/fileDoesNotExist
>  does not exist.
> chmod: could not get status for 
> '/home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChmod/nonExistingfiles*'
> chown: could not get status for 
> '/home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChown/fileDoesNotExist':
>  File 
> /home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChown/fileDoesNotExist
>  does not exist.
> chown: could not get status for 
> '/home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChown/nonExistingfiles*'
> chgrp: failed on 
> 'file:/home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChgrp/fileExists':
>  chgrp: changing group of 
> `/home/jenkins/jenkins-slave/workspace/Hadoop-branch1/trunk/build/test/data/testChgrp/fileExists':
>  Operation not permitted

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9487) Deprecation warnings in Configuration should go to their own log or otherwise be suppressible

2013-06-04 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675179#comment-13675179
 ] 

Sangjin Lee commented on HADOOP-9487:
-

I'm not suggesting an implementation yet. :) Could be a boolean variable in the 
config or some other flag... If it is part of the config, I presume one needs 
to make sure we don't have a circular logic (not that I think there is one).

> Deprecation warnings in Configuration should go to their own log or otherwise 
> be suppressible
> -
>
> Key: HADOOP-9487
> URL: https://issues.apache.org/jira/browse/HADOOP-9487
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: conf
>Affects Versions: 3.0.0
>Reporter: Steve Loughran
>
> Running local pig jobs triggers large quantities of warnings about deprecated 
> properties -something I don't care about as I'm not in a position to fix 
> without delving into Pig. 
> I can suppress them by changing the log level, but that can hide other 
> warnings that may actually matter.
> If there was a special Configuration.deprecated log for all deprecation 
> messages, this log could be suppressed by people who don't want noisy logs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9392) Token based authentication and Single Sign On

2013-06-04 Thread Kevin Minder (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675168#comment-13675168
 ] 

Kevin Minder commented on HADOOP-9392:
--

Logistics for remote attendance will also be announce publicly when we have 
that figured out.  We won't be making any decisions about security at either 
any prep or the Summit sessions and detailed summaries will be provided here 
for those that cannot attend. 

> Token based authentication and Single Sign On
> -
>
> Key: HADOOP-9392
> URL: https://issues.apache.org/jira/browse/HADOOP-9392
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Kai Zheng
>Assignee: Kai Zheng
> Fix For: 3.0.0
>
> Attachments: token-based-authn-plus-sso.pdf
>
>
> This is an umbrella entry for one of project Rhino’s topic, for details of 
> project Rhino, please refer to 
> https://github.com/intel-hadoop/project-rhino/. The major goal for this entry 
> as described in project Rhino was 
>  
> “Core, HDFS, ZooKeeper, and HBase currently support Kerberos authentication 
> at the RPC layer, via SASL. However this does not provide valuable attributes 
> such as group membership, classification level, organizational identity, or 
> support for user defined attributes. Hadoop components must interrogate 
> external resources for discovering these attributes and at scale this is 
> problematic. There is also no consistent delegation model. HDFS has a simple 
> delegation capability, and only Oozie can take limited advantage of it. We 
> will implement a common token based authentication framework to decouple 
> internal user and service authentication from external mechanisms used to 
> support it (like Kerberos)”
>  
> We’d like to start our work from Hadoop-Common and try to provide common 
> facilities by extending existing authentication framework which support:
> 1.Pluggable token provider interface 
> 2.Pluggable token verification protocol and interface
> 3.Security mechanism to distribute secrets in cluster nodes
> 4.Delegation model of user authentication

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9533) Centralized Hadoop SSO/Token Server

2013-06-04 Thread Kevin Minder (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9533?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675167#comment-13675167
 ] 

Kevin Minder commented on HADOOP-9533:
--

Logistics for remote attendance will also be announce publicly when we have 
that figured out.  We won't be making any decisions about security at either 
any prep or the Summit sessions and detailed summaries will be provided here 
for those that cannot attend. 

> Centralized Hadoop SSO/Token Server
> ---
>
> Key: HADOOP-9533
> URL: https://issues.apache.org/jira/browse/HADOOP-9533
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Larry McCay
> Attachments: HSSO-Interaction-Overview-rev-1.docx, 
> HSSO-Interaction-Overview-rev-1.pdf
>
>
> This is an umbrella Jira filing to oversee a set of proposals for introducing 
> a new master service for Hadoop Single Sign On (HSSO).
> There is an increasing need for pluggable authentication providers that 
> authenticate both users and services as well as validate tokens in order to 
> federate identities authenticated by trusted IDPs. These IDPs may be deployed 
> within the enterprise or third-party IDPs that are external to the enterprise.
> These needs speak to a specific pain point: which is a narrow integration 
> path into the enterprise identity infrastructure. Kerberos is a fine solution 
> for those that already have it in place or are willing to adopt its use but 
> there remains a class of user that finds this unacceptable and needs to 
> integrate with a wider variety of identity management solutions.
> Another specific pain point is that of rolling and distributing keys. A 
> related and integral part of the HSSO server is library called the Credential 
> Management Framework (CMF), which will be a common library for easing the 
> management of secrets, keys and credentials.
> Initially, the existing delegation, block access and job tokens will continue 
> to be utilized. There may be some changes required to leverage a PKI based 
> signature facility rather than shared secrets. This is a means to simplify 
> the solution for the pain point of distributing shared secrets.
> This project will primarily centralize the responsibility of authentication 
> and federation into a single service that is trusted across the Hadoop 
> cluster and optionally across multiple clusters. This greatly simplifies a 
> number of things in the Hadoop ecosystem:
> 1.a single token format that is used across all of Hadoop regardless of 
> authentication method
> 2.a single service to have pluggable providers instead of all services
> 3.a single token authority that would be trusted across the cluster/s and 
> through PKI encryption be able to easily issue cryptographically verifiable 
> tokens
> 4.automatic rolling of the token authority’s keys and publishing of the 
> public key for easy access by those parties that need to verify incoming 
> tokens
> 5.use of PKI for signatures eliminates the need for securely sharing and 
> distributing shared secrets
> In addition to serving as the internal Hadoop SSO service this service will 
> be leveraged by the Knox Gateway from the cluster perimeter in order to 
> acquire the Hadoop cluster tokens. The same token mechanism that is used for 
> internal services will be used to represent user identities. Providing for 
> interesting scenarios such as SSO across Hadoop clusters within an enterprise 
> and/or into the cloud.
> The HSSO service will be comprised of three major components and capabilities:
> 1.Federating IDP – authenticates users/services and issues the common 
> Hadoop token
> 2.Federating SP – validates the token of trusted external IDPs and issues 
> the common Hadoop token
> 3.Token Authority – management of the common Hadoop tokens – including: 
> a.Issuance 
> b.Renewal
> c.Revocation
> As this is a meta Jira for tracking this overall effort, the details of the 
> individual efforts will be submitted along with the child Jira filings.
> Hadoop-Common would seem to be the most appropriate home for such a service 
> and its related common facilities. We will also leverage and extend existing 
> common mechanisms as appropriate.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9533) Centralized Hadoop SSO/Token Server

2013-06-04 Thread Kevin Minder (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9533?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675130#comment-13675130
 ] 

Kevin Minder commented on HADOOP-9533:
--

I'm happy to announce that we have secured a time slot and dedicated space 
during Hadoop Summit NA dedicated to forward looking Hadoop security design 
collaboration.  Currently, a room has been allocated on the 26th from 1:45 to 
3:30 PT.  Specific location will be available at the Summit and any changes in 
date or time will be announced publicly to the best of our abilities.  In order 
to create a manageable agenda for this session, I'd like to schedule some prep 
meetings via meetup.com to start discussions and preparations with those that 
would be interested in co-organizing the session.

> Centralized Hadoop SSO/Token Server
> ---
>
> Key: HADOOP-9533
> URL: https://issues.apache.org/jira/browse/HADOOP-9533
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Larry McCay
> Attachments: HSSO-Interaction-Overview-rev-1.docx, 
> HSSO-Interaction-Overview-rev-1.pdf
>
>
> This is an umbrella Jira filing to oversee a set of proposals for introducing 
> a new master service for Hadoop Single Sign On (HSSO).
> There is an increasing need for pluggable authentication providers that 
> authenticate both users and services as well as validate tokens in order to 
> federate identities authenticated by trusted IDPs. These IDPs may be deployed 
> within the enterprise or third-party IDPs that are external to the enterprise.
> These needs speak to a specific pain point: which is a narrow integration 
> path into the enterprise identity infrastructure. Kerberos is a fine solution 
> for those that already have it in place or are willing to adopt its use but 
> there remains a class of user that finds this unacceptable and needs to 
> integrate with a wider variety of identity management solutions.
> Another specific pain point is that of rolling and distributing keys. A 
> related and integral part of the HSSO server is library called the Credential 
> Management Framework (CMF), which will be a common library for easing the 
> management of secrets, keys and credentials.
> Initially, the existing delegation, block access and job tokens will continue 
> to be utilized. There may be some changes required to leverage a PKI based 
> signature facility rather than shared secrets. This is a means to simplify 
> the solution for the pain point of distributing shared secrets.
> This project will primarily centralize the responsibility of authentication 
> and federation into a single service that is trusted across the Hadoop 
> cluster and optionally across multiple clusters. This greatly simplifies a 
> number of things in the Hadoop ecosystem:
> 1.a single token format that is used across all of Hadoop regardless of 
> authentication method
> 2.a single service to have pluggable providers instead of all services
> 3.a single token authority that would be trusted across the cluster/s and 
> through PKI encryption be able to easily issue cryptographically verifiable 
> tokens
> 4.automatic rolling of the token authority’s keys and publishing of the 
> public key for easy access by those parties that need to verify incoming 
> tokens
> 5.use of PKI for signatures eliminates the need for securely sharing and 
> distributing shared secrets
> In addition to serving as the internal Hadoop SSO service this service will 
> be leveraged by the Knox Gateway from the cluster perimeter in order to 
> acquire the Hadoop cluster tokens. The same token mechanism that is used for 
> internal services will be used to represent user identities. Providing for 
> interesting scenarios such as SSO across Hadoop clusters within an enterprise 
> and/or into the cloud.
> The HSSO service will be comprised of three major components and capabilities:
> 1.Federating IDP – authenticates users/services and issues the common 
> Hadoop token
> 2.Federating SP – validates the token of trusted external IDPs and issues 
> the common Hadoop token
> 3.Token Authority – management of the common Hadoop tokens – including: 
> a.Issuance 
> b.Renewal
> c.Revocation
> As this is a meta Jira for tracking this overall effort, the details of the 
> individual efforts will be submitted along with the child Jira filings.
> Hadoop-Common would seem to be the most appropriate home for such a service 
> and its related common facilities. We will also leverage and extend existing 
> common mechanisms as appropriate.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atl

[jira] [Commented] (HADOOP-9392) Token based authentication and Single Sign On

2013-06-04 Thread Kevin Minder (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13675132#comment-13675132
 ] 

Kevin Minder commented on HADOOP-9392:
--

I'm happy to announce that we have secured a time slot and dedicated space 
during Hadoop Summit NA dedicated to forward looking Hadoop security design 
collaboration.  Currently, a room has been allocated on the 26th from 1:45 to 
3:30 PT.  Specific location will be available at the Summit and any changes in 
date or time will be announced publicly to the best of our abilities.  In order 
to create a manageable agenda for this session, I'd like to schedule some prep 
meetings via meetup.com to start discussions and preparations with those that 
would be interested in co-organizing the session.


> Token based authentication and Single Sign On
> -
>
> Key: HADOOP-9392
> URL: https://issues.apache.org/jira/browse/HADOOP-9392
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Kai Zheng
>Assignee: Kai Zheng
> Fix For: 3.0.0
>
> Attachments: token-based-authn-plus-sso.pdf
>
>
> This is an umbrella entry for one of project Rhino’s topic, for details of 
> project Rhino, please refer to 
> https://github.com/intel-hadoop/project-rhino/. The major goal for this entry 
> as described in project Rhino was 
>  
> “Core, HDFS, ZooKeeper, and HBase currently support Kerberos authentication 
> at the RPC layer, via SASL. However this does not provide valuable attributes 
> such as group membership, classification level, organizational identity, or 
> support for user defined attributes. Hadoop components must interrogate 
> external resources for discovering these attributes and at scale this is 
> problematic. There is also no consistent delegation model. HDFS has a simple 
> delegation capability, and only Oozie can take limited advantage of it. We 
> will implement a common token based authentication framework to decouple 
> internal user and service authentication from external mechanisms used to 
> support it (like Kerberos)”
>  
> We’d like to start our work from Hadoop-Common and try to provide common 
> facilities by extending existing authentication framework which support:
> 1.Pluggable token provider interface 
> 2.Pluggable token verification protocol and interface
> 3.Security mechanism to distribute secrets in cluster nodes
> 4.Delegation model of user authentication

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HADOOP-9618:


Status: Patch Available  (was: Open)

> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674684#comment-13674684
 ] 

Todd Lipcon commented on HADOOP-9618:
-

BTW, the info reported from the beans seems to be off due to an OpenJDK bug. 
When I run the same test program with Oracle JDK 1.6.0_14 I get correct stats 
from the CMS MXBean:

13/06/04 11:36:33 INFO util.JvmPauseMonitor: Detected pause in JVM or host 
machine (eg GC): pause of approximately 3232ms
GC pool 'ParNew' had collection(s): count=1 time=56ms
GC pool 'ConcurrentMarkSweep' had collection(s): count=1 time=3665ms


> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9517) Document Hadoop Compatibility

2013-06-04 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674683#comment-13674683
 ] 

Karthik Kambatla commented on HADOOP-9517:
--

bq. Classes that subclass a Hadoop class to provide a plugin point MAY need 
recompiling on each major version, possibly with the handling of changes to 
methods.

If the Hadoop class being extended is Public and the stability is defined by 
the annotation, is that not sufficient indication to the user that it might 
need to be changed as that interface/class changes. For example, we recently 
added a SchedulingPolicy to FairScheduler annotated Public-Evolving: the 
policies written for the current version need to be updated as and when the 
SchedulingPolicy class changes. Once it becomes stable, we follow the standard 
deprecation rules for Public-Stable API that protects the policies. No?

I think it is important to detail how the API compatibility rules impact the 
user-level code. May be I am missing something here. Otherwise, we might not 
need specific policies for them?

> Document Hadoop Compatibility
> -
>
> Key: HADOOP-9517
> URL: https://issues.apache.org/jira/browse/HADOOP-9517
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Reporter: Arun C Murthy
>Assignee: Karthik Kambatla
> Attachments: hadoop-9517.patch, hadoop-9517.patch, hadoop-9517.patch, 
> hadoop-9517.patch, hadoop-9517-proposal-v1.patch, 
> hadoop-9517-proposal-v1.patch
>
>
> As we get ready to call hadoop-2 stable we need to better define 'Hadoop 
> Compatibility'.
> http://wiki.apache.org/hadoop/Compatibility is a start, let's document 
> requirements clearly and completely.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Todd Lipcon (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Todd Lipcon updated HADOOP-9618:


Attachment: hadoop-9618.txt

Attached patch implements this feature. I wasn't able to write an automated 
unit test for this, since it relies on forcing the JVM to pause, which isn't 
really doable in a junit context. I added a main function, though, which leaks 
memory and forces into GC. Here's the output from my machine:

{code}
todd@todd-w510:~/git/hadoop-common/hadoop-common-project/hadoop-common$ java 
-verbose:gc -Xmx1g -XX:+UseConcMarkSweepGC -cp 
/home/todd/confs/devconf.2nn-2/:target/classes/:/home/todd/git/hadoop-common/hadoop-dist/target/hadoop-2.0.0-cdh4.1.2/share/hadoop/common/lib/\*
 org.apache.hadoop.util.JvmPauseMonitor
WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please use 
org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties files.
[GC 16715K->9327K(83008K), 0.0964960 secs]
[GC 26351K->23551K(83008K), 0.0406780 secs]
[GC 40575K->38828K(83008K), 0.0375710 secs]
[GC 39126K(83008K), 0.0027320 secs]
[GC 55852K->55851K(83008K), 0.0422420 secs]
[GC 70138K->69911K(88000K), 0.0437410 secs]
[GC 86935K->86934K(105028K), 0.0479130 secs]
[GC 103958K->103957K(121924K), 0.0572590 secs]
[GC 120981K->120804K(138824K), 0.0521890 secs]
[GC 137828K->137826K(155720K), 0.0508530 secs]
[GC 154850K->154848K(172808K), 0.0519540 secs]
[GC 163351K->163177K(181064K), 0.0307160 secs]
[GC 180201K->180201K(198160K), 0.0554300 secs]
[GC 197225K->197223K(215248K), 0.0660260 secs]
[GC 214247K->214244K(232144K), 0.0611980 secs]
[GC 231268K->231151K(249040K), 0.0543830 secs]
[GC 247289K->247117K(265168K), 0.0616800 secs]
[Full GC 247117K->225823K(265168K), 1.0242170 secs]
13/06/04 11:32:14 INFO util.JvmPauseMonitor: Detected pause in JVM or host 
machine (eg GC): pause of approximately 1145ms
GC pool 'ParNew' had collection(s): count=2 time=117ms
GC pool 'ConcurrentMarkSweep' had collection(s): count=2 time=445ms
[GC 268895K->268138K(424824K), 0.1054190 secs]
[GC 268825K(424824K), 0.0061170 secs]
[GC 311210K->311208K(424824K), 0.1476640 secs]
[GC 352421K->351750K(424824K), 0.1371280 secs]
[GC 394822K->394691K(454648K), 0.1522430 secs]
[GC 437763K->437760K(481144K), 0.1447820 secs]
[GC 480832K->486355K(529720K), 0.1565950 secs]
[GC 529427K->529894K(573304K), 0.1264970 secs]
[GC 547400K->546694K(590008K), 0.0609490 secs]
[Full GC 546694K->507462K(590008K), 2.6322880 secs]
13/06/04 11:32:19 INFO util.JvmPauseMonitor: Detected pause in JVM or host 
machine (eg GC): pause of approximately 2735ms
GC pool 'ParNew' had collection(s): count=2 time=188ms
GC pool 'ConcurrentMarkSweep' had collection(s): count=2 time=298ms
{code}

> Add thread which detects JVM pauses
> ---
>
> Key: HADOOP-9618
> URL: https://issues.apache.org/jira/browse/HADOOP-9618
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: util
>Affects Versions: 3.0.0
>Reporter: Todd Lipcon
>Assignee: Todd Lipcon
> Attachments: hadoop-9618.txt
>
>
> Often times users struggle to understand what happened when a long JVM pause 
> (GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
> example, a long GC pause while logging an edit to the QJM may cause the edit 
> to timeout, or a long GC pause may make other IPCs to the NameNode timeout. 
> We should add a simple thread which loops on 1-second sleeps, and if the 
> sleep ever takes significantly longer than 1 second, log a WARN. This will 
> make GC pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HADOOP-9618) Add thread which detects JVM pauses

2013-06-04 Thread Todd Lipcon (JIRA)
Todd Lipcon created HADOOP-9618:
---

 Summary: Add thread which detects JVM pauses
 Key: HADOOP-9618
 URL: https://issues.apache.org/jira/browse/HADOOP-9618
 Project: Hadoop Common
  Issue Type: New Feature
  Components: util
Affects Versions: 3.0.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon


Often times users struggle to understand what happened when a long JVM pause 
(GC or otherwise) causes things to malfunction inside a Hadoop daemon. For 
example, a long GC pause while logging an edit to the QJM may cause the edit to 
timeout, or a long GC pause may make other IPCs to the NameNode timeout. We 
should add a simple thread which loops on 1-second sleeps, and if the sleep 
ever takes significantly longer than 1 second, log a WARN. This will make GC 
pauses obvious in logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9487) Deprecation warnings in Configuration should go to their own log or otherwise be suppressible

2013-06-04 Thread Chu Tong (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674576#comment-13674576
 ] 

Chu Tong commented on HADOOP-9487:
--

[~sjlee0], so I assume this boolean variable has to be specified in one of the 
configuration files?

> Deprecation warnings in Configuration should go to their own log or otherwise 
> be suppressible
> -
>
> Key: HADOOP-9487
> URL: https://issues.apache.org/jira/browse/HADOOP-9487
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: conf
>Affects Versions: 3.0.0
>Reporter: Steve Loughran
>
> Running local pig jobs triggers large quantities of warnings about deprecated 
> properties -something I don't care about as I'm not in a position to fix 
> without delving into Pig. 
> I can suppress them by changing the log level, but that can hide other 
> warnings that may actually matter.
> If there was a special Configuration.deprecated log for all deprecation 
> messages, this log could be suppressed by people who don't want noisy logs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9487) Deprecation warnings in Configuration should go to their own log or otherwise be suppressible

2013-06-04 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674568#comment-13674568
 ] 

Sangjin Lee commented on HADOOP-9487:
-

[~stayhf] I'm not sure which part of my comment you're referring to... When we 
run any mapreduce job that uses old properties, the log is flooded with the 
deprecation warnings to the point that it is often hard to make out other legit 
warnings or errors. I support the idea that provides an ability to suppress 
these deprecation warning messages if a certain boolean option is enabled.

> Deprecation warnings in Configuration should go to their own log or otherwise 
> be suppressible
> -
>
> Key: HADOOP-9487
> URL: https://issues.apache.org/jira/browse/HADOOP-9487
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: conf
>Affects Versions: 3.0.0
>Reporter: Steve Loughran
>
> Running local pig jobs triggers large quantities of warnings about deprecated 
> properties -something I don't care about as I'm not in a position to fix 
> without delving into Pig. 
> I can suppress them by changing the log level, but that can hide other 
> warnings that may actually matter.
> If there was a special Configuration.deprecated log for all deprecation 
> messages, this log could be suppressed by people who don't want noisy logs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9487) Deprecation warnings in Configuration should go to their own log or otherwise be suppressible

2013-06-04 Thread Chu Tong (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674560#comment-13674560
 ] 

Chu Tong commented on HADOOP-9487:
--

[~sjlee0], your idea seems a bit vague to me, can you please elaborate? thanks

> Deprecation warnings in Configuration should go to their own log or otherwise 
> be suppressible
> -
>
> Key: HADOOP-9487
> URL: https://issues.apache.org/jira/browse/HADOOP-9487
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: conf
>Affects Versions: 3.0.0
>Reporter: Steve Loughran
>
> Running local pig jobs triggers large quantities of warnings about deprecated 
> properties -something I don't care about as I'm not in a position to fix 
> without delving into Pig. 
> I can suppress them by changing the log level, but that can hide other 
> warnings that may actually matter.
> If there was a special Configuration.deprecated log for all deprecation 
> messages, this log could be suppressed by people who don't want noisy logs

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9533) Centralized Hadoop SSO/Token Server

2013-06-04 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9533?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674475#comment-13674475
 ] 

Daryn Sharp commented on HADOOP-9533:
-

Initial thoughts are passing a bearer token over the network creates a weakest 
link  - it's dangerous if a rogue or untrustworthy service is involved.  SASL 
can be leveraged to avoid passing the actual bearer token over the network, 
although that might complicate the desired design.

The nice thing about the SASL abstraction is mechanisms are easily swapped out 
w/o writing code.  While I'm not opposed to SSL/TLS, relying solely on SSL (at 
least openssl) for authentication may not be a good idea since it seems to be 
rife with security flaws.  I'd consider SSL an extra security layer, not the 
only security layer.

> Centralized Hadoop SSO/Token Server
> ---
>
> Key: HADOOP-9533
> URL: https://issues.apache.org/jira/browse/HADOOP-9533
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Larry McCay
> Attachments: HSSO-Interaction-Overview-rev-1.docx, 
> HSSO-Interaction-Overview-rev-1.pdf
>
>
> This is an umbrella Jira filing to oversee a set of proposals for introducing 
> a new master service for Hadoop Single Sign On (HSSO).
> There is an increasing need for pluggable authentication providers that 
> authenticate both users and services as well as validate tokens in order to 
> federate identities authenticated by trusted IDPs. These IDPs may be deployed 
> within the enterprise or third-party IDPs that are external to the enterprise.
> These needs speak to a specific pain point: which is a narrow integration 
> path into the enterprise identity infrastructure. Kerberos is a fine solution 
> for those that already have it in place or are willing to adopt its use but 
> there remains a class of user that finds this unacceptable and needs to 
> integrate with a wider variety of identity management solutions.
> Another specific pain point is that of rolling and distributing keys. A 
> related and integral part of the HSSO server is library called the Credential 
> Management Framework (CMF), which will be a common library for easing the 
> management of secrets, keys and credentials.
> Initially, the existing delegation, block access and job tokens will continue 
> to be utilized. There may be some changes required to leverage a PKI based 
> signature facility rather than shared secrets. This is a means to simplify 
> the solution for the pain point of distributing shared secrets.
> This project will primarily centralize the responsibility of authentication 
> and federation into a single service that is trusted across the Hadoop 
> cluster and optionally across multiple clusters. This greatly simplifies a 
> number of things in the Hadoop ecosystem:
> 1.a single token format that is used across all of Hadoop regardless of 
> authentication method
> 2.a single service to have pluggable providers instead of all services
> 3.a single token authority that would be trusted across the cluster/s and 
> through PKI encryption be able to easily issue cryptographically verifiable 
> tokens
> 4.automatic rolling of the token authority’s keys and publishing of the 
> public key for easy access by those parties that need to verify incoming 
> tokens
> 5.use of PKI for signatures eliminates the need for securely sharing and 
> distributing shared secrets
> In addition to serving as the internal Hadoop SSO service this service will 
> be leveraged by the Knox Gateway from the cluster perimeter in order to 
> acquire the Hadoop cluster tokens. The same token mechanism that is used for 
> internal services will be used to represent user identities. Providing for 
> interesting scenarios such as SSO across Hadoop clusters within an enterprise 
> and/or into the cloud.
> The HSSO service will be comprised of three major components and capabilities:
> 1.Federating IDP – authenticates users/services and issues the common 
> Hadoop token
> 2.Federating SP – validates the token of trusted external IDPs and issues 
> the common Hadoop token
> 3.Token Authority – management of the common Hadoop tokens – including: 
> a.Issuance 
> b.Renewal
> c.Revocation
> As this is a meta Jira for tracking this overall effort, the details of the 
> individual efforts will be submitted along with the child Jira filings.
> Hadoop-Common would seem to be the most appropriate home for such a service 
> and its related common facilities. We will also leverage and extend existing 
> common mechanisms as appropriate.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIR

[jira] [Commented] (HADOOP-9481) Broken conditional logic with HADOOP_SNAPPY_LIBRARY

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674401#comment-13674401
 ] 

Hudson commented on HADOOP-9481:


Integrated in Hadoop-Mapreduce-trunk #1446 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1446/])
HADOOP-9481. Move from trunk to release 2.1.0 section (Revision 1489261)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489261
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Broken conditional logic with HADOOP_SNAPPY_LIBRARY
> ---
>
> Key: HADOOP-9481
> URL: https://issues.apache.org/jira/browse/HADOOP-9481
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.1.0-beta
>Reporter: Vadim Bondarev
>Assignee: Vadim Bondarev
>Priority: Minor
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9481-trunk--N1.patch, HADOOP-9481-trunk--N4.patch
>
>
> The problem is a regression introduced by recent fix 
> https://issues.apache.org/jira/browse/HADOOP-8562 .
> That fix makes some improvements for Windows platform, but breaks native code 
> work on Unix.
> Namely, let's see the diff HADOOP-8562 of the file 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
>  :  
> {noformat}
> --- 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
> +++ 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
> @@ -16,12 +16,18 @@
>   * limitations under the License.
>   */
> -#include 
> +
> +#if defined HADOOP_SNAPPY_LIBRARY
> +
>  #include 
>  #include 
>  #include 
> +#ifdef UNIX
> +#include 
>  #include "config.h"
> +#endif // UNIX
> +
>  #include "org_apache_hadoop_io_compress_snappy.h"
>  #include "org_apache_hadoop_io_compress_snappy_SnappyCompressor.h"
> @@ -81,7 +87,7 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>UNLOCK_CLASS(env, clazz, "SnappyCompressor");
>if (uncompressed_bytes == 0) {
> -return 0;
> +return (jint)0;
>}
>// Get the output direct buffer
> @@ -90,7 +96,7 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>UNLOCK_CLASS(env, clazz, "SnappyCompressor");
>if (compressed_bytes == 0) {
> -return 0;
> +return (jint)0;
>}
>/* size_t should always be 4 bytes or larger. */
> @@ -109,3 +115,5 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>(*env)->SetIntField(env, thisj, SnappyCompressor_uncompressedDirectBufLen, 
> 0);
>return (jint)buf_len;
>  }
> +
> +#endif //define HADOOP_SNAPPY_LIBRARY
> {noformat}
> Here we see that all the class implementation got enclosed into "if defined 
> HADOOP_SNAPPY_LIBRARY" directive, and the point is that 
> "HADOOP_SNAPPY_LIBRARY" is *not* defined. 
> This causes the class implementation to be effectively empty, what, in turn, 
> causes the UnsatisfiedLinkError to be thrown in the runtime upon any attempt 
> to invoke the native methods implemented there.
> The actual intention of the authors of HADOOP-8562 was (as we suppose) to 
> invoke "include config.h", where "HADOOP_SNAPPY_LIBRARY" is defined. But 
> currently it is *not* included because it resides *inside* "if defined 
> HADOOP_SNAPPY_LIBRARY" block.
> Similar situation with "ifdef UNIX", because UNIX or WINDOWS variables are 
> defined in "org_apache_hadoop.h", which is indirectly included through 
> "include "org_apache_hadoop_io_compress_snappy.h"", and in the current code 
> this is done *after* code "ifdef UNIX", so in the current code the block 
> "ifdef UNIX" is *not* executed on UNIX.
> The suggested patch fixes the described problems by reordering the "include" 
> and "if" preprocessor directives accordingly, bringing the methods of class 
> org.apache.hadoop.io.compress.snappy.SnappyCompressor back to work again.
> Of course, Snappy native libraries must be installed to build and invoke 
> snappy native methods.
> (Note: there was a mistype in commit message: 8952 written in place of 8562: 
> HADOOP-8952. Enhancements to support Hadoop on Windows Server and Windows 
> Azure environments. Contributed by Ivan Mitic, Chuan Liu, Ramya Sunil, Bikas 
> Saha, Kanna Karanam, John Gordon, Brandon Li, Chris Nauroth, David Lao, 
> Sumadhur Reddy Bolli, Arpit Agarwal, Ahmed El Baz, Mike Liddell, Jing Zhao, 
> Thejas Nair, Steve Maine, Ganeshan Iyer, Raja Aluri, Giridharan Kesavan, 
> Ramya Bharathi Nimmagadda.
>git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1453486 
> 13f79535-47bb-0310-9956-ffa450edef68
> )

--
This mes

[jira] [Commented] (HADOOP-9614) smart-test-patch.sh hangs for new version of patch (2.7.1)

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9614?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674410#comment-13674410
 ] 

Hudson commented on HADOOP-9614:


Integrated in Hadoop-Mapreduce-trunk #1446 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1446/])
HADOOP-9614. smart-test-patch.sh hangs for new version of patch (2.7.1) 
(Ravi Prakash via jeagles) (Revision 1489136)

 Result = SUCCESS
jeagles : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489136
Files : 
* /hadoop/common/trunk/dev-support/smart-apply-patch.sh
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> smart-test-patch.sh hangs for new version of patch (2.7.1)
> --
>
> Key: HADOOP-9614
> URL: https://issues.apache.org/jira/browse/HADOOP-9614
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Ravi Prakash
>Assignee: Ravi Prakash
> Fix For: 3.0.0, 0.23.8, 2.0.5-alpha
>
> Attachments: HADOOP-9614.patch, HADOOP-9614.patch
>
>
> patch -p0 -E --dry-run prints "checking file " for the new version of 
> patch(2.7.1) rather than "patching file" as it did for older versions. This 
> causes TMP2 to become empty, which causes the script to hang on this command 
> forever:
> PREFIX_DIRS_AND_FILES=$(cut -d '/' -f 1 | sort | uniq)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9287) Parallel testing hadoop-common

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674408#comment-13674408
 ] 

Hudson commented on HADOOP-9287:


Integrated in Hadoop-Mapreduce-trunk #1446 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1446/])
Move HADOOP-9287 in CHANGES.txt after committing to branch-2 (Revision 
1489258)

 Result = SUCCESS
jlowe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489258
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Parallel testing hadoop-common
> --
>
> Key: HADOOP-9287
> URL: https://issues.apache.org/jira/browse/HADOOP-9287
> Project: Hadoop Common
>  Issue Type: Test
>  Components: test
>Affects Versions: 3.0.0
>Reporter: Tsuyoshi OZAWA
>Assignee: Andrey Klochkov
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9287.1.patch, HADOOP-9287-branch-2--N1.patch, 
> HADOOP-9287--N3.patch, HADOOP-9287--N3.patch, HADOOP-9287--N4.patch, 
> HADOOP-9287--N5.patch, HADOOP-9287--N6.patch, HADOOP-9287--N7.patch, 
> HADOOP-9287.patch, HADOOP-9287.patch
>
>
> The maven surefire plugin supports parallel testing feature. By using it, the 
> tests can be run more faster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9397) Incremental dist tar build fails

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9397?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674399#comment-13674399
 ] 

Hudson commented on HADOOP-9397:


Integrated in Hadoop-Mapreduce-trunk #1446 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1446/])
Move HADOOP-9397 to 2.1.0-beta after merging it into branch-2. (Revision 
1489026)

 Result = SUCCESS
jlowe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489026
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Incremental dist tar build fails
> 
>
> Key: HADOOP-9397
> URL: https://issues.apache.org/jira/browse/HADOOP-9397
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 3.0.0
>Reporter: Jason Lowe
>Assignee: Chris Nauroth
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9397.1.patch
>
>
> Building a dist tar build when the dist tarball already exists from a 
> previous build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9615) Hadoop Jar command not working when used with Spring ORM

2013-06-04 Thread Deepa Vasanthkumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674386#comment-13674386
 ] 

Deepa Vasanthkumar commented on HADOOP-9615:



As a workaround, I removed all configuration xmls from the jar file, and placed 
them into the working directory.
This resolved the issue for the timebeing. 


> Hadoop Jar command not working when used with Spring ORM
> 
>
> Key: HADOOP-9615
> URL: https://issues.apache.org/jira/browse/HADOOP-9615
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 2.0.0-alpha
> Environment: CentOS, 
>Reporter: Deepa Vasanthkumar
>  Labels: hadoop-2.0
>
> Unable to invoke 'hadoop jar' command for class, which contains Spring 
> persistance unit.
> The problem is that, the jar file uses Spring ORM for loading the persistance 
> configurations, and based on these configurations,
> i need to move the files to HDFS. 
> While invoking the jar with hadoop jar command (having spring orm injected) 
> the exception is as: 
> Exception in thread "main" 
> org.springframework.beans.factory.BeanCreationException: Error creating bean 
> with name 
> 'org.springframework.dao.annotation.PersistenceExceptionTranslationPostProcessor#0'
>  defined in class path resource [applicationContext.xml
> Error creating bean with name 'entityManagerFactory' defined in class path 
> resource [applicationContext.xml]: 
> Invocation of init method failed; nested exception is 
> java.lang.IllegalStateException: Conflicting persistence unit definitions for 
> name 'Persistance': file:/home/user/Desktop/ABC/apnJar.jar, 
> file:/tmp/hadoop-user/hadoop-unjar2841422106164401019/
> Caused by: java.lang.IllegalStateException: Conflicting persistence unit 
> definitions for name 'Persistance': 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9481) Broken conditional logic with HADOOP_SNAPPY_LIBRARY

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674333#comment-13674333
 ] 

Hudson commented on HADOOP-9481:


Integrated in Hadoop-Hdfs-trunk #1420 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1420/])
HADOOP-9481. Move from trunk to release 2.1.0 section (Revision 1489261)

 Result = FAILURE
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489261
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Broken conditional logic with HADOOP_SNAPPY_LIBRARY
> ---
>
> Key: HADOOP-9481
> URL: https://issues.apache.org/jira/browse/HADOOP-9481
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.1.0-beta
>Reporter: Vadim Bondarev
>Assignee: Vadim Bondarev
>Priority: Minor
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9481-trunk--N1.patch, HADOOP-9481-trunk--N4.patch
>
>
> The problem is a regression introduced by recent fix 
> https://issues.apache.org/jira/browse/HADOOP-8562 .
> That fix makes some improvements for Windows platform, but breaks native code 
> work on Unix.
> Namely, let's see the diff HADOOP-8562 of the file 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
>  :  
> {noformat}
> --- 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
> +++ 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
> @@ -16,12 +16,18 @@
>   * limitations under the License.
>   */
> -#include 
> +
> +#if defined HADOOP_SNAPPY_LIBRARY
> +
>  #include 
>  #include 
>  #include 
> +#ifdef UNIX
> +#include 
>  #include "config.h"
> +#endif // UNIX
> +
>  #include "org_apache_hadoop_io_compress_snappy.h"
>  #include "org_apache_hadoop_io_compress_snappy_SnappyCompressor.h"
> @@ -81,7 +87,7 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>UNLOCK_CLASS(env, clazz, "SnappyCompressor");
>if (uncompressed_bytes == 0) {
> -return 0;
> +return (jint)0;
>}
>// Get the output direct buffer
> @@ -90,7 +96,7 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>UNLOCK_CLASS(env, clazz, "SnappyCompressor");
>if (compressed_bytes == 0) {
> -return 0;
> +return (jint)0;
>}
>/* size_t should always be 4 bytes or larger. */
> @@ -109,3 +115,5 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>(*env)->SetIntField(env, thisj, SnappyCompressor_uncompressedDirectBufLen, 
> 0);
>return (jint)buf_len;
>  }
> +
> +#endif //define HADOOP_SNAPPY_LIBRARY
> {noformat}
> Here we see that all the class implementation got enclosed into "if defined 
> HADOOP_SNAPPY_LIBRARY" directive, and the point is that 
> "HADOOP_SNAPPY_LIBRARY" is *not* defined. 
> This causes the class implementation to be effectively empty, what, in turn, 
> causes the UnsatisfiedLinkError to be thrown in the runtime upon any attempt 
> to invoke the native methods implemented there.
> The actual intention of the authors of HADOOP-8562 was (as we suppose) to 
> invoke "include config.h", where "HADOOP_SNAPPY_LIBRARY" is defined. But 
> currently it is *not* included because it resides *inside* "if defined 
> HADOOP_SNAPPY_LIBRARY" block.
> Similar situation with "ifdef UNIX", because UNIX or WINDOWS variables are 
> defined in "org_apache_hadoop.h", which is indirectly included through 
> "include "org_apache_hadoop_io_compress_snappy.h"", and in the current code 
> this is done *after* code "ifdef UNIX", so in the current code the block 
> "ifdef UNIX" is *not* executed on UNIX.
> The suggested patch fixes the described problems by reordering the "include" 
> and "if" preprocessor directives accordingly, bringing the methods of class 
> org.apache.hadoop.io.compress.snappy.SnappyCompressor back to work again.
> Of course, Snappy native libraries must be installed to build and invoke 
> snappy native methods.
> (Note: there was a mistype in commit message: 8952 written in place of 8562: 
> HADOOP-8952. Enhancements to support Hadoop on Windows Server and Windows 
> Azure environments. Contributed by Ivan Mitic, Chuan Liu, Ramya Sunil, Bikas 
> Saha, Kanna Karanam, John Gordon, Brandon Li, Chris Nauroth, David Lao, 
> Sumadhur Reddy Bolli, Arpit Agarwal, Ahmed El Baz, Mike Liddell, Jing Zhao, 
> Thejas Nair, Steve Maine, Ganeshan Iyer, Raja Aluri, Giridharan Kesavan, 
> Ramya Bharathi Nimmagadda.
>git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1453486 
> 13f79535-47bb-0310-9956-ffa450edef68
> )

--
This message is au

[jira] [Commented] (HADOOP-9397) Incremental dist tar build fails

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9397?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674331#comment-13674331
 ] 

Hudson commented on HADOOP-9397:


Integrated in Hadoop-Hdfs-trunk #1420 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1420/])
Move HADOOP-9397 to 2.1.0-beta after merging it into branch-2. (Revision 
1489026)

 Result = FAILURE
jlowe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489026
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Incremental dist tar build fails
> 
>
> Key: HADOOP-9397
> URL: https://issues.apache.org/jira/browse/HADOOP-9397
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 3.0.0
>Reporter: Jason Lowe
>Assignee: Chris Nauroth
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9397.1.patch
>
>
> Building a dist tar build when the dist tarball already exists from a 
> previous build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9614) smart-test-patch.sh hangs for new version of patch (2.7.1)

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9614?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674342#comment-13674342
 ] 

Hudson commented on HADOOP-9614:


Integrated in Hadoop-Hdfs-trunk #1420 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1420/])
HADOOP-9614. smart-test-patch.sh hangs for new version of patch (2.7.1) 
(Ravi Prakash via jeagles) (Revision 1489136)

 Result = FAILURE
jeagles : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489136
Files : 
* /hadoop/common/trunk/dev-support/smart-apply-patch.sh
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> smart-test-patch.sh hangs for new version of patch (2.7.1)
> --
>
> Key: HADOOP-9614
> URL: https://issues.apache.org/jira/browse/HADOOP-9614
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Ravi Prakash
>Assignee: Ravi Prakash
> Fix For: 3.0.0, 0.23.8, 2.0.5-alpha
>
> Attachments: HADOOP-9614.patch, HADOOP-9614.patch
>
>
> patch -p0 -E --dry-run prints "checking file " for the new version of 
> patch(2.7.1) rather than "patching file" as it did for older versions. This 
> causes TMP2 to become empty, which causes the script to hang on this command 
> forever:
> PREFIX_DIRS_AND_FILES=$(cut -d '/' -f 1 | sort | uniq)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9287) Parallel testing hadoop-common

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674340#comment-13674340
 ] 

Hudson commented on HADOOP-9287:


Integrated in Hadoop-Hdfs-trunk #1420 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1420/])
Move HADOOP-9287 in CHANGES.txt after committing to branch-2 (Revision 
1489258)

 Result = FAILURE
jlowe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489258
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Parallel testing hadoop-common
> --
>
> Key: HADOOP-9287
> URL: https://issues.apache.org/jira/browse/HADOOP-9287
> Project: Hadoop Common
>  Issue Type: Test
>  Components: test
>Affects Versions: 3.0.0
>Reporter: Tsuyoshi OZAWA
>Assignee: Andrey Klochkov
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9287.1.patch, HADOOP-9287-branch-2--N1.patch, 
> HADOOP-9287--N3.patch, HADOOP-9287--N3.patch, HADOOP-9287--N4.patch, 
> HADOOP-9287--N5.patch, HADOOP-9287--N6.patch, HADOOP-9287--N7.patch, 
> HADOOP-9287.patch, HADOOP-9287.patch
>
>
> The maven surefire plugin supports parallel testing feature. By using it, the 
> tests can be run more faster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8545) Filesystem Implementation for OpenStack Swift

2013-06-04 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8545?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-8545:
---

Status: Open  (was: Patch Available)

> Filesystem Implementation for OpenStack Swift
> -
>
> Key: HADOOP-8545
> URL: https://issues.apache.org/jira/browse/HADOOP-8545
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs
>Affects Versions: 2.0.3-alpha, 1.2.0
>Reporter: Tim Miller
>Assignee: Dmitry Mezhensky
>  Labels: hadoop, patch
> Attachments: HADOOP-8545-026.patch, HADOOP-8545-027.patch, 
> HADOOP-8545-028.patch, HADOOP-8545-10.patch, HADOOP-8545-11.patch, 
> HADOOP-8545-12.patch, HADOOP-8545-13.patch, HADOOP-8545-14.patch, 
> HADOOP-8545-15.patch, HADOOP-8545-16.patch, HADOOP-8545-17.patch, 
> HADOOP-8545-18.patch, HADOOP-8545-19.patch, HADOOP-8545-1.patch, 
> HADOOP-8545-20.patch, HADOOP-8545-21.patch, HADOOP-8545-22.patch, 
> HADOOP-8545-23.patch, HADOOP-8545-24.patch, HADOOP-8545-25.patch, 
> HADOOP-8545-2.patch, HADOOP-8545-3.patch, HADOOP-8545-4.patch, 
> HADOOP-8545-5.patch, HADOOP-8545-6.patch, HADOOP-8545-7.patch, 
> HADOOP-8545-8.patch, HADOOP-8545-9.patch, HADOOP-8545-javaclouds-2.patch, 
> HADOOP-8545.patch, HADOOP-8545.patch
>
>
> ,Add a filesystem implementation for OpenStack Swift object store, similar to 
> the one which exists today for S3.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9614) smart-test-patch.sh hangs for new version of patch (2.7.1)

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9614?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674321#comment-13674321
 ] 

Hudson commented on HADOOP-9614:


Integrated in Hadoop-Hdfs-0.23-Build #628 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-0.23-Build/628/])
HADOOP-9614. smart-test-patch.sh hangs for new version of patch (2.7.1) 
(Ravi Prakash via jeagles) (Revision 1489159)

 Result = UNSTABLE
jeagles : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489159
Files : 
* /hadoop/common/branches/branch-0.23/dev-support/smart-apply-patch.sh
* 
/hadoop/common/branches/branch-0.23/hadoop-common-project/hadoop-common/CHANGES.txt


> smart-test-patch.sh hangs for new version of patch (2.7.1)
> --
>
> Key: HADOOP-9614
> URL: https://issues.apache.org/jira/browse/HADOOP-9614
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Ravi Prakash
>Assignee: Ravi Prakash
> Fix For: 3.0.0, 0.23.8, 2.0.5-alpha
>
> Attachments: HADOOP-9614.patch, HADOOP-9614.patch
>
>
> patch -p0 -E --dry-run prints "checking file " for the new version of 
> patch(2.7.1) rather than "patching file" as it did for older versions. This 
> causes TMP2 to become empty, which causes the script to hang on this command 
> forever:
> PREFIX_DIRS_AND_FILES=$(cut -d '/' -f 1 | sort | uniq)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9397) Incremental dist tar build fails

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9397?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674240#comment-13674240
 ] 

Hudson commented on HADOOP-9397:


Integrated in Hadoop-Yarn-trunk #230 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/230/])
Move HADOOP-9397 to 2.1.0-beta after merging it into branch-2. (Revision 
1489026)

 Result = FAILURE
jlowe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489026
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Incremental dist tar build fails
> 
>
> Key: HADOOP-9397
> URL: https://issues.apache.org/jira/browse/HADOOP-9397
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 3.0.0
>Reporter: Jason Lowe
>Assignee: Chris Nauroth
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9397.1.patch
>
>
> Building a dist tar build when the dist tarball already exists from a 
> previous build fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9614) smart-test-patch.sh hangs for new version of patch (2.7.1)

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9614?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674251#comment-13674251
 ] 

Hudson commented on HADOOP-9614:


Integrated in Hadoop-Yarn-trunk #230 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/230/])
HADOOP-9614. smart-test-patch.sh hangs for new version of patch (2.7.1) 
(Ravi Prakash via jeagles) (Revision 1489136)

 Result = FAILURE
jeagles : 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489136
Files : 
* /hadoop/common/trunk/dev-support/smart-apply-patch.sh
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> smart-test-patch.sh hangs for new version of patch (2.7.1)
> --
>
> Key: HADOOP-9614
> URL: https://issues.apache.org/jira/browse/HADOOP-9614
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 0.23.7, 2.0.4-alpha
>Reporter: Ravi Prakash
>Assignee: Ravi Prakash
> Fix For: 3.0.0, 0.23.8, 2.0.5-alpha
>
> Attachments: HADOOP-9614.patch, HADOOP-9614.patch
>
>
> patch -p0 -E --dry-run prints "checking file " for the new version of 
> patch(2.7.1) rather than "patching file" as it did for older versions. This 
> causes TMP2 to become empty, which causes the script to hang on this command 
> forever:
> PREFIX_DIRS_AND_FILES=$(cut -d '/' -f 1 | sort | uniq)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9481) Broken conditional logic with HADOOP_SNAPPY_LIBRARY

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674242#comment-13674242
 ] 

Hudson commented on HADOOP-9481:


Integrated in Hadoop-Yarn-trunk #230 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/230/])
HADOOP-9481. Move from trunk to release 2.1.0 section (Revision 1489261)

 Result = FAILURE
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489261
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Broken conditional logic with HADOOP_SNAPPY_LIBRARY
> ---
>
> Key: HADOOP-9481
> URL: https://issues.apache.org/jira/browse/HADOOP-9481
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.1.0-beta
>Reporter: Vadim Bondarev
>Assignee: Vadim Bondarev
>Priority: Minor
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9481-trunk--N1.patch, HADOOP-9481-trunk--N4.patch
>
>
> The problem is a regression introduced by recent fix 
> https://issues.apache.org/jira/browse/HADOOP-8562 .
> That fix makes some improvements for Windows platform, but breaks native code 
> work on Unix.
> Namely, let's see the diff HADOOP-8562 of the file 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
>  :  
> {noformat}
> --- 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
> +++ 
> hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/io/compress/snappy/SnappyCompressor.c
> @@ -16,12 +16,18 @@
>   * limitations under the License.
>   */
> -#include 
> +
> +#if defined HADOOP_SNAPPY_LIBRARY
> +
>  #include 
>  #include 
>  #include 
> +#ifdef UNIX
> +#include 
>  #include "config.h"
> +#endif // UNIX
> +
>  #include "org_apache_hadoop_io_compress_snappy.h"
>  #include "org_apache_hadoop_io_compress_snappy_SnappyCompressor.h"
> @@ -81,7 +87,7 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>UNLOCK_CLASS(env, clazz, "SnappyCompressor");
>if (uncompressed_bytes == 0) {
> -return 0;
> +return (jint)0;
>}
>// Get the output direct buffer
> @@ -90,7 +96,7 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>UNLOCK_CLASS(env, clazz, "SnappyCompressor");
>if (compressed_bytes == 0) {
> -return 0;
> +return (jint)0;
>}
>/* size_t should always be 4 bytes or larger. */
> @@ -109,3 +115,5 @@ JNIEXPORT jint JNICALL 
> Java_org_apache_hadoop_io_compress_snappy_SnappyCompresso
>(*env)->SetIntField(env, thisj, SnappyCompressor_uncompressedDirectBufLen, 
> 0);
>return (jint)buf_len;
>  }
> +
> +#endif //define HADOOP_SNAPPY_LIBRARY
> {noformat}
> Here we see that all the class implementation got enclosed into "if defined 
> HADOOP_SNAPPY_LIBRARY" directive, and the point is that 
> "HADOOP_SNAPPY_LIBRARY" is *not* defined. 
> This causes the class implementation to be effectively empty, what, in turn, 
> causes the UnsatisfiedLinkError to be thrown in the runtime upon any attempt 
> to invoke the native methods implemented there.
> The actual intention of the authors of HADOOP-8562 was (as we suppose) to 
> invoke "include config.h", where "HADOOP_SNAPPY_LIBRARY" is defined. But 
> currently it is *not* included because it resides *inside* "if defined 
> HADOOP_SNAPPY_LIBRARY" block.
> Similar situation with "ifdef UNIX", because UNIX or WINDOWS variables are 
> defined in "org_apache_hadoop.h", which is indirectly included through 
> "include "org_apache_hadoop_io_compress_snappy.h"", and in the current code 
> this is done *after* code "ifdef UNIX", so in the current code the block 
> "ifdef UNIX" is *not* executed on UNIX.
> The suggested patch fixes the described problems by reordering the "include" 
> and "if" preprocessor directives accordingly, bringing the methods of class 
> org.apache.hadoop.io.compress.snappy.SnappyCompressor back to work again.
> Of course, Snappy native libraries must be installed to build and invoke 
> snappy native methods.
> (Note: there was a mistype in commit message: 8952 written in place of 8562: 
> HADOOP-8952. Enhancements to support Hadoop on Windows Server and Windows 
> Azure environments. Contributed by Ivan Mitic, Chuan Liu, Ramya Sunil, Bikas 
> Saha, Kanna Karanam, John Gordon, Brandon Li, Chris Nauroth, David Lao, 
> Sumadhur Reddy Bolli, Arpit Agarwal, Ahmed El Baz, Mike Liddell, Jing Zhao, 
> Thejas Nair, Steve Maine, Ganeshan Iyer, Raja Aluri, Giridharan Kesavan, 
> Ramya Bharathi Nimmagadda.
>git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1453486 
> 13f79535-47bb-0310-9956-ffa450edef68
> )

--
This message is auto

[jira] [Commented] (HADOOP-9287) Parallel testing hadoop-common

2013-06-04 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674249#comment-13674249
 ] 

Hudson commented on HADOOP-9287:


Integrated in Hadoop-Yarn-trunk #230 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/230/])
Move HADOOP-9287 in CHANGES.txt after committing to branch-2 (Revision 
1489258)

 Result = FAILURE
jlowe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1489258
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


> Parallel testing hadoop-common
> --
>
> Key: HADOOP-9287
> URL: https://issues.apache.org/jira/browse/HADOOP-9287
> Project: Hadoop Common
>  Issue Type: Test
>  Components: test
>Affects Versions: 3.0.0
>Reporter: Tsuyoshi OZAWA
>Assignee: Andrey Klochkov
> Fix For: 3.0.0, 2.1.0-beta
>
> Attachments: HADOOP-9287.1.patch, HADOOP-9287-branch-2--N1.patch, 
> HADOOP-9287--N3.patch, HADOOP-9287--N3.patch, HADOOP-9287--N4.patch, 
> HADOOP-9287--N5.patch, HADOOP-9287--N6.patch, HADOOP-9287--N7.patch, 
> HADOOP-9287.patch, HADOOP-9287.patch
>
>
> The maven surefire plugin supports parallel testing feature. By using it, the 
> tests can be run more faster.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9392) Token based authentication and Single Sign On

2013-06-04 Thread Andrew Purtell (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674196#comment-13674196
 ] 

Andrew Purtell commented on HADOOP-9392:


When is the meetup?

> Token based authentication and Single Sign On
> -
>
> Key: HADOOP-9392
> URL: https://issues.apache.org/jira/browse/HADOOP-9392
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: security
>Reporter: Kai Zheng
>Assignee: Kai Zheng
> Fix For: 3.0.0
>
> Attachments: token-based-authn-plus-sso.pdf
>
>
> This is an umbrella entry for one of project Rhino’s topic, for details of 
> project Rhino, please refer to 
> https://github.com/intel-hadoop/project-rhino/. The major goal for this entry 
> as described in project Rhino was 
>  
> “Core, HDFS, ZooKeeper, and HBase currently support Kerberos authentication 
> at the RPC layer, via SASL. However this does not provide valuable attributes 
> such as group membership, classification level, organizational identity, or 
> support for user defined attributes. Hadoop components must interrogate 
> external resources for discovering these attributes and at scale this is 
> problematic. There is also no consistent delegation model. HDFS has a simple 
> delegation capability, and only Oozie can take limited advantage of it. We 
> will implement a common token based authentication framework to decouple 
> internal user and service authentication from external mechanisms used to 
> support it (like Kerberos)”
>  
> We’d like to start our work from Hadoop-Common and try to provide common 
> facilities by extending existing authentication framework which support:
> 1.Pluggable token provider interface 
> 2.Pluggable token verification protocol and interface
> 3.Security mechanism to distribute secrets in cluster nodes
> 4.Delegation model of user authentication

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9616) In branch-2, baseline of Javadoc Warnings (specified in test-patch.properties) is mismatch with Javadoc warnings in current codebase

2013-06-04 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9616?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674192#comment-13674192
 ] 

Junping Du commented on HADOOP-9616:


This patch is for branch-2 only so cannot applied on jenkins of trunk.

> In branch-2, baseline of Javadoc Warnings (specified in 
> test-patch.properties) is mismatch with  Javadoc warnings in current codebase
> -
>
> Key: HADOOP-9616
> URL: https://issues.apache.org/jira/browse/HADOOP-9616
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.0.4-alpha
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: HADOOP-9616.patch, patchJavadocWarnings-new.txt, 
> patchJavadocWarnings-old.txt
>
>
> Now the baseline is set to 13 warnings, but they are 29 warnings now. 16 
> warnings belongs to using Sun proprietary APIs, and 13 warnings is using 
> incorrect link in doc. I think we should at least fix 13 warnings and set the 
> baseline to 16.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-9616) In branch-2, baseline of Javadoc Warnings (specified in test-patch.properties) is mismatch with Javadoc warnings in current codebase

2013-06-04 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9616?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13674187#comment-13674187
 ] 

Hadoop QA commented on HADOOP-9616:
---

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12586073/patchJavadocWarnings-new.txt
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/2598//console

This message is automatically generated.

> In branch-2, baseline of Javadoc Warnings (specified in 
> test-patch.properties) is mismatch with  Javadoc warnings in current codebase
> -
>
> Key: HADOOP-9616
> URL: https://issues.apache.org/jira/browse/HADOOP-9616
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.0.4-alpha
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: HADOOP-9616.patch, patchJavadocWarnings-new.txt, 
> patchJavadocWarnings-old.txt
>
>
> Now the baseline is set to 13 warnings, but they are 29 warnings now. 16 
> warnings belongs to using Sun proprietary APIs, and 13 warnings is using 
> incorrect link in doc. I think we should at least fix 13 warnings and set the 
> baseline to 16.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9616) In branch-2, baseline of Javadoc Warnings (specified in test-patch.properties) is mismatch with Javadoc warnings in current codebase

2013-06-04 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated HADOOP-9616:
---

 Tags: javadoc
 Target Version/s: 2.1.0-beta
Affects Version/s: 2.0.4-alpha
   Status: Patch Available  (was: Open)

> In branch-2, baseline of Javadoc Warnings (specified in 
> test-patch.properties) is mismatch with  Javadoc warnings in current codebase
> -
>
> Key: HADOOP-9616
> URL: https://issues.apache.org/jira/browse/HADOOP-9616
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.0.4-alpha
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: HADOOP-9616.patch, patchJavadocWarnings-new.txt, 
> patchJavadocWarnings-old.txt
>
>
> Now the baseline is set to 13 warnings, but they are 29 warnings now. 16 
> warnings belongs to using Sun proprietary APIs, and 13 warnings is using 
> incorrect link in doc. I think we should at least fix 13 warnings and set the 
> baseline to 16.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9616) In branch-2, baseline of Javadoc Warnings (specified in test-patch.properties) is mismatch with Javadoc warnings in current codebase

2013-06-04 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated HADOOP-9616:
---

Attachment: patchJavadocWarnings-new.txt

Attach Javadoc checking output after applying this patch as *-new.txt.

> In branch-2, baseline of Javadoc Warnings (specified in 
> test-patch.properties) is mismatch with  Javadoc warnings in current codebase
> -
>
> Key: HADOOP-9616
> URL: https://issues.apache.org/jira/browse/HADOOP-9616
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: HADOOP-9616.patch, patchJavadocWarnings-new.txt, 
> patchJavadocWarnings-old.txt
>
>
> Now the baseline is set to 13 warnings, but they are 29 warnings now. 16 
> warnings belongs to using Sun proprietary APIs, and 13 warnings is using 
> incorrect link in doc. I think we should at least fix 13 warnings and set the 
> baseline to 16.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9616) In branch-2, baseline of Javadoc Warnings (specified in test-patch.properties) is mismatch with Javadoc warnings in current codebase

2013-06-04 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated HADOOP-9616:
---

Attachment: HADOOP-9616.patch

Attach a patch to fix these Javadoc issues.

> In branch-2, baseline of Javadoc Warnings (specified in 
> test-patch.properties) is mismatch with  Javadoc warnings in current codebase
> -
>
> Key: HADOOP-9616
> URL: https://issues.apache.org/jira/browse/HADOOP-9616
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: HADOOP-9616.patch, patchJavadocWarnings-old.txt
>
>
> Now the baseline is set to 13 warnings, but they are 29 warnings now. 16 
> warnings belongs to using Sun proprietary APIs, and 13 warnings is using 
> incorrect link in doc. I think we should at least fix 13 warnings and set the 
> baseline to 16.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-9616) In branch-2, baseline of Javadoc Warnings (specified in test-patch.properties) is mismatch with Javadoc warnings in current codebase

2013-06-04 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated HADOOP-9616:
---

Attachment: patchJavadocWarnings-old.txt

Attached previous Javadoc checking output as *-old.txt.

> In branch-2, baseline of Javadoc Warnings (specified in 
> test-patch.properties) is mismatch with  Javadoc warnings in current codebase
> -
>
> Key: HADOOP-9616
> URL: https://issues.apache.org/jira/browse/HADOOP-9616
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: HADOOP-9616.patch, patchJavadocWarnings-old.txt
>
>
> Now the baseline is set to 13 warnings, but they are 29 warnings now. 16 
> warnings belongs to using Sun proprietary APIs, and 13 warnings is using 
> incorrect link in doc. I think we should at least fix 13 warnings and set the 
> baseline to 16.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira