[jira] [Created] (HADOOP-11486) org.apache.hadoop.security.token.delegation.web.TestWebDelegationToken.testHttpUGI fails.

2015-01-16 Thread Spandan Dutta (JIRA)
Spandan Dutta created HADOOP-11486:
--

 Summary: 
org.apache.hadoop.security.token.delegation.web.TestWebDelegationToken.testHttpUGI
 fails.
 Key: HADOOP-11486
 URL: https://issues.apache.org/jira/browse/HADOOP-11486
 Project: Hadoop Common
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0
Reporter: Spandan Dutta


The jenkins job which was setup gave the following stack trace.

java.net.BindException: Address already in use
at java.net.PlainSocketImpl.socketBind(Native Method)
at 
java.net.AbstractPlainSocketImpl.bind(AbstractPlainSocketImpl.java:376)
at java.net.ServerSocket.bind(ServerSocket.java:376)
at java.net.ServerSocket.(ServerSocket.java:237)
at 
org.mortbay.jetty.bio.SocketConnector.newServerSocket(SocketConnector.java:80)
at org.mortbay.jetty.bio.SocketConnector.open(SocketConnector.java:73)
at 
org.mortbay.jetty.AbstractConnector.doStart(AbstractConnector.java:283)
at 
org.mortbay.jetty.bio.SocketConnector.doStart(SocketConnector.java:147)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at org.mortbay.jetty.Server.doStart(Server.java:235)
at 
org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50)
at 
org.apache.hadoop.security.token.delegation.web.TestWebDelegationToken.testHttpUGI(TestWebDelegationToken.java:934)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-10643) Add NativeS3Fs that delgates calls from FileContext apis to native s3 fs implementation

2015-01-16 Thread Sumit Kumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10643?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280942#comment-14280942
 ] 

Sumit Kumar commented on HADOOP-10643:
--

[~ste...@apache.org] I'm assuming your concern is on this portion of the patch 
(as part of AbstractFileSystem.java):
{noformat}
-// A file system implementation that requires authority must always
-// specify default port
-if (defaultPort < 0 && authorityNeeded) {
-  throw new HadoopIllegalArgumentException(
-  "FileSystem implementation error -  default port " + defaultPort
-  + " is not valid");
-}
{noformat}

If so, s3's urls have a specific requirement that they don't contain any port 
(so defaultPort becomes -1 in this case) and they don't have any authority in 
the url as well. Does this work?

> Add NativeS3Fs that delgates calls from FileContext apis to native s3 fs 
> implementation
> ---
>
> Key: HADOOP-10643
> URL: https://issues.apache.org/jira/browse/HADOOP-10643
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs/s3
>Affects Versions: 2.4.0
>Reporter: Sumit Kumar
>Assignee: Sumit Kumar
> Attachments: HADOOP-10643.patch
>
>
> The new set of file system related apis (FileContext/AbstractFileSystem) 
> already support local filesytem, hdfs, viewfs) however they don't support 
> s3n. This patch is to add that support using configurations like
> fs.AbstractFileSystem.s3n.impl = org.apache.hadoop.fs.s3native.NativeS3Fs
> This patch however doesn't provide a new implementation, instead relies on 
> DelegateToFileSystem abstract class to delegate all calls from FileContext 
> apis for s3n to the NativeS3FileSystem implementation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11364) [Java 8] Over usage of virtual memory

2015-01-16 Thread Mohammad Kamrul Islam (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11364?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280923#comment-14280923
 ] 

Mohammad Kamrul Islam commented on HADOOP-11364:


Sorry [~jira.shegalov] for the late reply. The failure was coming from distcp 
command which uses 1GB as mapreduce.map.memory.mb. I think distcp is map-only 
job.

But in other cases, we used higher memory.mb (2GB) and got the similar 
exception with max 4.2 GB VM.


> [Java 8] Over usage of virtual memory
> -
>
> Key: HADOOP-11364
> URL: https://issues.apache.org/jira/browse/HADOOP-11364
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Mohammad Kamrul Islam
>Assignee: Mohammad Kamrul Islam
>
> In our Hadoop 2 + Java8 effort , we found few jobs are being Killed by Hadoop 
> due to excessive virtual memory allocation.  Although the physical memory 
> usage is low.
> The most common error message is "Container [pid=??,containerID=container_??] 
> is running beyond virtual memory limits. Current usage: 365.1 MB of 1 GB 
> physical memory used; 3.2 GB of 2.1 GB virtual memory used. Killing 
> container."
> We see this problem for MR job as well as in spark driver/executor.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11479) hdfs crypto -createZone fails to impersonate the real user in a kerberised environment

2015-01-16 Thread Charles Lamb (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280870#comment-14280870
 ] 

Charles Lamb commented on HADOOP-11479:
---

[~ranadip],

I am pretty sure that the READ exception is being thrown when the NN is doing 
getMetadata. Here's the explanation:

There are KMS operation ACLs (hadoop.kms...) and KMS key ACLs 
(default.key.acl...). The KMS key ACLs are more coarse-grained (MANAGEMENT, 
GENERATE_EEK, DECRYPT_EEK, READ, ALL) than the KMS operation ACLs (which cover 
each public KMS method call individually).

So, by default, the HDFS user has READ permission on all keys 
(default.key.acl.READ=*). This gives that user access to the getKeyVersion, 
getKeyVersions, getMetadata, getKeysMetadata and getCurrentKey methods.

But, then, to lock down HDFS user access to key material, you also need to add 
the HDFS user to the blacklist for the following KMS operation ACLs:

CREATE, DELETE, ROLLOVER, GET, GET_KEYS, SET_KEY_MATERIAL, DECRYPT_EEK

(This is what setting the KMS Blacklist property in CM does: it is a shortcut 
to setting these seven KMS operation ACLs)
There is also a specific KMS operation ACL for GET_METADATA, but you don't want 
to set that in this case.

> hdfs crypto -createZone fails to impersonate the real user in a kerberised 
> environment
> --
>
> Key: HADOOP-11479
> URL: https://issues.apache.org/jira/browse/HADOOP-11479
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.6.0
> Environment: CentOS
>Reporter: Ranadip
> Attachments: KMS-test-acl.txt
>
>
> The problem occurs when KMS key level acl is created for the key before the 
> encryption zone is created. The command tried to create the encryption zone 
> using "hdfs" user's identity and not the real user's identity.
> Steps:
> In a kerberised environment:
> 1. Create key level ACL in KMS for a new key.
> 2. Create encryption key now. (Goes through fine)
> 3. Create encryption zone. (Fails)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11171) Enable using a proxy server to connect to S3a.

2015-01-16 Thread Thomas Demoor (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11171?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Demoor updated HADOOP-11171:
---
Attachment: HADOOP-11171-10.patch

By default failed connections are retried 10 times. Thi spatch reduces this to 
2 for the proxy tests causing the runtime to decrease from 5 mins to 8 secs. 
The huge difference in runtime is caused by the exponential backoff strategy 
used by the underlying aws-sdk.

> Enable using a proxy server to connect to S3a.
> --
>
> Key: HADOOP-11171
> URL: https://issues.apache.org/jira/browse/HADOOP-11171
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.4.0
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Attachments: HADOOP-11171-10.patch, HADOOP-11171-2.patch, 
> HADOOP-11171-3.patch, HADOOP-11171-4.patch, HADOOP-11171-5.patch, 
> HADOOP-11171-6.patch, HADOOP-11171-7.patch, HADOOP-11171-8.patch, 
> HADOOP-11171-9.patch, HADOOP-11171.patch
>
>
> This exposes the AWS SDK config for a proxy (host and port) to s3a through 
> config settings.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11460) Deprecrate shell vars

2015-01-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11460?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280825#comment-14280825
 ] 

Hadoop QA commented on HADOOP-11460:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12692800/HADOOP-11460-01.patch
  against trunk revision 000ca83.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs:

  org.apache.hadoop.ha.TestZKFailoverControllerStress
  
org.apache.hadoop.hdfs.server.blockmanagement.TestDatanodeManager

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/5418//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/5418//console

This message is automatically generated.

> Deprecrate shell vars
> -
>
> Key: HADOOP-11460
> URL: https://issues.apache.org/jira/browse/HADOOP-11460
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11460-00.patch, HADOOP-11460-01.patch
>
>
> It is a very common shell pattern in 3.x to effectively replace sub-project 
> specific vars with generics.  We should have a function that does this 
> replacement and provides a warning to the end user that the old shell var is 
> deprecated.  Additionally, we should use this shell function to deprecate the 
> shell vars that are holdovers already.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-9248) Allow configuration of Amazon S3 Endpoint

2015-01-16 Thread Jean-Martin Archer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280707#comment-14280707
 ] 

Jean-Martin Archer commented on HADOOP-9248:


As far as I am concerned, the s3a implementation makes this feature request 
unnecessary for our use cases.

> Allow configuration of Amazon S3 Endpoint
> -
>
> Key: HADOOP-9248
> URL: https://issues.apache.org/jira/browse/HADOOP-9248
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs/s3
> Environment: All environments connecting to S3
>Reporter: Timur Perelmutov
>
> http://wiki.apache.org/hadoop/AmazonS3 page describes configuration of Hadoop 
> with S3 as storage. Other systems like EMC Atmos now implement S3 Interface, 
> but in order to be able to connect to them, the endpoint needs to be 
> configurable. Please add a configuration parameter that would be propagated  
> to underlying jets3t library as s3service.s3-endpoint param.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-10181) GangliaContext does not work with multicast ganglia setup

2015-01-16 Thread Laurie Denness (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10181?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280617#comment-14280617
 ] 

Laurie Denness commented on HADOOP-10181:
-

+1 for this one.. We have 350 nodes so they span two subnets/VLANs, but we 
forward the Ganglia multicast between them. We just upgraded to CDH5 (Hadoop 
2.5) and now we've lost our Ganglia metrics for half of the cluster (because 
the TTL is set to 1) depending on which half of the cluster you ask (as they 
will not span the two subnets) 

(Given we now have a tonne of missing metrics I would say this is more than 
Minor, as the alternative is use Unicast Ganglia and that is not ideal for 
other reasons) 

> GangliaContext does not work with multicast ganglia setup
> -
>
> Key: HADOOP-10181
> URL: https://issues.apache.org/jira/browse/HADOOP-10181
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Andrew Otto
>Priority: Minor
>  Labels: ganglia, hadoop, metrics, multicast
>
> The GangliaContext class which is used to send Hadoop metrics to Ganglia uses 
> a DatagramSocket to send these metrics.  This works fine for Ganglia 
> multicast setups that are all on the same VLAN.  However, when working with 
> multiple VLANs, a packet sent via DatagramSocket to a multicast address will 
> end up with a TTL of 1.  Multicast TTL indicates the number of network hops 
> for which a particular multicast packet is valid.  The packets sent by 
> GangliaContext do not make it to ganglia aggregrators on the same multicast 
> group, but in different VLANs.
> To fix, we'd need a configuration property that specifies that multicast is 
> to be used, and another that allows setting of the multicast packet TTL.  
> With these set, we could then use MulticastSocket setTimeToLive() instead of 
> just plain ol' DatagramSocket.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11327) BloomFilter#not() omits the last bit, resulting in an incorrect filter

2015-01-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11327?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280589#comment-14280589
 ] 

Hadoop QA commented on HADOOP-11327:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12692798/HADOOP-11327.v2.txt
  against trunk revision 000ca83.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-common-project/hadoop-common.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/5419//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/5419//console

This message is automatically generated.

> BloomFilter#not() omits the last bit, resulting in an incorrect filter
> --
>
> Key: HADOOP-11327
> URL: https://issues.apache.org/jira/browse/HADOOP-11327
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 2.5.1
>Reporter: Tim Luo
>Assignee: Eric Payne
>Priority: Minor
> Attachments: HADOOP-11327.v1.txt, HADOOP-11327.v2.txt
>
>
> There's an off-by-one error in {{BloomFilter#not()}}:
> {{BloomFilter#not}} calls {{BitSet#flip(0, vectorSize - 1)}}, but according 
> to the javadoc for that method, {{toIndex}} is end-_exclusive_:
> {noformat}
> * @param  toIndex index after the last bit to flip
> {noformat}
> This means that the last bit in the bit array is not flipped.
> Specifically, this was discovered in the following scenario:
> 1. A new/empty {{BloomFilter}} was created with vectorSize=7.
> 2. Invoke {{bloomFilter.not()}}; now expecting a bloom filter with all 7 bits 
> (0 through 6) flipped to 1 and membershipTest(...) to always return true.
> 3. However, membershipTest(...) was found to often not return true, and upon 
> inspection, the BitSet only had bits 0 through 5 flipped.
> The fix should be simple: remove the "- 1" from the call to {{BitSet#flip}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-10062) race condition in MetricsSystemImpl#publishMetricsNow that causes incorrect results

2015-01-16 Thread Sangjin Lee (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10062?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sangjin Lee updated HADOOP-10062:
-
Priority: Major  (was: Minor)

> race condition in MetricsSystemImpl#publishMetricsNow that causes incorrect 
> results
> ---
>
> Key: HADOOP-10062
> URL: https://issues.apache.org/jira/browse/HADOOP-10062
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: metrics
>Affects Versions: 3.0.0
> Environment: CentOS 6.4, Oracle JDK 1.6.0_31, JDK1.7.0_45
>Reporter: Shinichi Yamashita
>Assignee: Sangjin Lee
> Attachments: HADOOP-10062-failed.txt, HADOOP-10062-success.txt, 
> HADOOP-10062.003.patch, HADOOP-10062.patch, HADOOP-10062.patch
>
>
> TestMetricsSystemInpl#testMultiThreadedPublish failed with "Metrics not 
> collected"
> {code}
> Running org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl
> Tests run: 6, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 1.688 sec <<< 
> FAILURE! - in org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl
> testMultiThreadedPublish(org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl)
>   Time elapsed: 0.056 sec  <<< FAILURE!
> java.lang.AssertionError: Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Passed
> at org.junit.Assert.fail(Assert.java:93)
> at org.junit.Assert.assertTrue(Assert.java:43)
> at 
> org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl.testMultiThreadedPublish(TestMetricsSystemImpl.java:232)
> Results :
> Failed tests:
>   TestMetricsSystemImpl.testMultiThreadedPublish:232 Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Metric not collected!
> Passed
> Tests run: 6, Failures: 1, Errors: 0, Skipped: 0
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11460) Deprecrate shell vars

2015-01-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-11460:
--
Status: Patch Available  (was: Open)

> Deprecrate shell vars
> -
>
> Key: HADOOP-11460
> URL: https://issues.apache.org/jira/browse/HADOOP-11460
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11460-00.patch, HADOOP-11460-01.patch
>
>
> It is a very common shell pattern in 3.x to effectively replace sub-project 
> specific vars with generics.  We should have a function that does this 
> replacement and provides a warning to the end user that the old shell var is 
> deprecated.  Additionally, we should use this shell function to deprecate the 
> shell vars that are holdovers already.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11460) Deprecrate shell vars

2015-01-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-11460:
--
Attachment: HADOOP-11460-01.patch

-01:
* Eliminate most (all?) of the usages of the deprecated env vars!

> Deprecrate shell vars
> -
>
> Key: HADOOP-11460
> URL: https://issues.apache.org/jira/browse/HADOOP-11460
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11460-00.patch, HADOOP-11460-01.patch
>
>
> It is a very common shell pattern in 3.x to effectively replace sub-project 
> specific vars with generics.  We should have a function that does this 
> replacement and provides a warning to the end user that the old shell var is 
> deprecated.  Additionally, we should use this shell function to deprecate the 
> shell vars that are holdovers already.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11460) Deprecrate shell vars

2015-01-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-11460:
--
Description: 
It is a very common shell pattern in 3.x to effectively replace sub-project 
specific vars with generics.  We should have a function that does this 
replacement and provides a warning to the end user that the old shell var is 
deprecated.  Additionally, we should use this shell function to deprecate the 
shell vars that are holdovers already.


  was:It is a very common shell pattern in 3.x to effectively replace 
sub-project specific vars with generics.  We should have a function that does 
this replacement and provides a warning to the end user that the old shell var 
is deprecated.  This will allow us an easier time to remove them in the future.


> Deprecrate shell vars
> -
>
> Key: HADOOP-11460
> URL: https://issues.apache.org/jira/browse/HADOOP-11460
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11460-00.patch
>
>
> It is a very common shell pattern in 3.x to effectively replace sub-project 
> specific vars with generics.  We should have a function that does this 
> replacement and provides a warning to the end user that the old shell var is 
> deprecated.  Additionally, we should use this shell function to deprecate the 
> shell vars that are holdovers already.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11171) Enable using a proxy server to connect to S3a.

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280510#comment-14280510
 ] 

Steve Loughran commented on HADOOP-11171:
-

well, the tests pass, but they take ~5 minutes
{code}
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 295.421 sec - 
in org.apache.hadoop.fs.s3a.TestS3AConfiguration

Results :

Tests run: 5, Failures: 0, Errors: 0, Skipped: 0

[INFO] 
[INFO] BUILD SUCCESS
[INFO] 
[INFO] Total time: 4:57.925s
{code}

Cause? Connection retries:
{code}
   java.lang.Thread.State: TIMED_WAITING (sleeping)
at java.lang.Thread.sleep(Native Method)
at 
com.amazonaws.http.AmazonHttpClient.pauseBeforeNextRetry(AmazonHttpClient.java:864)
at 
com.amazonaws.http.AmazonHttpClient.executeHelper(AmazonHttpClient.java:353)
at 
com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:232)
at 
com.amazonaws.services.s3.AmazonS3Client.invoke(AmazonS3Client.java:3528)
at 
com.amazonaws.services.s3.AmazonS3Client.headBucket(AmazonS3Client.java:1031)
at 
com.amazonaws.services.s3.AmazonS3Client.doesBucketExist(AmazonS3Client.java:994)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:282)
at 
org.apache.hadoop.fs.s3a.S3ATestUtils.createTestFileSystem(S3ATestUtils.java:48)
at 
org.apache.hadoop.fs.s3a.TestS3AConfiguration.TestAutomaticProxyPortSelection(TestS3AConfiguration.java:126)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74)
{code}

Is there a way to tune the failure policy of the AWS lib so it doesn't retry on 
these test runs?

> Enable using a proxy server to connect to S3a.
> --
>
> Key: HADOOP-11171
> URL: https://issues.apache.org/jira/browse/HADOOP-11171
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.4.0
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Attachments: HADOOP-11171-2.patch, HADOOP-11171-3.patch, 
> HADOOP-11171-4.patch, HADOOP-11171-5.patch, HADOOP-11171-6.patch, 
> HADOOP-11171-7.patch, HADOOP-11171-8.patch, HADOOP-11171-9.patch, 
> HADOOP-11171.patch
>
>
> This exposes the AWS SDK config for a proxy (host and port) to s3a through 
> config settings.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11327) BloomFilter#not() omits the last bit, resulting in an incorrect filter

2015-01-16 Thread Eric Payne (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Payne updated HADOOP-11327:

Attachment: HADOOP-11327.v2.txt

Thanks, [~jlowe], for the review and comments.

I have updated the test case in version 2 of the patch.

> BloomFilter#not() omits the last bit, resulting in an incorrect filter
> --
>
> Key: HADOOP-11327
> URL: https://issues.apache.org/jira/browse/HADOOP-11327
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: util
>Affects Versions: 2.5.1
>Reporter: Tim Luo
>Assignee: Eric Payne
>Priority: Minor
> Attachments: HADOOP-11327.v1.txt, HADOOP-11327.v2.txt
>
>
> There's an off-by-one error in {{BloomFilter#not()}}:
> {{BloomFilter#not}} calls {{BitSet#flip(0, vectorSize - 1)}}, but according 
> to the javadoc for that method, {{toIndex}} is end-_exclusive_:
> {noformat}
> * @param  toIndex index after the last bit to flip
> {noformat}
> This means that the last bit in the bit array is not flipped.
> Specifically, this was discovered in the following scenario:
> 1. A new/empty {{BloomFilter}} was created with vectorSize=7.
> 2. Invoke {{bloomFilter.not()}}; now expecting a bloom filter with all 7 bits 
> (0 through 6) flipped to 1 and membershipTest(...) to always return true.
> 3. However, membershipTest(...) was found to often not return true, and upon 
> inspection, the BitSet only had bits 0 through 5 flipped.
> The fix should be simple: remove the "- 1" from the call to {{BitSet#flip}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11460) Deprecrate shell vars

2015-01-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-11460:
--
Summary: Deprecrate shell vars  (was: Add a way to deprecrate shell vars)

> Deprecrate shell vars
> -
>
> Key: HADOOP-11460
> URL: https://issues.apache.org/jira/browse/HADOOP-11460
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11460-00.patch
>
>
> It is a very common shell pattern in 3.x to effectively replace sub-project 
> specific vars with generics.  We should have a function that does this 
> replacement and provides a warning to the end user that the old shell var is 
> deprecated.  This will allow us an easier time to remove them in the future.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-10542) Potential null pointer dereference in Jets3tFileSystemStore#retrieveBlock()

2015-01-16 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280446#comment-14280446
 ] 

Ted Yu commented on HADOOP-10542:
-

If in is null, do we return an empty fileBlock or raise exception ?
I lean toward the latter. e.g. raise IOE with message saying object with key 
blockToKey(block) wasn't found.

> Potential null pointer dereference in Jets3tFileSystemStore#retrieveBlock()
> ---
>
> Key: HADOOP-10542
> URL: https://issues.apache.org/jira/browse/HADOOP-10542
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.6.0
>Reporter: Ted Yu
>Priority: Minor
>
> {code}
>   in = get(blockToKey(block), byteRangeStart);
>   out = new BufferedOutputStream(new FileOutputStream(fileBlock));
>   byte[] buf = new byte[bufferSize];
>   int numRead;
>   while ((numRead = in.read(buf)) >= 0) {
> {code}
> get() may return null.
> The while loop dereferences in without null check.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-10533) S3 input stream NPEs in MapReduce job

2015-01-16 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10533?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HADOOP-10533:

Summary: S3 input stream NPEs in MapReduce job  (was: S3 input stream NPEs 
in MapReduce jon)

> S3 input stream NPEs in MapReduce job
> -
>
> Key: HADOOP-10533
> URL: https://issues.apache.org/jira/browse/HADOOP-10533
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 1.0.0, 1.0.3, 3.0.0, 2.4.0
> Environment: Hadoop with default configurations
>Reporter: Benjamin Kim
>Assignee: Steve Loughran
>Priority: Minor
> Fix For: 2.5.0
>
>
> I'm running a wordcount MR as follows
> hadoop jar WordCount.jar wordcount.WordCountDriver 
> s3n://bucket/wordcount/input s3n://bucket/wordcount/output
>  
> s3n://bucket/wordcount/input is a s3 object that contains other input files.
> However I get following NPE error
> 12/10/02 18:56:23 INFO mapred.JobClient:  map 0% reduce 0%
> 12/10/02 18:56:54 INFO mapred.JobClient:  map 50% reduce 0%
> 12/10/02 18:56:56 INFO mapred.JobClient: Task Id : 
> attempt_201210021853_0001_m_01_0, Status : FAILED
> java.lang.NullPointerException
> at 
> org.apache.hadoop.fs.s3native.NativeS3FileSystem$NativeS3FsInputStream.close(NativeS3FileSystem.java:106)
> at java.io.BufferedInputStream.close(BufferedInputStream.java:451)
> at java.io.FilterInputStream.close(FilterInputStream.java:155)
> at org.apache.hadoop.util.LineReader.close(LineReader.java:83)
> at 
> org.apache.hadoop.mapreduce.lib.input.LineRecordReader.close(LineRecordReader.java:144)
> at 
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.close(MapTask.java:497)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:765)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
> at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:396)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
> at org.apache.hadoop.mapred.Child.main(Child.java:249)
> MR runs fine if i specify more specific input path such as 
> s3n://bucket/wordcount/input/file.txt
> MR fails if I pass s3 folder as a parameter
> In summary,
> This works
>  hadoop jar ./hadoop-examples-1.0.3.jar wordcount 
> /user/hadoop/wordcount/input/ s3n://bucket/wordcount/output/
> This doesn't work
>  hadoop jar ./hadoop-examples-1.0.3.jar wordcount 
> s3n://bucket/wordcount/input/ s3n://bucket/wordcount/output/
> (both input path are directories)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11171) Enable using a proxy server to connect to S3a.

2015-01-16 Thread Thomas Demoor (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11171?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Demoor updated HADOOP-11171:
---
Attachment: HADOOP-11171-9.patch

Rebased on trunk and reran the tests.

Didn't apply as this shares the testclass with my endpoint feature that you 
merged in earlier today.

> Enable using a proxy server to connect to S3a.
> --
>
> Key: HADOOP-11171
> URL: https://issues.apache.org/jira/browse/HADOOP-11171
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.4.0
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Attachments: HADOOP-11171-2.patch, HADOOP-11171-3.patch, 
> HADOOP-11171-4.patch, HADOOP-11171-5.patch, HADOOP-11171-6.patch, 
> HADOOP-11171-7.patch, HADOOP-11171-8.patch, HADOOP-11171-9.patch, 
> HADOOP-11171.patch
>
>
> This exposes the AWS SDK config for a proxy (host and port) to s3a through 
> config settings.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11463) Replace method-local TransferManager object with S3AFileSystem#transfers

2015-01-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280421#comment-14280421
 ] 

Hadoop QA commented on HADOOP-11463:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12692782/hadoop-11463-002.patch
  against trunk revision 000ca83.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-tools/hadoop-aws.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/5417//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/5417//console

This message is automatically generated.

> Replace method-local TransferManager object with S3AFileSystem#transfers
> 
>
> Key: HADOOP-11463
> URL: https://issues.apache.org/jira/browse/HADOOP-11463
> Project: Hadoop Common
>  Issue Type: Task
>  Components: fs/s3
>Affects Versions: 2.7.0
>Reporter: Ted Yu
>Assignee: Ted Yu
> Attachments: hadoop-11463-001.patch, hadoop-11463-002.patch
>
>
> This is continuation of HADOOP-11446.
> The following changes are made according to Thomas Demoor's comments:
> 1. Replace method-local TransferManager object with S3AFileSystem#transfers
> 2. Do not shutdown TransferManager after purging existing multipart file - 
> otherwise the current transfer is unable to proceed
> 3. Shutdown TransferManager instance in the close method of S3AFileSystem



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11350) The size of header buffer of HttpServer is too small when HTTPS is enabled

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280413#comment-14280413
 ] 

Hudson commented on HADOOP-11350:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #2026 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2026/])
HADOOP-11350. The size of header buffer of HttpServer is too small when HTTPS 
is enabled. Contributed by Benoy Antony. (wheat9: rev 
3ab3a6498812c9fa0c53dae02ce696033062af87)
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestSSLHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/HttpServerFunctionalTest.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java


> The size of header buffer of HttpServer is too small when HTTPS is enabled
> --
>
> Key: HADOOP-11350
> URL: https://issues.apache.org/jira/browse/HADOOP-11350
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HADOOP-11350.patch, HADOOP-11350.patch, 
> HADOOP-11350.patch, HADOOP-11350.patch
>
>
> {code}
> curl -k  -vvv -i -L --negotiate -u : https://:50070
> < HTTP/1.1 413 FULL head
> HTTP/1.1 413 FULL head
> < Connection: close
> Connection: close
> < Server: Jetty(6.1.26)
> Server: Jetty(6.1.26)
> {code}
> For some users, the spnego token too large for the default header buffer used 
> by Jetty. 
> Though the issue is fixed for HTTP connections (via HADOOP-8816), HTTPS 
> connections needs to be fixed as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280410#comment-14280410
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #2026 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2026/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11318) Update the document for hadoop fs -stat

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280411#comment-14280411
 ] 

Hudson commented on HADOOP-11318:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #2026 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2026/])
HADOOP-11318. Update the document for hadoop fs -stat (aajisaka: rev 
ce29074685abaa88ae9586fe8277aca23f9d54d6)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Stat.java
* hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm
* hadoop-common-project/hadoop-common/CHANGES.txt


> Update the document for hadoop fs -stat
> ---
>
> Key: HADOOP-11318
> URL: https://issues.apache.org/jira/browse/HADOOP-11318
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.5.1
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-11318-001.patch, HADOOP-11318-002.patch, 
> HADOOP-11318-002.patch, HADOOP-11318-003.patch
>
>
> In FileSystemShell.apt.vm, 
> {code}
> stat
>Usage: <<>>
>Returns the stat information on the path.
> {code}
> Now {{-stat}} accepts the below formats.
>  *   %b: Size of file in blocks
>  *   %g: Group name of owner
>  *   %n: Filename
>  *   %o: Block size
>  *   %r: replication
>  *   %u: User name of owner
>  *   %y: UTC date as "-MM-dd HH:mm:ss"
>  *   %Y: Milliseconds since January 1, 1970 UTC
> They should be documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8757) Metrics should disallow names with invalid characters

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280418#comment-14280418
 ] 

Hudson commented on HADOOP-8757:


FAILURE: Integrated in Hadoop-Mapreduce-trunk #2026 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2026/])
HADOOP-8757. Metrics should disallow names with invalid characters (rchiang via 
rkanter) (rkanter: rev b6ff9c03a4f8aba945e562a7ff60b0fc6a1cd040)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMetricsRegistry.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java


> Metrics should disallow names with invalid characters
> -
>
> Key: HADOOP-8757
> URL: https://issues.apache.org/jira/browse/HADOOP-8757
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Todd Lipcon
>Assignee: Ray Chiang
>Priority: Minor
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-8757-01.patch, HADOOP-8757-02.patch, 
> HADOOP-8757.003.patch
>
>
> Just spent a couple hours trying to figure out why a metric I added didn't 
> show up in JMX, only to eventually realize it was because I had a whitespace 
> in the property name. This didn't cause any errors to be logged -- the metric 
> just didn't show up in JMX. We should check that the name is valid and log an 
> error, or replace invalid characters with something like an underscore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11479) hdfs crypto -createZone fails to impersonate the real user in a kerberised environment

2015-01-16 Thread Ranadip (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280359#comment-14280359
 ] 

Ranadip commented on HADOOP-11479:
--

Hi Charles,

Yes, you are right on the user setup and namenode process is started by "hdfs". 
This is a kerberos enabled environment, so the real user is obtained from the 
Kerberos ticket. The Kerberos user, though not "hdfs" is a member of the hdfs 
supergroup. So, Hadoop should see the user as a superuser when executing this 
command. Moreover, the encryption zone is being created inside the Kerberos 
user's home directory where the user executing the command is in fact the owner 
of the EZ being created.
However, providing key level access, specially READ access to an application 
user does ring some alarm bells. In effect, with this limitation that hdfs user 
needs those key accesses for all encryption zones, it means if the hdfs user 
credentials are compromised, the "baddie" can have access to the encrypted data 
as well (or am I missing something?) Should the process rather not use the real 
user (authenticated by Kerberos) from the UserGroupInformation  check key level 
authorisations with KMS?

> hdfs crypto -createZone fails to impersonate the real user in a kerberised 
> environment
> --
>
> Key: HADOOP-11479
> URL: https://issues.apache.org/jira/browse/HADOOP-11479
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.6.0
> Environment: CentOS
>Reporter: Ranadip
> Attachments: KMS-test-acl.txt
>
>
> The problem occurs when KMS key level acl is created for the key before the 
> encryption zone is created. The command tried to create the encryption zone 
> using "hdfs" user's identity and not the real user's identity.
> Steps:
> In a kerberised environment:
> 1. Create key level ACL in KMS for a new key.
> 2. Create encryption key now. (Goes through fine)
> 3. Create encryption zone. (Fails)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8757) Metrics should disallow names with invalid characters

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280357#comment-14280357
 ] 

Hudson commented on HADOOP-8757:


FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #76 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/76/])
HADOOP-8757. Metrics should disallow names with invalid characters (rchiang via 
rkanter) (rkanter: rev b6ff9c03a4f8aba945e562a7ff60b0fc6a1cd040)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMetricsRegistry.java


> Metrics should disallow names with invalid characters
> -
>
> Key: HADOOP-8757
> URL: https://issues.apache.org/jira/browse/HADOOP-8757
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Todd Lipcon
>Assignee: Ray Chiang
>Priority: Minor
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-8757-01.patch, HADOOP-8757-02.patch, 
> HADOOP-8757.003.patch
>
>
> Just spent a couple hours trying to figure out why a metric I added didn't 
> show up in JMX, only to eventually realize it was because I had a whitespace 
> in the property name. This didn't cause any errors to be logged -- the metric 
> just didn't show up in JMX. We should check that the name is valid and log an 
> error, or replace invalid characters with something like an underscore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11318) Update the document for hadoop fs -stat

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280350#comment-14280350
 ] 

Hudson commented on HADOOP-11318:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #76 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/76/])
HADOOP-11318. Update the document for hadoop fs -stat (aajisaka: rev 
ce29074685abaa88ae9586fe8277aca23f9d54d6)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Stat.java
* hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm


> Update the document for hadoop fs -stat
> ---
>
> Key: HADOOP-11318
> URL: https://issues.apache.org/jira/browse/HADOOP-11318
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.5.1
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-11318-001.patch, HADOOP-11318-002.patch, 
> HADOOP-11318-002.patch, HADOOP-11318-003.patch
>
>
> In FileSystemShell.apt.vm, 
> {code}
> stat
>Usage: <<>>
>Returns the stat information on the path.
> {code}
> Now {{-stat}} accepts the below formats.
>  *   %b: Size of file in blocks
>  *   %g: Group name of owner
>  *   %n: Filename
>  *   %o: Block size
>  *   %r: replication
>  *   %u: User name of owner
>  *   %y: UTC date as "-MM-dd HH:mm:ss"
>  *   %Y: Milliseconds since January 1, 1970 UTC
> They should be documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280349#comment-14280349
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #76 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/76/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11350) The size of header buffer of HttpServer is too small when HTTPS is enabled

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280352#comment-14280352
 ] 

Hudson commented on HADOOP-11350:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #76 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/76/])
HADOOP-11350. The size of header buffer of HttpServer is too small when HTTPS 
is enabled. Contributed by Benoy Antony. (wheat9: rev 
3ab3a6498812c9fa0c53dae02ce696033062af87)
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/HttpServerFunctionalTest.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestSSLHttpServer.java


> The size of header buffer of HttpServer is too small when HTTPS is enabled
> --
>
> Key: HADOOP-11350
> URL: https://issues.apache.org/jira/browse/HADOOP-11350
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HADOOP-11350.patch, HADOOP-11350.patch, 
> HADOOP-11350.patch, HADOOP-11350.patch
>
>
> {code}
> curl -k  -vvv -i -L --negotiate -u : https://:50070
> < HTTP/1.1 413 FULL head
> HTTP/1.1 413 FULL head
> < Connection: close
> Connection: close
> < Server: Jetty(6.1.26)
> Server: Jetty(6.1.26)
> {code}
> For some users, the spnego token too large for the default header buffer used 
> by Jetty. 
> Though the issue is fixed for HTTP connections (via HADOOP-8816), HTTPS 
> connections needs to be fixed as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11463) Replace method-local TransferManager object with S3AFileSystem#transfers

2015-01-16 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HADOOP-11463:

Attachment: hadoop-11463-002.patch

> Replace method-local TransferManager object with S3AFileSystem#transfers
> 
>
> Key: HADOOP-11463
> URL: https://issues.apache.org/jira/browse/HADOOP-11463
> Project: Hadoop Common
>  Issue Type: Task
>  Components: fs/s3
>Affects Versions: 2.7.0
>Reporter: Ted Yu
>Assignee: Ted Yu
> Attachments: hadoop-11463-001.patch, hadoop-11463-002.patch
>
>
> This is continuation of HADOOP-11446.
> The following changes are made according to Thomas Demoor's comments:
> 1. Replace method-local TransferManager object with S3AFileSystem#transfers
> 2. Do not shutdown TransferManager after purging existing multipart file - 
> otherwise the current transfer is unable to proceed
> 3. Shutdown TransferManager instance in the close method of S3AFileSystem



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8757) Metrics should disallow names with invalid characters

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280312#comment-14280312
 ] 

Hudson commented on HADOOP-8757:


FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #72 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/72/])
HADOOP-8757. Metrics should disallow names with invalid characters (rchiang via 
rkanter) (rkanter: rev b6ff9c03a4f8aba945e562a7ff60b0fc6a1cd040)
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMetricsRegistry.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java


> Metrics should disallow names with invalid characters
> -
>
> Key: HADOOP-8757
> URL: https://issues.apache.org/jira/browse/HADOOP-8757
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Todd Lipcon
>Assignee: Ray Chiang
>Priority: Minor
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-8757-01.patch, HADOOP-8757-02.patch, 
> HADOOP-8757.003.patch
>
>
> Just spent a couple hours trying to figure out why a metric I added didn't 
> show up in JMX, only to eventually realize it was because I had a whitespace 
> in the property name. This didn't cause any errors to be logged -- the metric 
> just didn't show up in JMX. We should check that the name is valid and log an 
> error, or replace invalid characters with something like an underscore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11318) Update the document for hadoop fs -stat

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280304#comment-14280304
 ] 

Hudson commented on HADOOP-11318:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #72 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/72/])
HADOOP-11318. Update the document for hadoop fs -stat (aajisaka: rev 
ce29074685abaa88ae9586fe8277aca23f9d54d6)
* hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Stat.java
* hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* hadoop-common-project/hadoop-common/CHANGES.txt


> Update the document for hadoop fs -stat
> ---
>
> Key: HADOOP-11318
> URL: https://issues.apache.org/jira/browse/HADOOP-11318
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.5.1
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-11318-001.patch, HADOOP-11318-002.patch, 
> HADOOP-11318-002.patch, HADOOP-11318-003.patch
>
>
> In FileSystemShell.apt.vm, 
> {code}
> stat
>Usage: <<>>
>Returns the stat information on the path.
> {code}
> Now {{-stat}} accepts the below formats.
>  *   %b: Size of file in blocks
>  *   %g: Group name of owner
>  *   %n: Filename
>  *   %o: Block size
>  *   %r: replication
>  *   %u: User name of owner
>  *   %y: UTC date as "-MM-dd HH:mm:ss"
>  *   %Y: Milliseconds since January 1, 1970 UTC
> They should be documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11483) HardLink.java should use the jdk7 createLink method

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280309#comment-14280309
 ] 

Hudson commented on HADOOP-11483:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #72 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/72/])
HADOOP-11483. HardLink.java should use the jdk7 createLink method (aajisaka: 
rev db51548f706ccd2d0200745ab89e27610c6d10bc)
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHardLink.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HardLink.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> HardLink.java should use the jdk7 createLink method
> ---
>
> Key: HADOOP-11483
> URL: https://issues.apache.org/jira/browse/HADOOP-11483
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.7.0
>Reporter: Colin Patrick McCabe
>Assignee: Akira AJISAKA
> Fix For: 2.7.0
>
> Attachments: HDFS-7556-001.patch
>
>
> Now that we are using jdk7, HardLink.java should use the jdk7 createLink 
> method rather than our shell commands or JNI methods.
> Note that we cannot remove all of the JNI / shell commands unless we remove 
> the code which is checking the link count, something that jdk7 doesn't 
> provide (at least, I don't think it does)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11350) The size of header buffer of HttpServer is too small when HTTPS is enabled

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280306#comment-14280306
 ] 

Hudson commented on HADOOP-11350:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #72 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/72/])
HADOOP-11350. The size of header buffer of HttpServer is too small when HTTPS 
is enabled. Contributed by Benoy Antony. (wheat9: rev 
3ab3a6498812c9fa0c53dae02ce696033062af87)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpServer.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestSSLHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/HttpServerFunctionalTest.java


> The size of header buffer of HttpServer is too small when HTTPS is enabled
> --
>
> Key: HADOOP-11350
> URL: https://issues.apache.org/jira/browse/HADOOP-11350
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HADOOP-11350.patch, HADOOP-11350.patch, 
> HADOOP-11350.patch, HADOOP-11350.patch
>
>
> {code}
> curl -k  -vvv -i -L --negotiate -u : https://:50070
> < HTTP/1.1 413 FULL head
> HTTP/1.1 413 FULL head
> < Connection: close
> Connection: close
> < Server: Jetty(6.1.26)
> Server: Jetty(6.1.26)
> {code}
> For some users, the spnego token too large for the default header buffer used 
> by Jetty. 
> Though the issue is fixed for HTTP connections (via HADOOP-8816), HTTPS 
> connections needs to be fixed as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280303#comment-14280303
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #72 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/72/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11350) The size of header buffer of HttpServer is too small when HTTPS is enabled

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280285#comment-14280285
 ] 

Hudson commented on HADOOP-11350:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #2007 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2007/])
HADOOP-11350. The size of header buffer of HttpServer is too small when HTTPS 
is enabled. Contributed by Benoy Antony. (wheat9: rev 
3ab3a6498812c9fa0c53dae02ce696033062af87)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestSSLHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/HttpServerFunctionalTest.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpServer.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> The size of header buffer of HttpServer is too small when HTTPS is enabled
> --
>
> Key: HADOOP-11350
> URL: https://issues.apache.org/jira/browse/HADOOP-11350
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HADOOP-11350.patch, HADOOP-11350.patch, 
> HADOOP-11350.patch, HADOOP-11350.patch
>
>
> {code}
> curl -k  -vvv -i -L --negotiate -u : https://:50070
> < HTTP/1.1 413 FULL head
> HTTP/1.1 413 FULL head
> < Connection: close
> Connection: close
> < Server: Jetty(6.1.26)
> Server: Jetty(6.1.26)
> {code}
> For some users, the spnego token too large for the default header buffer used 
> by Jetty. 
> Though the issue is fixed for HTTP connections (via HADOOP-8816), HTTPS 
> connections needs to be fixed as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8757) Metrics should disallow names with invalid characters

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280291#comment-14280291
 ] 

Hudson commented on HADOOP-8757:


FAILURE: Integrated in Hadoop-Hdfs-trunk #2007 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2007/])
HADOOP-8757. Metrics should disallow names with invalid characters (rchiang via 
rkanter) (rkanter: rev b6ff9c03a4f8aba945e562a7ff60b0fc6a1cd040)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMetricsRegistry.java


> Metrics should disallow names with invalid characters
> -
>
> Key: HADOOP-8757
> URL: https://issues.apache.org/jira/browse/HADOOP-8757
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Todd Lipcon
>Assignee: Ray Chiang
>Priority: Minor
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-8757-01.patch, HADOOP-8757-02.patch, 
> HADOOP-8757.003.patch
>
>
> Just spent a couple hours trying to figure out why a metric I added didn't 
> show up in JMX, only to eventually realize it was because I had a whitespace 
> in the property name. This didn't cause any errors to be logged -- the metric 
> just didn't show up in JMX. We should check that the name is valid and log an 
> error, or replace invalid characters with something like an underscore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280282#comment-14280282
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #2007 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2007/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11318) Update the document for hadoop fs -stat

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280283#comment-14280283
 ] 

Hudson commented on HADOOP-11318:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #2007 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2007/])
HADOOP-11318. Update the document for hadoop fs -stat (aajisaka: rev 
ce29074685abaa88ae9586fe8277aca23f9d54d6)
* hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* hadoop-common-project/hadoop-common/CHANGES.txt
* hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Stat.java


> Update the document for hadoop fs -stat
> ---
>
> Key: HADOOP-11318
> URL: https://issues.apache.org/jira/browse/HADOOP-11318
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.5.1
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-11318-001.patch, HADOOP-11318-002.patch, 
> HADOOP-11318-002.patch, HADOOP-11318-003.patch
>
>
> In FileSystemShell.apt.vm, 
> {code}
> stat
>Usage: <<>>
>Returns the stat information on the path.
> {code}
> Now {{-stat}} accepts the below formats.
>  *   %b: Size of file in blocks
>  *   %g: Group name of owner
>  *   %n: Filename
>  *   %o: Block size
>  *   %r: replication
>  *   %u: User name of owner
>  *   %y: UTC date as "-MM-dd HH:mm:ss"
>  *   %Y: Milliseconds since January 1, 1970 UTC
> They should be documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11483) HardLink.java should use the jdk7 createLink method

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280288#comment-14280288
 ] 

Hudson commented on HADOOP-11483:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #2007 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2007/])
HADOOP-11483. HardLink.java should use the jdk7 createLink method (aajisaka: 
rev db51548f706ccd2d0200745ab89e27610c6d10bc)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HardLink.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHardLink.java


> HardLink.java should use the jdk7 createLink method
> ---
>
> Key: HADOOP-11483
> URL: https://issues.apache.org/jira/browse/HADOOP-11483
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.7.0
>Reporter: Colin Patrick McCabe
>Assignee: Akira AJISAKA
> Fix For: 2.7.0
>
> Attachments: HDFS-7556-001.patch
>
>
> Now that we are using jdk7, HardLink.java should use the jdk7 createLink 
> method rather than our shell commands or JNI methods.
> Note that we cannot remove all of the JNI / shell commands unless we remove 
> the code which is checking the link count, something that jdk7 doesn't 
> provide (at least, I don't think it does)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11463) Replace method-local TransferManager object with S3AFileSystem#transfers

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-11463:

  Component/s: fs/s3
 Target Version/s: 2.7.0
Affects Version/s: 2.7.0

> Replace method-local TransferManager object with S3AFileSystem#transfers
> 
>
> Key: HADOOP-11463
> URL: https://issues.apache.org/jira/browse/HADOOP-11463
> Project: Hadoop Common
>  Issue Type: Task
>  Components: fs/s3
>Affects Versions: 2.7.0
>Reporter: Ted Yu
>Assignee: Ted Yu
> Attachments: hadoop-11463-001.patch
>
>
> This is continuation of HADOOP-11446.
> The following changes are made according to Thomas Demoor's comments:
> 1. Replace method-local TransferManager object with S3AFileSystem#transfers
> 2. Do not shutdown TransferManager after purging existing multipart file - 
> otherwise the current transfer is unable to proceed
> 3. Shutdown TransferManager instance in the close method of S3AFileSystem



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11463) Replace method-local TransferManager object with S3AFileSystem#transfers

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11463?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280258#comment-14280258
 ] 

Steve Loughran commented on HADOOP-11463:
-

-1

I want one minor change, which is the {close()}} operation must check for 
transfers != null before doing its work. There's some stuff done in the 
{{initialize()}} method before the field is assigned, so there's a risk that if 
something went wrong there, close() would fail.

Once that's done I'll approve the patch

> Replace method-local TransferManager object with S3AFileSystem#transfers
> 
>
> Key: HADOOP-11463
> URL: https://issues.apache.org/jira/browse/HADOOP-11463
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Ted Yu
>Assignee: Ted Yu
> Attachments: hadoop-11463-001.patch
>
>
> This is continuation of HADOOP-11446.
> The following changes are made according to Thomas Demoor's comments:
> 1. Replace method-local TransferManager object with S3AFileSystem#transfers
> 2. Do not shutdown TransferManager after purging existing multipart file - 
> otherwise the current transfer is unable to proceed
> 3. Shutdown TransferManager instance in the close method of S3AFileSystem



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-10542) Potential null pointer dereference in Jets3tFileSystemStore#retrieveBlock()

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280247#comment-14280247
 ] 

Steve Loughran commented on HADOOP-10542:
-

this still exists in 2.6. Ted: do you have a patch for this?

> Potential null pointer dereference in Jets3tFileSystemStore#retrieveBlock()
> ---
>
> Key: HADOOP-10542
> URL: https://issues.apache.org/jira/browse/HADOOP-10542
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.6.0
>Reporter: Ted Yu
>Priority: Minor
>
> {code}
>   in = get(blockToKey(block), byteRangeStart);
>   out = new BufferedOutputStream(new FileOutputStream(fileBlock));
>   byte[] buf = new byte[bufferSize];
>   int numRead;
>   while ((numRead = in.read(buf)) >= 0) {
> {code}
> get() may return null.
> The while loop dereferences in without null check.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-10542) Potential null pointer dereference in Jets3tFileSystemStore#retrieveBlock()

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-10542:

 Target Version/s: 2.7.0
Affects Version/s: 2.6.0

> Potential null pointer dereference in Jets3tFileSystemStore#retrieveBlock()
> ---
>
> Key: HADOOP-10542
> URL: https://issues.apache.org/jira/browse/HADOOP-10542
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.6.0
>Reporter: Ted Yu
>Priority: Minor
>
> {code}
>   in = get(blockToKey(block), byteRangeStart);
>   out = new BufferedOutputStream(new FileOutputStream(fileBlock));
>   byte[] buf = new byte[bufferSize];
>   int numRead;
>   while ((numRead = in.read(buf)) >= 0) {
> {code}
> get() may return null.
> The while loop dereferences in without null check.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HADOOP-10037) s3n read truncated, but doesn't throw exception

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10037?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-10037.
-
   Resolution: Cannot Reproduce
Fix Version/s: 2.6.0

closing as Cannot Reproduce, as it appears to have gone away for you.

# Hadoop 2.6 is using a much later version of jets3t
# Hadoop 2.6 also offers a (compatible) s3a fiesystem which uses the AWS SDK 
instead. 

If you do see this problem, try using s3a to see if it occurs there

> s3n read truncated, but doesn't throw exception 
> 
>
> Key: HADOOP-10037
> URL: https://issues.apache.org/jira/browse/HADOOP-10037
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 2.0.0-alpha
> Environment: Ubuntu Linux 13.04 running on Amazon EC2 (cc2.8xlarge)
>Reporter: David Rosenstrauch
> Fix For: 2.6.0
>
> Attachments: S3ReadFailedOnTruncation.html, S3ReadSucceeded.html
>
>
> For months now we've been finding that we've been experiencing frequent data 
> truncation issues when reading from S3 using the s3n:// protocol.  I finally 
> was able to gather some debugging output on the issue in a job I ran last 
> night, and so can finally file a bug report.
> The job I ran last night was on a 16-node cluster (all of them AWS EC2 
> cc2.8xlarge machines, running Ubuntu 13.04 and Cloudera CDH4.3.0).  The job 
> was a Hadoop streaming job, which reads through a large number (i.e., 
> ~55,000) of files on S3, each of them approximately 300K bytes in size.
> All of the files contain 46 columns of data in each record.  But I added in 
> an extra check in my mapper code to count and verify the number of columns in 
> every record - throwing an error and crashing the map task if the column 
> count is wrong.
> If you look in the attached task logs, you'll see 2 attempts on the same 
> task.  The first one fails due to data truncated (i.e., my job intentionally 
> fails the map task due to the current record failing the column count check). 
>  The task then gets retried on a different machine and runs to a succesful 
> completion.
> You can see further evidence of the truncation further down in the task logs, 
> where it displays the count of the records read:  the failed task says 32953 
> records read, while the successful task says 63133.
> Any idea what the problem might be here and/or how to work around it?  This 
> issue is a very common occurrence on our clusters.  E.g., in the job I ran 
> last night before I had gone to bed I had already encountered 8 such 
> failuers, and the job was only 10% complete.  (~25,000 out of ~250,000 tasks.)
> I realize that it's common for I/O errors to occur - possibly even frequently 
> - in a large Hadoop job.  But I would think that if an I/O failure (like a 
> truncated read) did occur, that something in the underlying infrastructure 
> code (i.e., either in NativeS3FileSystem or in jets3t) should detect the 
> error and throw an IOException accordingly.  It shouldn't be up to the 
> calling code to detect such failures, IMO.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-9248) Allow configuration of Amazon S3 Endpoint

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280235#comment-14280235
 ] 

Steve Loughran commented on HADOOP-9248:


HADOOP-11261 fixes this for the new S3a filesystem. On the assumption that once 
s3a has bedded in and people are confident it works in production, they can 
migrate off s3n —do we need this feature any more?

> Allow configuration of Amazon S3 Endpoint
> -
>
> Key: HADOOP-9248
> URL: https://issues.apache.org/jira/browse/HADOOP-9248
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs/s3
> Environment: All environments connecting to S3
>Reporter: Timur Perelmutov
>
> http://wiki.apache.org/hadoop/AmazonS3 page describes configuration of Hadoop 
> with S3 as storage. Other systems like EMC Atmos now implement S3 Interface, 
> but in order to be able to connect to them, the endpoint needs to be 
> configurable. Please add a configuration parameter that would be propagated  
> to underlying jets3t library as s3service.s3-endpoint param.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8870) NullPointerException when glob doesn't return files

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8870?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280232#comment-14280232
 ] 

Steve Loughran commented on HADOOP-8870:


This doesn't occur branch-2 due to HADOOP-8906. Shall I close this as a "Cannot 
reproduce?"

> NullPointerException when glob doesn't return files
> ---
>
> Key: HADOOP-8870
> URL: https://issues.apache.org/jira/browse/HADOOP-8870
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs, fs/s3
>Affects Versions: 0.20.205.0, 1.0.3
>Reporter: Jaka Jancar
>
> Reading
> {code}s3n://bucket/{a/,b/,c/}{code}
> if one of the globs matches nothing, I get:
> {code}
> Exception in thread "main" java.lang.NullPointerException
>   at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:992)
>   at 
> org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:177)
>   at 
> org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:208)
>   at spark.HadoopRDD.(HadoopRDD.scala:51)
>   at spark.SparkContext.hadoopFile(SparkContext.scala:186)
>   at spark.SparkContext.textFile(SparkContext.scala:155)
>   at 
> com.celtra.analyzer.LogAnalyzer.analyzeSufficientS3Logs(LogAnalyzer.scala:52)
>   at com.celtra.analyzer.App$.main(App.scala:164)
>   at com.celtra.analyzer.App.main(App.scala)
> {code}
> I'm not sure whether this is specific to S3 or all filesystems.
> This was occuring in 0.20.205 and I confirmed it's still present in 1.0.3.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-7352) Contracts of LocalFileSystem and DistributedFileSystem should require FileSystem::listStatus throw IOException not return null upon access error

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7352?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280227#comment-14280227
 ] 

Steve Loughran commented on HADOOP-7352:


HDFS & the FS spec do effectively require an IOE of some form to be raised here.

Localfs does return null if a list operation fails due to an invalid path or 
other access problem. This is a bug: it must be escalated to an IOE for 
consistency with everything else.

If someone were to write a patch I'd review it.

> Contracts of LocalFileSystem and DistributedFileSystem should require 
> FileSystem::listStatus throw IOException not return null upon access error
> 
>
> Key: HADOOP-7352
> URL: https://issues.apache.org/jira/browse/HADOOP-7352
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs, fs/s3
>Reporter: Matt Foley
>Assignee: Matt Foley
>
> In HADOOP-6201 and HDFS-538 it was agreed that FileSystem::listStatus should 
> throw FileNotFoundException instead of returning null, when the target 
> directory did not exist.
> However, in LocalFileSystem implementation today, FileSystem::listStatus 
> still may return null, when the target directory exists but does not grant 
> read permission.  This causes NPE in many callers, for all the reasons cited 
> in HADOOP-6201 and HDFS-538.  See HADOOP-7327 and its linked issues for 
> examples.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HADOOP-9577) Actual data loss using s3n (against US Standard region)

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9577?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-9577.

Resolution: Won't Fix

I'm going to close this as something we don't currently plan to fix in the 
Hadoop core codebase, given that Netflix S3mper and EMR itself both offer a 
solution, namely support on Amazon Dynamo for a consistent metadata store.

The other way to get guaranteed create consistency is "don't use US East", 
which has no consistency guarantees —whereas everything else offers Create , 
but not Update or Delete

> Actual data loss using s3n (against US Standard region)
> ---
>
> Key: HADOOP-9577
> URL: https://issues.apache.org/jira/browse/HADOOP-9577
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 1.0.3
>Reporter: Joshua Caplan
>Priority: Critical
>
>  The implementation of needsTaskCommit() assumes that the FileSystem used for 
> writing temporary outputs is consistent.  That happens not to be the case 
> when using the S3 native filesystem in the US Standard region.  It is 
> actually quite common in larger jobs for the exists() call to return false 
> even if the task attempt wrote output minutes earlier, which essentially 
> cancels the commit operation with no error.  That's real life data loss right 
> there, folks.
> The saddest part is that the Hadoop APIs do not seem to provide any 
> legitimate means for the various RecordWriters to communicate with the 
> OutputCommitter.  In my projects I have created a static map of semaphores 
> keyed by TaskAttemptID, which all my custom RecordWriters have to be aware 
> of.  That's pretty lame.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HADOOP-4436) S3 object names with arbitrary slashes confuse NativeS3FileSystem

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-4436?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-4436.

   Resolution: Won't Fix
Fix Version/s: 2.7.0

closing as WONTFIX...s3n, s3a and openstack clients all assume that a "/" paths 
represent directory delimiters in paths, an assumption that goes pretty deep. 
Sorry

> S3 object names with arbitrary slashes confuse NativeS3FileSystem
> -
>
> Key: HADOOP-4436
> URL: https://issues.apache.org/jira/browse/HADOOP-4436
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 0.18.1
>Reporter: David Phillips
> Fix For: 2.7.0
>
>
> Consider a bucket with the following object names:
> * /
> * /foo
> * foo//bar
> NativeS3FileSystem treats an object named "/" as a directory.  Doing an "fs 
> -lsr" causes an infinite loop.
> I suggest we change NativeS3FileSystem to handle these by ignoring any such 
> "invalid" names.  Thoughts?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11262) Enable YARN to use S3A

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11262?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280205#comment-14280205
 ] 

Steve Loughran commented on HADOOP-11262:
-

h3.  TestS3A

# replace .* import with specific class imports
# you shouldn't have to set the s3a.impl file; it should instead go into 
core-default.xml
# {{testS3AStatus()}}: assertTrue/false should include an error message, 
ideally with the errant values
# {{testS3ACreateFileInSubDir}} should close FS in a try/finally block. You can 
use the Java7 construct to do this now.


> Enable YARN to use S3A 
> ---
>
> Key: HADOOP-11262
> URL: https://issues.apache.org/jira/browse/HADOOP-11262
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>  Labels: amazon, s3
> Attachments: HADOOP-11262-2.patch, HADOOP-11262-3.patch, 
> HADOOP-11262.patch
>
>
> Uses DelegateToFileSystem to expose S3A as an AbstractFileSystem.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-10643) Add NativeS3Fs that delgates calls from FileContext apis to native s3 fs implementation

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10643?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280198#comment-14280198
 ] 

Steve Loughran commented on HADOOP-10643:
-

Sumit -why did you cut a check from {{AbstractFileSystem}}? That could have 
implications elsewhere.

This should be retained, otherwise we've added a new way for things to fail, 
different error messages, etc.

Can you add a new protected method to validate the URI; something which does 
the check —and which your code overrides if it really, really doesn't want a 
port to be specified. That won't impact anything else.



> Add NativeS3Fs that delgates calls from FileContext apis to native s3 fs 
> implementation
> ---
>
> Key: HADOOP-10643
> URL: https://issues.apache.org/jira/browse/HADOOP-10643
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs/s3
>Affects Versions: 2.4.0
>Reporter: Sumit Kumar
>Assignee: Sumit Kumar
> Attachments: HADOOP-10643.patch
>
>
> The new set of file system related apis (FileContext/AbstractFileSystem) 
> already support local filesytem, hdfs, viewfs) however they don't support 
> s3n. This patch is to add that support using configurations like
> fs.AbstractFileSystem.s3n.impl = org.apache.hadoop.fs.s3native.NativeS3Fs
> This patch however doesn't provide a new implementation, instead relies on 
> DelegateToFileSystem abstract class to delegate all calls from FileContext 
> apis for s3n to the NativeS3FileSystem implementation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11171) Enable using a proxy server to connect to S3a.

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280194#comment-14280194
 ] 

Steve Loughran commented on HADOOP-11171:
-

Thomas, this doesn't apply any more. Can you sync it up with trunk/branch-2 and 
I'll test & apply it

> Enable using a proxy server to connect to S3a.
> --
>
> Key: HADOOP-11171
> URL: https://issues.apache.org/jira/browse/HADOOP-11171
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 2.4.0
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Attachments: HADOOP-11171-2.patch, HADOOP-11171-3.patch, 
> HADOOP-11171-4.patch, HADOOP-11171-5.patch, HADOOP-11171-6.patch, 
> HADOOP-11171-7.patch, HADOOP-11171-8.patch, HADOOP-11171.patch
>
>
> This exposes the AWS SDK config for a proxy (host and port) to s3a through 
> config settings.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11350) The size of header buffer of HttpServer is too small when HTTPS is enabled

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280096#comment-14280096
 ] 

Hudson commented on HADOOP-11350:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #809 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/809/])
HADOOP-11350. The size of header buffer of HttpServer is too small when HTTPS 
is enabled. Contributed by Benoy Antony. (wheat9: rev 
3ab3a6498812c9fa0c53dae02ce696033062af87)
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/HttpServerFunctionalTest.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestSSLHttpServer.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> The size of header buffer of HttpServer is too small when HTTPS is enabled
> --
>
> Key: HADOOP-11350
> URL: https://issues.apache.org/jira/browse/HADOOP-11350
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HADOOP-11350.patch, HADOOP-11350.patch, 
> HADOOP-11350.patch, HADOOP-11350.patch
>
>
> {code}
> curl -k  -vvv -i -L --negotiate -u : https://:50070
> < HTTP/1.1 413 FULL head
> HTTP/1.1 413 FULL head
> < Connection: close
> Connection: close
> < Server: Jetty(6.1.26)
> Server: Jetty(6.1.26)
> {code}
> For some users, the spnego token too large for the default header buffer used 
> by Jetty. 
> Though the issue is fixed for HTTP connections (via HADOOP-8816), HTTPS 
> connections needs to be fixed as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8757) Metrics should disallow names with invalid characters

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280102#comment-14280102
 ] 

Hudson commented on HADOOP-8757:


FAILURE: Integrated in Hadoop-Yarn-trunk #809 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/809/])
HADOOP-8757. Metrics should disallow names with invalid characters (rchiang via 
rkanter) (rkanter: rev b6ff9c03a4f8aba945e562a7ff60b0fc6a1cd040)
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMetricsRegistry.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java


> Metrics should disallow names with invalid characters
> -
>
> Key: HADOOP-8757
> URL: https://issues.apache.org/jira/browse/HADOOP-8757
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Todd Lipcon
>Assignee: Ray Chiang
>Priority: Minor
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-8757-01.patch, HADOOP-8757-02.patch, 
> HADOOP-8757.003.patch
>
>
> Just spent a couple hours trying to figure out why a metric I added didn't 
> show up in JMX, only to eventually realize it was because I had a whitespace 
> in the property name. This didn't cause any errors to be logged -- the metric 
> just didn't show up in JMX. We should check that the name is valid and log an 
> error, or replace invalid characters with something like an underscore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280093#comment-14280093
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #809 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/809/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11318) Update the document for hadoop fs -stat

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280094#comment-14280094
 ] 

Hudson commented on HADOOP-11318:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #809 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/809/])
HADOOP-11318. Update the document for hadoop fs -stat (aajisaka: rev 
ce29074685abaa88ae9586fe8277aca23f9d54d6)
* hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Stat.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm


> Update the document for hadoop fs -stat
> ---
>
> Key: HADOOP-11318
> URL: https://issues.apache.org/jira/browse/HADOOP-11318
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.5.1
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-11318-001.patch, HADOOP-11318-002.patch, 
> HADOOP-11318-002.patch, HADOOP-11318-003.patch
>
>
> In FileSystemShell.apt.vm, 
> {code}
> stat
>Usage: <<>>
>Returns the stat information on the path.
> {code}
> Now {{-stat}} accepts the below formats.
>  *   %b: Size of file in blocks
>  *   %g: Group name of owner
>  *   %n: Filename
>  *   %o: Block size
>  *   %r: replication
>  *   %u: User name of owner
>  *   %y: UTC date as "-MM-dd HH:mm:ss"
>  *   %Y: Milliseconds since January 1, 1970 UTC
> They should be documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11483) HardLink.java should use the jdk7 createLink method

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280099#comment-14280099
 ] 

Hudson commented on HADOOP-11483:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #809 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/809/])
HADOOP-11483. HardLink.java should use the jdk7 createLink method (aajisaka: 
rev db51548f706ccd2d0200745ab89e27610c6d10bc)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HardLink.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHardLink.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> HardLink.java should use the jdk7 createLink method
> ---
>
> Key: HADOOP-11483
> URL: https://issues.apache.org/jira/browse/HADOOP-11483
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.7.0
>Reporter: Colin Patrick McCabe
>Assignee: Akira AJISAKA
> Fix For: 2.7.0
>
> Attachments: HDFS-7556-001.patch
>
>
> Now that we are using jdk7, HardLink.java should use the jdk7 createLink 
> method rather than our shell commands or JNI methods.
> Note that we cannot remove all of the JNI / shell commands unless we remove 
> the code which is checking the link count, something that jdk7 doesn't 
> provide (at least, I don't think it does)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-8757) Metrics should disallow names with invalid characters

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280088#comment-14280088
 ] 

Hudson commented on HADOOP-8757:


FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #75 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/75/])
HADOOP-8757. Metrics should disallow names with invalid characters (rchiang via 
rkanter) (rkanter: rev b6ff9c03a4f8aba945e562a7ff60b0fc6a1cd040)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/lib/TestMetricsRegistry.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/lib/MetricsRegistry.java


> Metrics should disallow names with invalid characters
> -
>
> Key: HADOOP-8757
> URL: https://issues.apache.org/jira/browse/HADOOP-8757
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Affects Versions: 2.0.0-alpha, 3.0.0
>Reporter: Todd Lipcon
>Assignee: Ray Chiang
>Priority: Minor
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-8757-01.patch, HADOOP-8757-02.patch, 
> HADOOP-8757.003.patch
>
>
> Just spent a couple hours trying to figure out why a metric I added didn't 
> show up in JMX, only to eventually realize it was because I had a whitespace 
> in the property name. This didn't cause any errors to be logged -- the metric 
> just didn't show up in JMX. We should check that the name is valid and log an 
> error, or replace invalid characters with something like an underscore.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11483) HardLink.java should use the jdk7 createLink method

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280085#comment-14280085
 ] 

Hudson commented on HADOOP-11483:
-

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #75 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/75/])
HADOOP-11483. HardLink.java should use the jdk7 createLink method (aajisaka: 
rev db51548f706ccd2d0200745ab89e27610c6d10bc)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HardLink.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestHardLink.java


> HardLink.java should use the jdk7 createLink method
> ---
>
> Key: HADOOP-11483
> URL: https://issues.apache.org/jira/browse/HADOOP-11483
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 2.7.0
>Reporter: Colin Patrick McCabe
>Assignee: Akira AJISAKA
> Fix For: 2.7.0
>
> Attachments: HDFS-7556-001.patch
>
>
> Now that we are using jdk7, HardLink.java should use the jdk7 createLink 
> method rather than our shell commands or JNI methods.
> Note that we cannot remove all of the JNI / shell commands unless we remove 
> the code which is checking the link count, something that jdk7 doesn't 
> provide (at least, I don't think it does)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280079#comment-14280079
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #75 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/75/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11318) Update the document for hadoop fs -stat

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280080#comment-14280080
 ] 

Hudson commented on HADOOP-11318:
-

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #75 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/75/])
HADOOP-11318. Update the document for hadoop fs -stat (aajisaka: rev 
ce29074685abaa88ae9586fe8277aca23f9d54d6)
* hadoop-common-project/hadoop-common/src/site/apt/FileSystemShell.apt.vm
* hadoop-common-project/hadoop-common/src/test/resources/testConf.xml
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/shell/Stat.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> Update the document for hadoop fs -stat
> ---
>
> Key: HADOOP-11318
> URL: https://issues.apache.org/jira/browse/HADOOP-11318
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.5.1
>Reporter: Akira AJISAKA
>Assignee: Akira AJISAKA
>  Labels: newbie
> Fix For: 2.7.0
>
> Attachments: HADOOP-11318-001.patch, HADOOP-11318-002.patch, 
> HADOOP-11318-002.patch, HADOOP-11318-003.patch
>
>
> In FileSystemShell.apt.vm, 
> {code}
> stat
>Usage: <<>>
>Returns the stat information on the path.
> {code}
> Now {{-stat}} accepts the below formats.
>  *   %b: Size of file in blocks
>  *   %g: Group name of owner
>  *   %n: Filename
>  *   %o: Block size
>  *   %r: replication
>  *   %u: User name of owner
>  *   %y: UTC date as "-MM-dd HH:mm:ss"
>  *   %Y: Milliseconds since January 1, 1970 UTC
> They should be documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11350) The size of header buffer of HttpServer is too small when HTTPS is enabled

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11350?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280082#comment-14280082
 ] 

Hudson commented on HADOOP-11350:
-

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #75 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/75/])
HADOOP-11350. The size of header buffer of HttpServer is too small when HTTPS 
is enabled. Contributed by Benoy Antony. (wheat9: rev 
3ab3a6498812c9fa0c53dae02ce696033062af87)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestSSLHttpServer.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpServer.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/HttpServerFunctionalTest.java


> The size of header buffer of HttpServer is too small when HTTPS is enabled
> --
>
> Key: HADOOP-11350
> URL: https://issues.apache.org/jira/browse/HADOOP-11350
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Benoy Antony
>Assignee: Benoy Antony
> Fix For: 2.7.0
>
> Attachments: HADOOP-11350.patch, HADOOP-11350.patch, 
> HADOOP-11350.patch, HADOOP-11350.patch
>
>
> {code}
> curl -k  -vvv -i -L --negotiate -u : https://:50070
> < HTTP/1.1 413 FULL head
> HTTP/1.1 413 FULL head
> < Connection: close
> Connection: close
> < Server: Jetty(6.1.26)
> Server: Jetty(6.1.26)
> {code}
> For some users, the spnego token too large for the default header buffer used 
> by Jetty. 
> Though the issue is fixed for HTTP connections (via HADOOP-8816), HTTPS 
> connections needs to be fixed as well. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280063#comment-14280063
 ] 

Hudson commented on HADOOP-11261:
-

FAILURE: Integrated in Hadoop-trunk-Commit #6876 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/6876/])
HADOOP-11261 Set custom endpoint for S3A. (Thomas Demoor via stevel) (stevel: 
rev 000ca83ea3aeb3687625c857bcc0762fd2887db2)
* 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java
* 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/TestS3AConfiguration.java
* hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Constants.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-11261.
-
   Resolution: Fixed
Fix Version/s: 2.7.0

> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280055#comment-14280055
 ] 

Steve Loughran commented on HADOOP-11261:
-

All tests pass, reviewed code looks good

+1, committing

Thomas: you are going to need to write a site documentation patch which covers 
this and the other configuration parameters you've got in pending changes

> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Fix For: 2.7.0
>
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11261) Set custom endpoint for S3A

2015-01-16 Thread Steve Loughran (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-11261:

Assignee: Thomas Demoor

> Set custom endpoint for S3A
> ---
>
> Key: HADOOP-11261
> URL: https://issues.apache.org/jira/browse/HADOOP-11261
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Thomas Demoor
>Assignee: Thomas Demoor
>  Labels: amazon, s3
> Attachments: HADOOP-11261-2.patch, HADOOP-11261-3.patch, 
> JIRA-11261.patch
>
>
> Use a config setting to allow customizing the used AWS region. 
> It also enables using a custom url pointing to an S3-compatible object store.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11335) KMS ACL in meta data or database

2015-01-16 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11335?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14280013#comment-14280013
 ] 

Arun Suresh commented on HADOOP-11335:
--

Thanks for working this [~dian.fu]

I have taken an first pass at the latest patch.. my initial comments follow 
(will send out more comments once I do further reviews).

1) JavaKeyStoreProvider:
   * createKey() : I think it might be a bit odd that we have metadata with 
version == 0 
   * deleteKey() : Am sorry, I might be missing something but, I did not quite 
understand why we can't just delete the metadata from the cache when the key is 
deleted

2) KeyProvider:
   * The Metadata Class now has a dependency on {{KeyOpType}} which is defined 
in {{KeyProviderAuthorizationExtension}}. The Extension classes were meant to 
add functionality to a KeyProvider. It seems a bit weird that the KeyProvider 
class should have a dependency on an Extension.
   * Not very confortable adding setters to Metadata. I Guess the original 
implementaion conciously made a choice to not allow modification of metadata 
once it is created (except for the version)

3) KeyProviderExtension:
   * Do we really need the read and write locks ? The undelying KeyProvider 
should take care of the synchronization. (for eg. {{JavaKeyStoreProvider}}) 
does infact use write and read locks for createKey etc.. this would probably 
lead to unnecessary double locking. 

4) KeyProviderAuthorizationExtension:
   * Same as above.. do we really need the read and write locks ? I feel the 
Extension class should handle its own concurrency semantics

5) MetadataKeyAuthorizer
   * Remove commented code

Looking at the commented code in {{MetadataKeyAuthorizer}}, I see that you had 
initially toyed with having an extended {{MetadataWithACL}} class. Any reason 
why you did not pursue that design ? It seems to me like that could have been a 
way to probably avoid having to modify the {{JavaKeyStoreProvider}} and 
{{KeyProvider}}. One suggestion would have been to templatized {{KeyProvider}} 
like so :
{noformat}
  public class KeyProvider
  ...
{noformat}
and have different implemenetation of a {{KeyProvider}} like :
{noformat}
  public classs KeyProviderWithACls extends KeyProvider
  ...  
{noformat}  

> KMS ACL in meta data or database
> 
>
> Key: HADOOP-11335
> URL: https://issues.apache.org/jira/browse/HADOOP-11335
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: kms
>Affects Versions: 2.6.0
>Reporter: Jerry Chen
>Assignee: Dian Fu
>  Labels: Security
> Attachments: HADOOP-11335.001.patch, HADOOP-11335.002.patch, KMS ACL 
> in metadata or database.pdf
>
>   Original Estimate: 504h
>  Remaining Estimate: 504h
>
> Currently Hadoop KMS has implemented ACL for keys and the per key ACL are 
> stored in the configuration file kms-acls.xml.
> The management of ACL in configuration file would not be easy in enterprise 
> usage and it is put difficulties for backup and recovery.
> It is ideal to store the ACL for keys in the key meta data similar to what 
> file system ACL does.  In this way, the backup and recovery that works on 
> keys should work for ACL for keys too.
> On the other hand, with the ACL in meta data, the ACL of each key can be 
> easily manipulate with API or command line tool and take effect instantly.  
> This is very important for enterprise level access control management.  This 
> feature can be addressed by separate JIRA. While with the configuration file, 
> these would be hard to provide.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11485) Pluggable shell integration

2015-01-16 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11485?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14279946#comment-14279946
 ] 

Allen Wittenauer commented on HADOOP-11485:
---

I'm making HADOOP-11460 required, as it is nearly impossible to give 3rd 
parties guidance on what they can/cannot change without removing some of the 
cruft generated by previous shell work.

> Pluggable shell integration
> ---
>
> Key: HADOOP-11485
> URL: https://issues.apache.org/jira/browse/HADOOP-11485
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>Assignee: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11485-00.patch
>
>
> It would be useful to provide a way for core and non-core Hadoop components 
> to plug into the shell infrastructure.  This would allow us to pull the HDFS, 
> MapReduce, and YARN shell functions out of hadoop-functions.sh.  
> Additionally, it should let 3rd parties such as HBase influence things like 
> classpaths at runtime.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11460) Add a way to deprecrate shell vars

2015-01-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11460?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-11460:
--
Attachment: HADOOP-11460-00.patch

-00: 
* initial version
* deprecate a bunch of hdfs, mapred, and yarn vars

> Add a way to deprecrate shell vars
> --
>
> Key: HADOOP-11460
> URL: https://issues.apache.org/jira/browse/HADOOP-11460
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>  Labels: scripts, shell
> Attachments: HADOOP-11460-00.patch
>
>
> It is a very common shell pattern in 3.x to effectively replace sub-project 
> specific vars with generics.  We should have a function that does this 
> replacement and provides a warning to the end user that the old shell var is 
> deprecated.  This will allow us an easier time to remove them in the future.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)