[jira] [Commented] (HADOOP-10149) Create ByteBuffer-based cipher API

2013-12-12 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13847237#comment-13847237
 ] 

Yi Liu commented on HADOOP-10149:
-

Hi Owen,  as for “we should use the standard javax.security.Cipher API with a 
customer provider that is based on openssl.”, we had already done the work, you 
can find it on 10150. 

> Create ByteBuffer-based cipher API
> --
>
> Key: HADOOP-10149
> URL: https://issues.apache.org/jira/browse/HADOOP-10149
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Owen O'Malley
>Assignee: Owen O'Malley
>
> As part of HDFS-5143, [~hitliuyi] included a ByteBuffer-based API for 
> encryption and decryption. Especially, because of the zero-copy work this 
> seems like an important piece of work. 
> This API should be discussed independently instead of just as part of 
> HDFS-5143.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13847064#comment-13847064
 ] 

Ted Yu commented on HADOOP-10163:
-

I will try to digest the above code.

However, for JIRA with long discussion, e.g. HBASE-8755, searching comments for 
attachment Id seems inefficient.

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-8753) LocalDirAllocator throws "ArithmeticException: / by zero" when there is no available space on configured local dir

2013-12-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846897#comment-13846897
 ] 

Hudson commented on HADOOP-8753:


FAILURE: Integrated in Hadoop-trunk-Commit #4876 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4876/])
HADOOP-8753. LocalDirAllocator throws ArithmeticException: divide by zero when 
there is no available space on configured local dir. Contributed by Benoy 
Antony. (hitesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1550570)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/LocalDirAllocator.java


> LocalDirAllocator throws "ArithmeticException: / by zero" when there is no 
> available space on configured local dir
> --
>
> Key: HADOOP-8753
> URL: https://issues.apache.org/jira/browse/HADOOP-8753
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.0.2-alpha
>Reporter: Nishan Shetty, Huawei
>Assignee: Benoy Antony
>Priority: Minor
> Fix For: 2.4.0
>
> Attachments: HADOOP-8753.1.copy.patch, HADOOP-8753.1.patch, 
> YARN-16.patch
>
>
> 12/08/09 13:59:49 INFO mapreduce.Job: Task Id : 
> attempt_1344492468506_0023_m_00_0, Status : FAILED
> java.lang.ArithmeticException: / by zero
> at 
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:371)
> at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:150)
> at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:131)
> at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:115)
> at 
> org.apache.hadoop.yarn.server.nodemanager.LocalDirsHandlerService.getLocalPathForWrite(LocalDirsHandlerService.java:257)
> at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:849)
> Instead of throwing exception directly we can log a warning saying no 
> available space on configured local dir



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-8753) LocalDirAllocator throws "ArithmeticException: / by zero" when there is no available space on configured local dir

2013-12-12 Thread Hitesh Shah (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hitesh Shah updated HADOOP-8753:


   Resolution: Fixed
Fix Version/s: 2.4.0
   Status: Resolved  (was: Patch Available)

Committed to trunk and branch-2. Thanks for the patience [~benoyantony]

> LocalDirAllocator throws "ArithmeticException: / by zero" when there is no 
> available space on configured local dir
> --
>
> Key: HADOOP-8753
> URL: https://issues.apache.org/jira/browse/HADOOP-8753
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.0.2-alpha
>Reporter: Nishan Shetty, Huawei
>Assignee: Benoy Antony
>Priority: Minor
> Fix For: 2.4.0
>
> Attachments: HADOOP-8753.1.copy.patch, HADOOP-8753.1.patch, 
> YARN-16.patch
>
>
> 12/08/09 13:59:49 INFO mapreduce.Job: Task Id : 
> attempt_1344492468506_0023_m_00_0, Status : FAILED
> java.lang.ArithmeticException: / by zero
> at 
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:371)
> at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:150)
> at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:131)
> at 
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:115)
> at 
> org.apache.hadoop.yarn.server.nodemanager.LocalDirsHandlerService.getLocalPathForWrite(LocalDirsHandlerService.java:257)
> at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:849)
> Instead of throwing exception directly we can log a warning saying no 
> available space on configured local dir



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846852#comment-13846852
 ] 

Brock Noland commented on HADOOP-10163:
---

also later we match the file name.

{noformat}
# validate the patch name, parse branch if needed
shopt -s nocasematch
PATCH_NAME=$(basename $PATCH_URL)
# Test examples:
# HIVE-123.patch HIVE-123.1.patch HIVE-123.D123.patch HIVE-123.D123.1.patch 
HIVE-123-tez.patch HIVE-123.1-tez.patch
# HIVE-.patch, HIVE-.XX.patch  HIVE-.XX-branch.patch 
HIVE--branch.patch
if [[ $PATCH_NAME =~ ^HIVE-[0-9]+(\.[0-9]+)?(-[a-z0-9-]+)?\.(patch|patch.\txt)$ 
]]
then
  if [[ -n "${BASH_REMATCH[2]}" ]]
  then
BRANCH=${BASH_REMATCH[2]#*-}
  else
echo "Assuming branch $BRANCH"
  fi
# HIVE-.D.patch or HIVE-.D.XX.patch
elif [[ $PATCH_NAME =~ ^(HIVE-[0-9]+\.)?D[0-9]+(\.[0-9]+)?\.(patch|patch.\txt)$ 
]]
then
  echo "Assuming branch $BRANCH"
else
  echo "Patch $PATCH_NAME does not appear to be a patch"
  exit 1
fi
{noformat}

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846844#comment-13846844
 ] 

Brock Noland commented on HADOOP-10163:
---

Here is what we do with hive where $JIRA_TEXT is a file containing the entire 
HTML page:

{noformat}
# pull attachments from JIRA (hack stolen from hadoop since rest api doesn't 
show attachments)
PATCH_URL=$(grep -o '"/jira/secure/attachment/[0-9]*/[^"]*' $JIRA_TEXT | \
  grep -v -e 'htm[l]*$' | sort | tail -1 | \
  grep -o '/jira/secure/attachment/[0-9]*/[^"]*')
if [[ -z "$PATCH_URL" ]]
then
  echo "Unable to find attachment for $JIRA_NAME"
  exit 1
fi
# ensure attachment has not already been tested
ATTACHMENT_ID=$(basename $(dirname $PATCH_URL))
if grep -q "ATTACHMENT ID: $ATTACHMENT_ID" $JIRA_TEXT
then
  echo "Attachment $ATTACHMENT_ID is already tested for $JIRA_NAME"
  exit 1
fi
{noformat}

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846820#comment-13846820
 ] 

Ted Yu commented on HADOOP-10163:
-

The above scenario actually may happen to Hive:
Patch A was tested on day 1.
B.htm is attached to same JIRA on day 2.
test-patch.sh would filter out B.htm and run patch A again.

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846765#comment-13846765
 ] 

Ted Yu commented on HADOOP-10163:
-

Here is proposed change from HBASE-10044 :
{code}
+relativePatchURL=`$GREP -o '"/jira/secure/attachment/[0-9]*/[^"]*' 
$PATCH_DIR/jira | $EGREP '(\.txt$|\.patch$|\.diff$)' | sort | tail -1 | $GREP 
-o '/jira/secure/attachment/[0-9]*/[^"]*'`
{code}
Suppose patch A was tested on day 1 - JIRA stayed in Patch Available status.
Some more comments were then added.
Tread dump named thread.out was attached.
QA bot would be triggered due to the new attachment. However the above change 
would filter out the dump file and go back to patch A which has been run.

The attachment ID of last patch should be known so that test-patch.sh can 
decide that no new patch has been attached and bail out.

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846759#comment-13846759
 ] 

Brock Noland commented on HADOOP-10163:
---

We don't iterate jira comments, it's a simple (jira text contains "ATTACHMENT 
ID XXX").

Not sure if this holds for other projects but in hive a specific attachment is 
only tested once.

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10110) hadoop-auth has a build break due to missing dependency

2013-12-12 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846691#comment-13846691
 ] 

Arpit Agarwal commented on HADOOP-10110:


What's the target branch for 2.0.x? Thanks.

> hadoop-auth has a build break due to missing dependency
> ---
>
> Key: HADOOP-10110
> URL: https://issues.apache.org/jira/browse/HADOOP-10110
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 3.0.0, 2.0.6-alpha
>Reporter: Chuan Liu
>Assignee: Chuan Liu
>Priority: Blocker
> Fix For: 3.0.0, 2.3.0
>
> Attachments: HADOOP-10110.patch
>
>
> We have a build break in hadoop-auth if build with maven cache cleaned. The 
> error looks like the follows. The problem exists on both Windows and Linux. 
> If you have old jetty jars in your maven cache, you won't see the error.
> {noformat}
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> [INFO] Total time: 1:29.469s
> [INFO] Finished at: Mon Nov 18 12:30:36 PST 2013
> [INFO] Final Memory: 37M/120M
> [INFO] 
> 
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:2.5.1:testCompile 
> (default-testCompile) on project hadoop-auth: Compilation failure: 
> Compilation failure:
> [ERROR] 
> /home/chuan/trunk/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java:[84,13]
>  cannot access org.mortbay.component.AbstractLifeCycle
> [ERROR] class file for org.mortbay.component.AbstractLifeCycle not found
> [ERROR] server = new Server(0);
> [ERROR] 
> /home/chuan/trunk/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java:[94,29]
>  cannot access org.mortbay.component.LifeCycle
> [ERROR] class file for org.mortbay.component.LifeCycle not found
> [ERROR] server.getConnectors()[0].setHost(host);
> [ERROR] 
> /home/chuan/trunk/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java:[96,10]
>  cannot find symbol
> [ERROR] symbol  : method start()
> [ERROR] location: class org.mortbay.jetty.Server
> [ERROR] 
> /home/chuan/trunk/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java:[102,12]
>  cannot find symbol
> [ERROR] symbol  : method stop()
> [ERROR] location: class org.mortbay.jetty.Server
> [ERROR] -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn  -rf :hadoop-auth
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


How can I make a custom counter in the MapReduce engine?

2013-12-12 Thread Yoonmin Nam
I tried to make a my counter to check the operation time of Maptask.run and
ReduceTask.run using global counter or even LOG.info().

In a pseudo-distributed mode, using builded mapreduce-core-snapshot.jar
which I modified in the MapTask.java and ReduceTask.java to make my goal, it
works perfectly as I expected.

However, in a real cluster, the source code I injected isn't work and just
system-default counter variables are showing.

I think there are different logging logic or security issue between
pseudo-distributed and cluster mode, but I am not sure about that.

Is there anyone who can let me know about it?

Thanks!

 



[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846646#comment-13846646
 ] 

Ted Yu commented on HADOOP-10163:
-

Do you have sample code on how to iterate through JIRA comments ?

Thanks

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846634#comment-13846634
 ] 

Brock Noland commented on HADOOP-10163:
---

Yep

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10162) Fix symlink-related test failures in TestFileContextResolveAfs and TestStat in branch-2

2013-12-12 Thread Mit Desai (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846622#comment-13846622
 ] 

Mit Desai commented on HADOOP-10162:


Thanks Colin. I will keep this in mind from now on :-)

> Fix symlink-related test failures in TestFileContextResolveAfs and TestStat 
> in branch-2
> ---
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Fix For: 2.4.0
>
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10044) Improve the javadoc of rpc code

2013-12-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846602#comment-13846602
 ] 

Hudson commented on HADOOP-10044:
-

SUCCESS: Integrated in Hadoop-trunk-Commit #4874 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4874/])
HADOOP-10044 Improve the javadoc of rpc code (sanjay Radia) (sradia: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1550486)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/RpcConstants.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java


> Improve the javadoc of rpc code
> ---
>
> Key: HADOOP-10044
> URL: https://issues.apache.org/jira/browse/HADOOP-10044
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
>Priority: Minor
> Attachments: HADOOP-10044.20131014.patch, hadoop-10044.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Brock Noland (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846585#comment-13846585
 ] 

Brock Noland commented on HADOOP-10163:
---

We handled this scenario in Hive by placing "ATTACHMENT ID " in 
the HiveQA comment:

https://issues.apache.org/jira/browse/HIVE-5973?focusedCommentId=13846235&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13846235

If Hadoop/HBase did that then they could eliminate duplicate runs as we have 
done in Hive.

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Ted Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846592#comment-13846592
 ] 

Ted Yu commented on HADOOP-10163:
-

That approach would imply scanning comments until "ATTACHMENT ID " is found, 
right ?

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10044) Improve the javadoc of rpc code

2013-12-12 Thread Sanjay Radia (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10044?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sanjay Radia updated HADOOP-10044:
--

  Resolution: Fixed
Target Version/s: 2.3.0
  Status: Resolved  (was: Patch Available)

The failed test timeout is unrelated (and I also ran it successfully).
Committed.

> Improve the javadoc of rpc code
> ---
>
> Key: HADOOP-10044
> URL: https://issues.apache.org/jira/browse/HADOOP-10044
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Sanjay Radia
>Assignee: Sanjay Radia
>Priority: Minor
> Attachments: HADOOP-10044.20131014.patch, hadoop-10044.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10162) Fix symlink-related test failures in TestFileContextResolveAfs and TestStat in branch-2

2013-12-12 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HADOOP-10162:
--

  Resolution: Fixed
   Fix Version/s: 2.4.0
Target Version/s: 2.4.0
  Status: Resolved  (was: Patch Available)

> Fix symlink-related test failures in TestFileContextResolveAfs and TestStat 
> in branch-2
> ---
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Fix For: 2.4.0
>
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10162) Fix symlink-related test failures in TestFileContextResolveAfs and TestStat in branch-2

2013-12-12 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846582#comment-13846582
 ] 

Colin Patrick McCabe commented on HADOOP-10162:
---

Thanks for looking at this, Mit.  Since this patch is for branch-2, it does not 
require a Jenkins run and we usually don't press "patch available."  I renamed 
this to "Fix symlink-related test failures in TestFileContextResolveAfs and 
TestStat in branch-2" since that's really what this patch is about (the 
disabling patch is already backported to these branches)  I have run the tests 
and verified that it fixes them as promised.  +1.

> Fix symlink-related test failures in TestFileContextResolveAfs and TestStat 
> in branch-2
> ---
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10162) Fix symlink-related test failures in TestFileContextResolveAfs and TestStat in branch-2

2013-12-12 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HADOOP-10162:
--

Summary: Fix symlink-related test failures in TestFileContextResolveAfs and 
TestStat in branch-2  (was: BACKPORT HADOOP-10052 to Branch-2)

> Fix symlink-related test failures in TestFileContextResolveAfs and TestStat 
> in branch-2
> ---
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10163) Attachment Id for last tested patch should be passed to test-patch.sh

2013-12-12 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10163?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated HADOOP-10163:


Summary: Attachment Id for last tested patch should be passed to 
test-patch.sh  (was: Enhance jenkinsPrecommitAdmin.py to pass attachment Id for 
last tested patch)

> Attachment Id for last tested patch should be passed to test-patch.sh
> -
>
> Key: HADOOP-10163
> URL: https://issues.apache.org/jira/browse/HADOOP-10163
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>
> In HBASE-10044, attempt was made to filter attachments according to known 
> file extensions.
> However, that change alone wouldn't work because when non-patch is attached, 
> QA bot doesn't provide attachment Id for last tested patch.
> This results in the modified test-patch.sh to seek backward and launch 
> duplicate test run for last tested patch.
> If attachment Id for last tested patch is provided, test-patch.sh can decide 
> whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Created] (HADOOP-10163) Enhance jenkinsPrecommitAdmin.py to pass attachment Id for last tested patch

2013-12-12 Thread Ted Yu (JIRA)
Ted Yu created HADOOP-10163:
---

 Summary: Enhance jenkinsPrecommitAdmin.py to pass attachment Id 
for last tested patch
 Key: HADOOP-10163
 URL: https://issues.apache.org/jira/browse/HADOOP-10163
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Ted Yu


In HBASE-10044, attempt was made to filter attachments according to known file 
extensions.
However, that change alone wouldn't work because when non-patch is attached, QA 
bot doesn't provide attachment Id for last tested patch.
This results in the modified test-patch.sh to seek backward and launch 
duplicate test run for last tested patch.

If attachment Id for last tested patch is provided, test-patch.sh can decide 
whether there is need to run test.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10162) BACKPORT HADOOP-10052 to Branch-2

2013-12-12 Thread Mit Desai (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846452#comment-13846452
 ] 

Mit Desai commented on HADOOP-10162:


It was supposed to fail as this change is only for branch 2

> BACKPORT HADOOP-10052 to Branch-2
> -
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10162) BACKPORT HADOOP-10052 to Branch-2

2013-12-12 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846435#comment-13846435
 ] 

Hadoop QA commented on HADOOP-10162:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12618431/HADOOP-10162.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:red}-1 javac{color:red}.  The patch appears to cause the build to 
fail.

Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/3357//console

This message is automatically generated.

> BACKPORT HADOOP-10052 to Branch-2
> -
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10162) BACKPORT HADOOP-10052 to Branch-2

2013-12-12 Thread Mit Desai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mit Desai updated HADOOP-10162:
---

Status: Patch Available  (was: Open)

> BACKPORT HADOOP-10052 to Branch-2
> -
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10162) BACKPORT HADOOP-10052 to Branch-2

2013-12-12 Thread Mit Desai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mit Desai updated HADOOP-10162:
---

Attachment: HADOOP-10162.patch

Patch posted ONLY for branch-2. This change does not go into trunk.

> BACKPORT HADOOP-10052 to Branch-2
> -
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HADOOP-10162.patch
>
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Updated] (HADOOP-10162) BACKPORT HADOOP-10052 to Branch-2

2013-12-12 Thread Jonathan Eagles (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jonathan Eagles updated HADOOP-10162:
-

Assignee: Mit Desai  (was: Jonathan Eagles)

> BACKPORT HADOOP-10052 to Branch-2
> -
>
> Key: HADOOP-10162
> URL: https://issues.apache.org/jira/browse/HADOOP-10162
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 2.4.0
>Reporter: Mit Desai
>Assignee: Mit Desai
>
> We need to backport HADOOP-10052  to branch-2 as I found that the 
> TestFileContextResolveAfs is failing after HADOOP-10020 went in.
> Also the test TestStat is failing for the same reason. It needs to be fixed 
> as well.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (HADOOP-10087) UserGroupInformation.getGroupNames() fails to return primary group first when JniBasedUnixGroupsMappingWithFallback is used

2013-12-12 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10087?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13846372#comment-13846372
 ] 

Hudson commented on HADOOP-10087:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1610 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1610/])
HADOOP-10087. UserGroupInformation.getGroupNames() fails to return primary 
group first when JniBasedUnixGroupsMappingWithFallback is used (cmccabe) 
(cmccabe: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1550229)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/native/src/org/apache/hadoop/security/hadoop_user_info.c


> UserGroupInformation.getGroupNames() fails to return primary group first when 
> JniBasedUnixGroupsMappingWithFallback is used
> ---
>
> Key: HADOOP-10087
> URL: https://issues.apache.org/jira/browse/HADOOP-10087
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.1.0-beta, 2.2.0
> Environment: SUSE Linux Enterprise Server 11 (x86_64)
>Reporter: Yu Gao
>Assignee: Colin Patrick McCabe
>  Labels: security
> Fix For: 2.3.0
>
> Attachments: HADOOP-10087.001.patch, HADOOP-10087.002.patch
>
>
> When JniBasedUnixGroupsMappingWithFallback is used as the group mapping 
> resolution provider, UserGroupInformation.getGroupNames() fails to return the 
> primary group first in the list as documented.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)