[jira] [Created] (HADOOP-13870) Incorrect behavior of copyFromLocalFile on implementations of FilterFileSystem

2016-12-06 Thread Joe Ellis (JIRA)
Joe Ellis created HADOOP-13870:
--

 Summary: Incorrect behavior of copyFromLocalFile on 
implementations of FilterFileSystem
 Key: HADOOP-13870
 URL: https://issues.apache.org/jira/browse/HADOOP-13870
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Reporter: Joe Ellis


This may be an incorrect assumption on my part, but it was my belief that 
overriding the create method on a FilterFileSystem was sufficient to intercept 
all calls that would write data through the FileSystem. This is apparently not 
true because calling copyFromLocalFile on the FilterFileSystem eventually 
invokes the create method on the wrapped FileSystem. I would expect open -> 
create -> copy(opened, created) to be functionally equivalent to copyFromLocal 
when using the same input and output paths, but this is not the case.

{code:java}
import java.io.IOException;
import java.net.URI;
import java.nio.charset.StandardCharsets;
import org.apache.commons.io.IOUtils;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataInputStream;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.FilterFileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.util.Progressable;
import org.junit.Before;
import org.junit.Test;

public final class CopyFromLocalFileTest {

private static final Path DATA_PATH = new Path("file:///tmp/test_in");
private static final Path OUT_PATH = new Path("file:///tmp/test_out");

private FileSystem localFs;
private FileSystem wrappedFs;

@Before
public void before() throws IOException {
localFs = FileSystem.get(URI.create("file:///"), new Configuration());
wrappedFs = new FailingFileSystem(localFs);

FSDataOutputStream tmpFile = localFs.create(DATA_PATH);
byte[] bytes = "data".getBytes(StandardCharsets.UTF_8);
tmpFile.write(bytes);
tmpFile.close();
}

@Test
public void test_correct() throws IOException {
FSDataInputStream in = wrappedFs.open(DATA_PATH);
FSDataOutputStream out = wrappedFs.create(OUT_PATH);
IOUtils.copy(in, out);
}

@Test
public void test_incorrect() throws IOException {
wrappedFs.copyFromLocalFile(DATA_PATH, OUT_PATH);
}

private static final class FailingFileSystem extends FilterFileSystem {

public FailingFileSystem(FileSystem fs) {
super(fs);
}

@Override
public FSDataOutputStream create(Path f, FsPermission permission, 
boolean overwrite, int bufferSize,
short replication, long blockSize, Progressable progress) 
throws IOException {
throw new IOException("fail");
}

}
}
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-13869) using HADOOP_USER_CLASSPATH_FIRST inconsistently

2016-12-06 Thread Fei Hui (JIRA)
Fei Hui created HADOOP-13869:


 Summary: using HADOOP_USER_CLASSPATH_FIRST inconsistently
 Key: HADOOP-13869
 URL: https://issues.apache.org/jira/browse/HADOOP-13869
 Project: Hadoop Common
  Issue Type: Bug
  Components: scripts
Reporter: Fei Hui


I find HADOOP_USER_CLASSPATH_FIRST is used inconsistently.
I know it doesn't mattter because it affects classpath once 
HADOOP_USER_CLASSPATH_FIRST is not empty
BUT Maybe it's better that using  HADOOP_USER_CLASSPATH_FIRST uniformly



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/ppc64le

2016-12-06 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/178/

[Dec 5, 2016 4:44:40 PM] (weichiu) HDFS-11181. Fuse wrapper has a typo. 
Contributed by Wei-Chiu Chuang.
[Dec 5, 2016 5:35:17 PM] (xiao) HADOOP-13847. KMSWebApp should close 
KeyProviderCryptoExtension.
[Dec 5, 2016 6:48:25 PM] (weichiu) HDFS-11201. Spelling errors in the logging, 
help, assertions and
[Dec 5, 2016 6:54:43 PM] (weichiu) Revert "HDFS-11201. Spelling errors in the 
logging, help, assertions and
[Dec 5, 2016 7:39:34 PM] (jianhe) YARN-5559. Analyse 2.8.0/3.0.0 jdiff reports 
and fix any issues.
[Dec 5, 2016 7:48:58 PM] (liuml07) HDFS-11094. Send back HAState along with 
NamespaceInfo during a
[Dec 5, 2016 8:04:07 PM] (liuml07) HADOOP-13675. Bug in return value for 
delete() calls in WASB.
[Dec 5, 2016 8:44:20 PM] (xyao) Revert "HADOOP-10930. Refactor: Wrap Datanode 
IO related operations.
[Dec 6, 2016 1:23:38 AM] (naganarasimha_gr) YARN-5921. Incorrect 
synchronization in
[Dec 6, 2016 1:36:00 AM] (xiao) HADOOP-13864. KMS should not require truststore 
password. Contributed by
[Dec 6, 2016 2:13:53 AM] (wang) HDFS-10581. Hide redundant table on NameNode 
WebUI when no nodes are
[Dec 6, 2016 5:01:47 AM] (aajisaka) HADOOP-13835. Move Google Test Framework 
code from mapreduce to
[Dec 6, 2016 7:09:35 AM] (wang) Revert "HDFS-11156. Add new op 
GETFILEBLOCKLOCATIONS to WebHDFS REST
[Dec 6, 2016 7:18:18 AM] (wang) HADOOP-13861. Spelling errors in logging and 
exceptions for code.




-1 overall


The following subsystems voted -1:
compile unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc javac


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.hdfs.TestEncryptedTransfer 
   hadoop.hdfs.server.datanode.TestDataNodeLifeline 
   hadoop.hdfs.TestDFSShellGenericOptions 
   hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewer 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.yarn.server.nodemanager.recovery.TestNMLeveldbStateStoreService 
   hadoop.yarn.server.nodemanager.TestNodeManagerShutdown 
   hadoop.yarn.server.timeline.TestRollingLevelDB 
   hadoop.yarn.server.timeline.TestTimelineDataManager 
   hadoop.yarn.server.timeline.TestLeveldbTimelineStore 
   hadoop.yarn.server.timeline.webapp.TestTimelineWebServices 
   hadoop.yarn.server.timeline.recovery.TestLeveldbTimelineStateStore 
   hadoop.yarn.server.timeline.TestRollingLevelDBTimelineStore 
   
hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer 
   hadoop.yarn.server.timelineservice.storage.common.TestRowKeys 
   hadoop.yarn.server.timelineservice.storage.common.TestKeyConverters 
   hadoop.yarn.server.timelineservice.storage.common.TestSeparator 
   hadoop.yarn.server.resourcemanager.recovery.TestLeveldbRMStateStore 
   hadoop.yarn.server.resourcemanager.TestRMRestart 
   hadoop.yarn.server.resourcemanager.TestResourceTrackerService 
   hadoop.yarn.server.TestMiniYarnClusterNodeUtilization 
   hadoop.yarn.server.TestContainerManagerSecurity 
   hadoop.yarn.server.timeline.TestLevelDBCacheTimelineStore 
   hadoop.yarn.server.timeline.TestOverrideTimelineStoreYarnClient 
   hadoop.yarn.server.timeline.TestEntityGroupFSTimelineStore 
   hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageApps 
   
hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowRunCompaction
 
   
hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageEntities 
   hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowRun 
   
hadoop.yarn.server.timelineservice.storage.TestPhoenixOfflineAggregationWriterImpl
 
   
hadoop.yarn.server.timelineservice.reader.TestTimelineReaderWebServicesHBaseStorage
 
   
hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowActivity 
   hadoop.yarn.applications.distributedshell.TestDistributedShell 
   hadoop.mapred.TestShuffleHandler 
   hadoop.mapreduce.v2.hs.TestHistoryServerLeveldbStateStoreService 

Timed out junit tests :

   org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache 
  

   compile:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/178/artifact/out/patch-compile-root.txt
  [172K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/178/artifact/out/patch-compile-root.txt
  [172K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/178/artifact/out/patch-compile-root.txt
  [172K]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/178/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [204K]
   

[jira] [Created] (HADOOP-13868) Configure multi-part copies and uploads separately

2016-12-06 Thread Sean Mackrory (JIRA)
Sean Mackrory created HADOOP-13868:
--

 Summary: Configure multi-part copies and uploads separately
 Key: HADOOP-13868
 URL: https://issues.apache.org/jira/browse/HADOOP-13868
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Sean Mackrory
Assignee: Sean Mackrory


I've been looking at a big performance regression when writing to S3 from Spark 
that appears to have been introduced with HADOOP-12891.

In the Amazon SDK, the default threshold for multi-part copies is 320x the 
threshold for multi-part uploads (and the block size is 20x bigger), so I don't 
think it's wise for us 

I did some quick tests and it seems to me the sweet spot when multi-part copies 
start being faster is around 512MB. It wasn't as significant, but using 
104857600 (Amazon's default) for the blocksize was also slightly better.

I propose we do the following, although they're independent.

(1) Split the configuration. Ideally, I'd like to have 
fs.s3a.multipart.copy.threshold and fs.s3a.multipart.upload.threshold (and 
corresponding properties for the block size). But then there's the question of 
what to do with the existing fs.s3a.multipart.* properties. Deprecation? Leave 
it as a short-hand for configuring both (that's overridden by the more specific 
properties?).

(2) Consider increasing the default values. In my tests, 256 MB seemed to be 
where multipart uploads came into their own, and 512 MB was where multipart 
copies started outperforming the alternative. Would be interested to hear what 
other people have seen.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-12854) Move to netty 4.1.x release

2016-12-06 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12854?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu resolved HADOOP-12854.
-
Resolution: Duplicate

Dup of HADOOP-13866

> Move to netty 4.1.x release
> ---
>
> Key: HADOOP-12854
> URL: https://issues.apache.org/jira/browse/HADOOP-12854
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: build
>Affects Versions: 2.8.0
>Reporter: Steve Loughran
>
> Netty is getting close to having a final release of a 4.1 netty-all artifact; 
> HDFS currently pulls in 4.1.0.Beta5
> Once a 4.1 release is out, switch to it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Reopened] (HADOOP-13866) Upgrade netty-all to 4.1.1.Final

2016-12-06 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-13866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu reopened HADOOP-13866:
-

> Upgrade netty-all to 4.1.1.Final
> 
>
> Key: HADOOP-13866
> URL: https://issues.apache.org/jira/browse/HADOOP-13866
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Ted Yu
>Assignee: Ted Yu
> Attachments: HADOOP-13866.v1.patch, HADOOP-13866.v2.patch
>
>
> netty-all 4.1.1.Final is stable release which we should upgrade to.
> See bottom of HADOOP-12927 for related discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2016-12-06 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/

[Dec 5, 2016 4:44:40 PM] (weichiu) HDFS-11181. Fuse wrapper has a typo. 
Contributed by Wei-Chiu Chuang.
[Dec 5, 2016 5:35:17 PM] (xiao) HADOOP-13847. KMSWebApp should close 
KeyProviderCryptoExtension.
[Dec 5, 2016 6:48:25 PM] (weichiu) HDFS-11201. Spelling errors in the logging, 
help, assertions and
[Dec 5, 2016 6:54:43 PM] (weichiu) Revert "HDFS-11201. Spelling errors in the 
logging, help, assertions and
[Dec 5, 2016 7:39:34 PM] (jianhe) YARN-5559. Analyse 2.8.0/3.0.0 jdiff reports 
and fix any issues.
[Dec 5, 2016 7:48:58 PM] (liuml07) HDFS-11094. Send back HAState along with 
NamespaceInfo during a
[Dec 5, 2016 8:04:07 PM] (liuml07) HADOOP-13675. Bug in return value for 
delete() calls in WASB.
[Dec 5, 2016 8:44:20 PM] (xyao) Revert "HADOOP-10930. Refactor: Wrap Datanode 
IO related operations.
[Dec 6, 2016 1:23:38 AM] (naganarasimha_gr) YARN-5921. Incorrect 
synchronization in
[Dec 6, 2016 1:36:00 AM] (xiao) HADOOP-13864. KMS should not require truststore 
password. Contributed by
[Dec 6, 2016 2:13:53 AM] (wang) HDFS-10581. Hide redundant table on NameNode 
WebUI when no nodes are
[Dec 6, 2016 5:01:47 AM] (aajisaka) HADOOP-13835. Move Google Test Framework 
code from mapreduce to




-1 overall


The following subsystems voted -1:
asflicense findbugs unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.hdfs.TestRollingUpgrade 
   hadoop.yarn.server.timeline.webapp.TestTimelineWebServices 
   hadoop.yarn.server.resourcemanager.TestRMRestart 
   hadoop.yarn.server.TestContainerManagerSecurity 
   hadoop.yarn.server.TestMiniYarnClusterNodeUtilization 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-compile-javac-root.txt
  [168K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-checkstyle-root.txt
  [16M]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-patch-pylint.txt
  [20K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-patch-shellcheck.txt
  [28K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-patch-shelldocs.txt
  [16K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/whitespace-eol.txt
  [11M]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/whitespace-tabs.txt
  [1.3M]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-tests.txt
  [4.0K]

   javadoc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/diff-javadoc-javadoc-root.txt
  [2.2M]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [152K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-applicationhistoryservice.txt
  [12K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
  [72K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-tests.txt
  [316K]

   asflicense:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/247/artifact/out/patch-asflicense-problems.txt
  [4.0K]

Powered by Apache Yetus 0.4.0-SNAPSHOT   http://yetus.apache.org



-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org