[jira] [Created] (HADOOP-14103) sort out hadoop-aws contract-test-options.xml
Steve Loughran created HADOOP-14103: --- Summary: sort out hadoop-aws contract-test-options.xml Key: HADOOP-14103 URL: https://issues.apache.org/jira/browse/HADOOP-14103 Project: Hadoop Common Issue Type: Sub-task Components: fs/s3, test Affects Versions: 2.8.0 Reporter: Steve Loughran Priority: Minor The doc update of HADOOP-14099 has shown that there's confusion about whether we need a src/test/resources/contract-test-options.xml file. It's documented as needed, branch-2 has it in .gitignore; trunk doesn't. I think it's needed for the contract tests, which the S3A test base extends (And therefore needs). However, we can just put in an SCM managed one and have it just XInclude auth-keys.xml I propose: do that, fix up the testing docs to match -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org
[jira] [Resolved] (HADOOP-14101) MultiObjectDeleteException thrown when writing directly to s3a
[ https://issues.apache.org/jira/browse/HADOOP-14101?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Steve Loughran resolved HADOOP-14101. - Resolution: Duplicate > MultiObjectDeleteException thrown when writing directly to s3a > -- > > Key: HADOOP-14101 > URL: https://issues.apache.org/jira/browse/HADOOP-14101 > Project: Hadoop Common > Issue Type: Bug > Components: fs/s3 >Affects Versions: 2.7.0, 2.7.1, 2.7.2, 2.7.3 >Reporter: Luke Miner > Labels: s3 > > When I write large jobs to s3a in spark, I get non-deterministic > {{MultiObjectDeleteException}}. This seems to happen at the end of the job > when the temporary files are being cleaned up. It is fixed when > {{fs.s3a.multiobjectdelete.enable}} is set to {{false}}. > {code} > Exception in thread "main" > com.amazonaws.services.s3.model.MultiObjectDeleteException: Status Code: 0, > AWS Service: null, AWS Request ID: null, AWS Error Code: null, AWS Error > Message: One or more objects could not be deleted, S3 Extended Request ID: > null > at > com.amazonaws.services.s3.AmazonS3Client.deleteObjects(AmazonS3Client.java:1745) > at org.apache.hadoop.fs.s3a.S3AFileSystem.delete(S3AFileSystem.java:674) > at > org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:90) > at > org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:60) > at > org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:58) > at > org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74) > at > org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115) > at > org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:115) > at > org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:136) > at > org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at > org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:133) > at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:114) > at > org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:86) > at > org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:86) > at > org.apache.spark.sql.execution.datasources.DataSource.write(DataSource.scala:510) > at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:211) > at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:194) > at > org.apache.spark.sql.DataFrameWriter.parquet(DataFrameWriter.scala:488) > at Json2Pq$.main(json2pq.scala:179) > at Json2Pq.main(json2pq.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:736) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:185) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:210) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:124) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > {code} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org
Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/ [Feb 21, 2017 6:55:37 AM] (varunsaxena) YARN-6159. Documentation changes for TimelineV2Client (Naganarasimha G R [Feb 21, 2017 12:43:19 PM] (vinayakumarb) HDFS-11177. 'storagepolicies -getStoragePolicy' command should accept [Feb 21, 2017 6:04:25 PM] (epayne) HDFS-11404. Increase timeout on [Feb 21, 2017 6:28:49 PM] (stevel) HADOOP-13826. S3A Deadlock in multipart copy due to thread pool limits. [Feb 21, 2017 9:44:42 PM] (liuml07) HADOOP-14017. User friendly name for ADLS user and group. Contributed by [Feb 21, 2017 9:55:27 PM] (wang) HDFS-11405. Rename "erasurecode" CLI subcommand to "ec". Contributed by [Feb 21, 2017 10:54:20 PM] (wang) HDFS-11406. Remove unused getStartInstance and getFinalizeInstance in [Feb 21, 2017 11:32:46 PM] (liuml07) HDFS-11430. Separate class InnerNode from class NetworkTopology and make [Feb 22, 2017 6:24:32 AM] (sunilg) YARN-6184. Introduce loading icon in each page of new YARN UI. -1 overall The following subsystems voted -1: asflicense unit The following subsystems voted -1 but were configured to be filtered/ignored: cc javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: Failed junit tests : hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure hadoop.hdfs.web.TestWebHdfsTimeouts hadoop.yarn.server.timeline.webapp.TestTimelineWebServices hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer hadoop.yarn.server.TestContainerManagerSecurity hadoop.yarn.server.TestMiniYarnClusterNodeUtilization cc: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/diff-compile-cc-root.txt [4.0K] javac: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/diff-compile-javac-root.txt [168K] pylint: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/diff-patch-pylint.txt [20K] shellcheck: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/diff-patch-shellcheck.txt [24K] shelldocs: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/diff-patch-shelldocs.txt [12K] whitespace: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/whitespace-eol.txt [11M] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/whitespace-tabs.txt [1.3M] javadoc: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/diff-javadoc-javadoc-root.txt [2.2M] unit: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt [292K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-applicationhistoryservice.txt [12K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt [56K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-tests.txt [324K] asflicense: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/325/artifact/out/patch-asflicense-problems.txt [4.0K] Powered by Apache Yetus 0.5.0-SNAPSHOT http://yetus.apache.org - To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org
Apache Hadoop qbt Report: trunk+JDK8 on Linux/ppc64le
For more details, see https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/ [Feb 21, 2017 6:04:25 PM] (epayne) HDFS-11404. Increase timeout on [Feb 21, 2017 6:28:49 PM] (stevel) HADOOP-13826. S3A Deadlock in multipart copy due to thread pool limits. [Feb 21, 2017 9:44:42 PM] (liuml07) HADOOP-14017. User friendly name for ADLS user and group. Contributed by [Feb 21, 2017 9:55:27 PM] (wang) HDFS-11405. Rename "erasurecode" CLI subcommand to "ec". Contributed by [Feb 21, 2017 10:54:20 PM] (wang) HDFS-11406. Remove unused getStartInstance and getFinalizeInstance in [Feb 21, 2017 11:32:46 PM] (liuml07) HDFS-11430. Separate class InnerNode from class NetworkTopology and make [Feb 22, 2017 6:24:32 AM] (sunilg) YARN-6184. Introduce loading icon in each page of new YARN UI. [Feb 22, 2017 11:43:48 AM] (stevel) HADOOP-14099 Split S3 testing documentation out into its own file. -1 overall The following subsystems voted -1: compile unit The following subsystems voted -1 but were configured to be filtered/ignored: cc javac The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: Failed junit tests : hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewer hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureToleration hadoop.hdfs.web.TestWebHdfsTimeouts hadoop.yarn.server.timeline.TestRollingLevelDB hadoop.yarn.server.timeline.TestTimelineDataManager hadoop.yarn.server.timeline.TestLeveldbTimelineStore hadoop.yarn.server.timeline.webapp.TestTimelineWebServices hadoop.yarn.server.timeline.recovery.TestLeveldbTimelineStateStore hadoop.yarn.server.timeline.TestRollingLevelDBTimelineStore hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer hadoop.yarn.server.resourcemanager.recovery.TestLeveldbRMStateStore hadoop.yarn.server.resourcemanager.TestRMAdminService hadoop.yarn.server.TestMiniYarnClusterNodeUtilization hadoop.yarn.server.TestContainerManagerSecurity hadoop.yarn.server.timeline.TestLevelDBCacheTimelineStore hadoop.yarn.server.timeline.TestOverrideTimelineStoreYarnClient hadoop.yarn.server.timeline.TestEntityGroupFSTimelineStore hadoop.yarn.applications.distributedshell.TestDistributedShell hadoop.mapred.TestShuffleHandler hadoop.mapreduce.v2.hs.TestHistoryServerLeveldbStateStoreService Timed out junit tests : org.apache.hadoop.hdfs.TestLeaseRecovery2 org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache compile: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-compile-root.txt [124K] cc: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-compile-root.txt [124K] javac: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-compile-root.txt [124K] unit: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt [264K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-nodemanager.txt [16K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-applicationhistoryservice.txt [52K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt [72K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-tests.txt [324K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timeline-pluginstorage.txt [28K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-applications_hadoop-yarn-applications-distributedshell.txt [12K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-ui.txt [4.0K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapreduce-client-shuffle.txt [8.0K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/237/artifact/out/patch-unit-hadoop-mapreduce-project_hadoop-mapreduce-client_hadoop-mapre
[jira] [Created] (HADOOP-14104) Client should always ask namenode for kms provider path.
Rushabh S Shah created HADOOP-14104: --- Summary: Client should always ask namenode for kms provider path. Key: HADOOP-14104 URL: https://issues.apache.org/jira/browse/HADOOP-14104 Project: Hadoop Common Issue Type: Bug Components: kms Reporter: Rushabh S Shah Assignee: Rushabh S Shah According to current implementation of kms provider in client conf, there can only be one kms. In multi-cluster environment, if a client is reading encrypted data from multiple clusters it will only get kms token for local cluster. -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org
[jira] [Created] (HADOOP-14106) Fix deprecated FileSystem inefficient calls in unit test
Mingliang Liu created HADOOP-14106: -- Summary: Fix deprecated FileSystem inefficient calls in unit test Key: HADOOP-14106 URL: https://issues.apache.org/jira/browse/HADOOP-14106 Project: Hadoop Common Issue Type: Improvement Reporter: Mingliang Liu [HADOOP-13321] deprecates FileSystem APIs that promote inefficient call patterns. There are existing code patterns in tests and this JIRA is to address them. {code} [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractFSContractTestBase.java:[372,10] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[249,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[251,16] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[264,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[266,16] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[280,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[282,16] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[288,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[290,16] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[306,14] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[308,16] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[313,14] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[315,16] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[340,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[342,16] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[351,14] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[353,16] [deprecation] isDirectory(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[400,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[759,14] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java:[761,16] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FileSystemContractBaseTest.java:[124,18] [deprecation] isFile(Path) in FileSystem has been deprecated [WARNING] /testptch/hadoop/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FileSystemContract
[jira] [Created] (HADOOP-14107) ITestS3GuardListConsistency fails intermittently
Mingliang Liu created HADOOP-14107: -- Summary: ITestS3GuardListConsistency fails intermittently Key: HADOOP-14107 URL: https://issues.apache.org/jira/browse/HADOOP-14107 Project: Hadoop Common Issue Type: Sub-task Components: fs/s3 Affects Versions: HADOOP-13345 Reporter: Mingliang Liu {code} mvn -Dit.test='ITestS3GuardListConsistency' -Dtest=none -Dscale -Ds3guard -Ddynamo -q clean verify --- T E S T S --- Running org.apache.hadoop.fs.s3a.ITestS3GuardListConsistency Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.544 sec <<< FAILURE! - in org.apache.hadoop.fs.s3a.ITestS3GuardListConsistency testListStatusWriteBack(org.apache.hadoop.fs.s3a.ITestS3GuardListConsistency) Time elapsed: 3.147 sec <<< FAILURE! java.lang.AssertionError: Unexpected number of results from metastore. Metastore should only know about /XYZ: DirListingMetadata{path=s3a://mliu-s3guard/test/ListStatusWriteBack, listMap={s3a://mliu-s3guard/test/ListStatusWriteBack/XYZ=PathMetadata{fileStatus=S3AFileStatus{path=s3a://mliu-s3guard/test/ListStatusWriteBack/XYZ; isDirectory=true; modification_time=0; access_time=0; owner=mliu; group=mliu; permission=rwxrwxrwx; isSymlink=false} isEmptyDirectory=true}, s3a://mliu-s3guard/test/ListStatusWriteBack/123=PathMetadata{fileStatus=S3AFileStatus{path=s3a://mliu-s3guard/test/ListStatusWriteBack/123; isDirectory=true; modification_time=0; access_time=0; owner=mliu; group=mliu; permission=rwxrwxrwx; isSymlink=false} isEmptyDirectory=true}}, isAuthoritative=false} at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at org.apache.hadoop.fs.s3a.ITestS3GuardListConsistency.testListStatusWriteBack(ITestS3GuardListConsistency.java:127) {code} See discussion on the parent JIRA [HADOOP-13345]. -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org
[jira] [Created] (HADOOP-14108) CLI MiniCluster: add an option to specify NameNode HTTP port
Takanobu Asanuma created HADOOP-14108: - Summary: CLI MiniCluster: add an option to specify NameNode HTTP port Key: HADOOP-14108 URL: https://issues.apache.org/jira/browse/HADOOP-14108 Project: Hadoop Common Issue Type: Improvement Reporter: Takanobu Asanuma Assignee: Takanobu Asanuma Priority: Minor About CLI MiniCluster, NameNode HTTP port is randomly determined. If you want to see nn-web-ui or do fsck, you need to look for the port number from the minicluster's log. It would be useful if users can specify the port number. -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org