Re: [PR] HDFS-17249. Fix TestDFSUtil.testIsValidName() unit test failure [hadoop]

2023-11-09 Thread via GitHub


hadoop-yetus commented on PR #6249:
URL: https://github.com/apache/hadoop/pull/6249#issuecomment-1805222187

   :confetti_ball: **+1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   9m 11s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  0s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  0s |  |  detect-secrets was not available.  
|
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +0 :ok: |  mvndep  |  14m 17s |  |  Maven dependency ordering for branch  |
   | +1 :green_heart: |  mvninstall  |  20m 12s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   3m 11s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  compile  |   3m  4s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  checkstyle  |   0m 53s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   1m 35s |  |  trunk passed  |
   | +1 :green_heart: |  javadoc  |   1m 25s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   1m 45s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   3m 28s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  25m 34s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 22s |  |  Maven dependency ordering for patch  |
   | +1 :green_heart: |  mvninstall  |   1m 37s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   3m 50s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javac  |   3m 50s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   3m 21s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  javac  |   3m 21s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | -0 :warning: |  checkstyle  |   0m 45s | 
[/results-checkstyle-hadoop-hdfs-project.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6249/3/artifact/out/results-checkstyle-hadoop-hdfs-project.txt)
 |  hadoop-hdfs-project: The patch generated 2 new + 81 unchanged - 0 fixed = 
83 total (was 81)  |
   | +1 :green_heart: |  mvnsite  |   1m 18s |  |  the patch passed  |
   | +1 :green_heart: |  javadoc  |   1m  5s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   1m 25s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   3m 31s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  21m 39s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |   1m 59s |  |  hadoop-hdfs-client in the patch 
passed.  |
   | +1 :green_heart: |  unit  | 199m  6s |  |  hadoop-hdfs in the patch 
passed.  |
   | +1 :green_heart: |  asflicense  |   0m 38s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 326m 54s |  |  |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.43 ServerAPI=1.43 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6249/3/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/6249 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets |
   | uname | Linux 77c4ea86da58 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 
13:30:12 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / 8eec4d1ef52ce1d3368cd97ea5081590681089b5 |
   | Default Java | Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 |
   | Multi-JDK versions | 
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 
/usr/lib/jvm/java-8-openjdk-amd64:Private 
Build-1.8.0_382-8u382-ga-1~20.04.1-b05 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6249/3/testReport/ |
   | Max. process+thread count | 3576 (vs. ulimit of 5500) |
   | modules | C: hadoop-hdfs-project/hadoop-hdfs-client 
hadoop-hdfs-project/hadoop-hdfs U: hadoop-hdfs-project |
   | Console output | 
https://ci-hadoop.apache.org/job/hadoop-multibran

Re: [PR] HDFS-15413. add dfs.client.read.striped.datanode.max.attempts to fix read ecfile timeout [hadoop]

2023-11-09 Thread via GitHub


ayushtkn commented on code in PR #5829:
URL: https://github.com/apache/hadoop/pull/5829#discussion_r1388949742


##
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSStripedInputStreamWithTimeout.java:
##
@@ -0,0 +1,168 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdfs;
+
+import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.ErasureCodingPolicy;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
+import org.apache.hadoop.hdfs.protocol.LocatedStripedBlock;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.datanode.DataNodeTestUtils;
+import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset;
+import org.apache.hadoop.io.erasurecode.CodecUtil;
+import org.apache.hadoop.io.erasurecode.ErasureCodeNative;
+import 
org.apache.hadoop.io.erasurecode.rawcoder.NativeRSRawErasureCoderFactory;
+import org.apache.hadoop.test.GenericTestUtils;
+import org.junit.After;
+import org.junit.Assert;
+import org.junit.Before;
+import org.junit.Rule;
+import org.junit.Test;
+import org.junit.rules.Timeout;
+
+import java.io.IOException;
+import java.util.Arrays;
+
+public class TestDFSStripedInputStreamWithTimeout {
+
+  public static final Logger LOG =
+  LoggerFactory.getLogger(TestDFSStripedInputStreamWithTimeout.class);
+
+  private MiniDFSCluster cluster;
+  private Configuration conf = new Configuration();
+  private DistributedFileSystem fs;
+  private final Path dirPath = new Path("/striped");
+  private Path filePath = new Path(dirPath, "file");
+  private ErasureCodingPolicy ecPolicy;
+  private short dataBlocks;
+  private short parityBlocks;
+  private int cellSize;
+  private final int stripesPerBlock = 2;
+  private int blockSize;
+  private int blockGroupSize;
+
+  @Rule
+  public Timeout globalTimeout = new Timeout(30);
+
+  public ErasureCodingPolicy getEcPolicy() {
+return StripedFileTestUtil.getDefaultECPolicy();
+  }
+
+  @Before
+  public void setup() throws IOException {
+/*
+ * Initialize erasure coding policy.
+ */
+ecPolicy = getEcPolicy();
+dataBlocks = (short) ecPolicy.getNumDataUnits();
+parityBlocks = (short) ecPolicy.getNumParityUnits();
+cellSize = ecPolicy.getCellSize();
+blockSize = stripesPerBlock * cellSize;
+blockGroupSize = dataBlocks * blockSize;
+System.out.println("EC policy = " + ecPolicy);
+
+conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, blockSize);

Review Comment:
   This is deprecated config I believe, We should use 
``HdfsClientConfigKeys.DFS_CLIENT_SOCKET_TIMEOUT_KEY``



##
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSStripedInputStreamWithTimeout.java:
##
@@ -0,0 +1,168 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ * 
+ * http://www.apache.org/licenses/LICENSE-2.0
+ * 
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hdfs;
+
+import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.ErasureCodin

[jira] [Updated] (HADOOP-18956) Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and ZKSignerSecretProvider

2023-11-09 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated HADOOP-18956:

Labels: pull-request-available  (was: )

> Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and 
> ZKSignerSecretProvider
> --
>
> Key: HADOOP-18956
> URL: https://issues.apache.org/jira/browse/HADOOP-18956
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Zita Dombi
>Assignee: István Fajth
>Priority: Major
>  Labels: pull-request-available
>
> HADOOP-18709 added support for Zookeeper to communicate with SSL/TLS enabled 
> in hadoop-common. With those changes we have the necessary parameters, that 
> we need to set to enable SSL/TLS in a ZK Client. That change also did changes 
> in ZKCuratorManager, so with that it is easy to set the SSL/TLS, for Yarn it 
> was done in YARN-11468.
> In DelegationTokenAuthenticationFilter currently we are using 
> CuratorFrameworkFactory, it'd be good to change it to use ZKCuratorManager 
> and with that we should support SSL/TLS enablement.
> *UPDATE*
> So as I investigated this a bit more, it wouldn't be so easy to move to using 
> ZKCuratorManager. 
> DelegationTokenAuthenticationFilter uses ZK from two places: in 
> ZKDelegationTokenSecretManager and in ZKSignerSecretProvider. In both places 
> it uses CuratorFrameworkFactory, but the attributes and creation 
> differentiates from ZKCuratorManager. 
> In ZKDelegationTokenSecretManager it would be easy to add the new config and 
> based on that create ZK with CuratorFrameworkFactory. But 
> ZKSignerSecretProvider is in hadoop-auth module and with my change it would 
> need hadoop-common, so it would introduce circular dependency between modules 
> 'hadoop-auth' and 'hadoop-common'. I'm still working on a straightforward 
> solution. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[PR] HADOOP-18956. Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and ZKSignerSecretProvider [hadoop]

2023-11-09 Thread via GitHub


fapifta opened a new pull request, #6263:
URL: https://github.com/apache/hadoop/pull/6263

   ### Description of PR
   Zookeeper based delegation token and Zookeeper based signer secret provider 
is not able to use an SSL/TLS based communication even if Zookeeper is able to 
handle such connections.
   The pull request standardizes how ZKDelegationTokenSecretManager, and 
ZKSignerSecretProvider creates their respective CuratorFramework instance (as 
they are anyway interconnected).
   In the new code, they both utilize 
o.a.h.security.authentication.util.ZookeeperClient class to configure the 
client based on their respective configuration values.
   
   The change also introduces a new property in CommonConfigurationKeys, that 
affects YARN as well.
   The new key is `hadoop.zk.ssl.enabled`. With this new configuration the 
following dynamics are true:
   - if `hadoop.zk.ssl.enabled` is set then YARN will respect that and won't 
use the value in `yarn.resourcemanager.zk-client-ssl.enabled` to decide if SSL 
is enabled.
   - if `hadoop.zk.ssl.enabled` is set then the ZKDelegationTokenSecretManager 
will respect that and won't use the value in `zk-dt-secret-manager.ssl.enabled` 
to decide if SSL is enabled.
   - ZKSignerSecretProvider can not respect hadoop-common configurations, as it 
is in hadoop-auth, and common depends on auth not vice versa, so it will 
independently use the configured `signer.secret.provider.zookeeper.ssl.enabled` 
property wich defaults to false.
   - DFSZKFailoverController will also respect `hadoop.zk.ssl.enabled` prior to 
evaluating `dfs.ha.zkfc.client.ssl.enabled`.
   
   `hadoop.zk.ssl.enabled` does not have a default value set.
   
   The intent is to make it possible to enable SSL/TLS towards Zookeeper at 
once, or for all 4 places separately if one wish to do so.
   ZkDelegationTokenSecretManager, and ZKSignerSecretProvider has their own 
Truststore and Keystore overrides, those take precedence over the related 
hadoop.zk.* properties, so custom keystores and truststores can be configured 
even if the common setup is already set. (DFSZKFailoverController and YARN uses 
the hadoop.zk.* properties and does not have custom properties to set the 
truststore and keystore as it was implemented earlier).
   
   ### How was this patch tested?
   Added a JUnit test that checks how the new class introduced to create the 
CuratorFramework instance sets up the builder. From that point on it is 
Curator's responsibility to use the configuration as expected.
   Additionally some other tests should cover the functionality that should be 
provided exactly the same way as before.
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18956) Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and ZKSignerSecretProvider

2023-11-09 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17784665#comment-17784665
 ] 

ASF GitHub Bot commented on HADOOP-18956:
-

fapifta opened a new pull request, #6263:
URL: https://github.com/apache/hadoop/pull/6263

   ### Description of PR
   Zookeeper based delegation token and Zookeeper based signer secret provider 
is not able to use an SSL/TLS based communication even if Zookeeper is able to 
handle such connections.
   The pull request standardizes how ZKDelegationTokenSecretManager, and 
ZKSignerSecretProvider creates their respective CuratorFramework instance (as 
they are anyway interconnected).
   In the new code, they both utilize 
o.a.h.security.authentication.util.ZookeeperClient class to configure the 
client based on their respective configuration values.
   
   The change also introduces a new property in CommonConfigurationKeys, that 
affects YARN as well.
   The new key is `hadoop.zk.ssl.enabled`. With this new configuration the 
following dynamics are true:
   - if `hadoop.zk.ssl.enabled` is set then YARN will respect that and won't 
use the value in `yarn.resourcemanager.zk-client-ssl.enabled` to decide if SSL 
is enabled.
   - if `hadoop.zk.ssl.enabled` is set then the ZKDelegationTokenSecretManager 
will respect that and won't use the value in `zk-dt-secret-manager.ssl.enabled` 
to decide if SSL is enabled.
   - ZKSignerSecretProvider can not respect hadoop-common configurations, as it 
is in hadoop-auth, and common depends on auth not vice versa, so it will 
independently use the configured `signer.secret.provider.zookeeper.ssl.enabled` 
property wich defaults to false.
   - DFSZKFailoverController will also respect `hadoop.zk.ssl.enabled` prior to 
evaluating `dfs.ha.zkfc.client.ssl.enabled`.
   
   `hadoop.zk.ssl.enabled` does not have a default value set.
   
   The intent is to make it possible to enable SSL/TLS towards Zookeeper at 
once, or for all 4 places separately if one wish to do so.
   ZkDelegationTokenSecretManager, and ZKSignerSecretProvider has their own 
Truststore and Keystore overrides, those take precedence over the related 
hadoop.zk.* properties, so custom keystores and truststores can be configured 
even if the common setup is already set. (DFSZKFailoverController and YARN uses 
the hadoop.zk.* properties and does not have custom properties to set the 
truststore and keystore as it was implemented earlier).
   
   ### How was this patch tested?
   Added a JUnit test that checks how the new class introduced to create the 
CuratorFramework instance sets up the builder. From that point on it is 
Curator's responsibility to use the configuration as expected.
   Additionally some other tests should cover the functionality that should be 
provided exactly the same way as before.
   




> Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and 
> ZKSignerSecretProvider
> --
>
> Key: HADOOP-18956
> URL: https://issues.apache.org/jira/browse/HADOOP-18956
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Zita Dombi
>Assignee: István Fajth
>Priority: Major
>
> HADOOP-18709 added support for Zookeeper to communicate with SSL/TLS enabled 
> in hadoop-common. With those changes we have the necessary parameters, that 
> we need to set to enable SSL/TLS in a ZK Client. That change also did changes 
> in ZKCuratorManager, so with that it is easy to set the SSL/TLS, for Yarn it 
> was done in YARN-11468.
> In DelegationTokenAuthenticationFilter currently we are using 
> CuratorFrameworkFactory, it'd be good to change it to use ZKCuratorManager 
> and with that we should support SSL/TLS enablement.
> *UPDATE*
> So as I investigated this a bit more, it wouldn't be so easy to move to using 
> ZKCuratorManager. 
> DelegationTokenAuthenticationFilter uses ZK from two places: in 
> ZKDelegationTokenSecretManager and in ZKSignerSecretProvider. In both places 
> it uses CuratorFrameworkFactory, but the attributes and creation 
> differentiates from ZKCuratorManager. 
> In ZKDelegationTokenSecretManager it would be easy to add the new config and 
> based on that create ZK with CuratorFrameworkFactory. But 
> ZKSignerSecretProvider is in hadoop-auth module and with my change it would 
> need hadoop-common, so it would introduce circular dependency between modules 
> 'hadoop-auth' and 'hadoop-common'. I'm still working on a straightforward 
> solution. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18956) Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and ZKSignerSecretProvider

2023-11-09 Thread Jira


 [ 
https://issues.apache.org/jira/browse/HADOOP-18956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

István Fajth updated HADOOP-18956:
--
Summary: Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and 
ZKSignerSecretProvider  (was: Zookeeper SSL/TLS support in 
DelegationTokenAuthenticationFilter)

> Zookeeper SSL/TLS support in ZKDelegationTokenSecretManager and 
> ZKSignerSecretProvider
> --
>
> Key: HADOOP-18956
> URL: https://issues.apache.org/jira/browse/HADOOP-18956
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Zita Dombi
>Assignee: István Fajth
>Priority: Major
>
> HADOOP-18709 added support for Zookeeper to communicate with SSL/TLS enabled 
> in hadoop-common. With those changes we have the necessary parameters, that 
> we need to set to enable SSL/TLS in a ZK Client. That change also did changes 
> in ZKCuratorManager, so with that it is easy to set the SSL/TLS, for Yarn it 
> was done in YARN-11468.
> In DelegationTokenAuthenticationFilter currently we are using 
> CuratorFrameworkFactory, it'd be good to change it to use ZKCuratorManager 
> and with that we should support SSL/TLS enablement.
> *UPDATE*
> So as I investigated this a bit more, it wouldn't be so easy to move to using 
> ZKCuratorManager. 
> DelegationTokenAuthenticationFilter uses ZK from two places: in 
> ZKDelegationTokenSecretManager and in ZKSignerSecretProvider. In both places 
> it uses CuratorFrameworkFactory, but the attributes and creation 
> differentiates from ZKCuratorManager. 
> In ZKDelegationTokenSecretManager it would be easy to add the new config and 
> based on that create ZK with CuratorFrameworkFactory. But 
> ZKSignerSecretProvider is in hadoop-auth module and with my change it would 
> need hadoop-common, so it would introduce circular dependency between modules 
> 'hadoop-auth' and 'hadoop-common'. I'm still working on a straightforward 
> solution. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18931) FileSystem.getFileSystemClass() to log at debug the jar the .class came from

2023-11-09 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18931?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17784609#comment-17784609
 ] 

ASF GitHub Bot commented on HADOOP-18931:
-

virajjasani commented on PR #6197:
URL: https://github.com/apache/hadoop/pull/6197#issuecomment-1804710842

   @goiri could you please review this PR?




> FileSystem.getFileSystemClass() to log at debug the jar the .class came from
> 
>
> Key: HADOOP-18931
> URL: https://issues.apache.org/jira/browse/HADOOP-18931
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs
>Affects Versions: 3.3.6
>Reporter: Steve Loughran
>Priority: Minor
>  Labels: pull-request-available
>
> we want to be able to log the jar the filesystem implementation class, so 
> that we can identify which version of a module the class came from.
> this is to help track down problems where different machines in the cluster 
> or the .tar.gz bundle is out of date. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HADOOP-18931. FileSystem.getFileSystemClass() to log at debug the jar the .class came from [hadoop]

2023-11-09 Thread via GitHub


virajjasani commented on PR #6197:
URL: https://github.com/apache/hadoop/pull/6197#issuecomment-1804710842

   @goiri could you please review this PR?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18967) Allow secure mode to be enabled with no downtime

2023-11-09 Thread Charles Connell (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18967?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Charles Connell updated HADOOP-18967:
-
Description: 
My employer (HubSpot) recently completed transitioning all of the Hadoop 
clusters underlying our HBase databases into secure mode. It was important to 
us that we be able to make this change without impacting the functionality of 
our SaaS product. To accomplish this, we added some new settings to our fork of 
Hadoop, and fixed a latent bug. This ticket is my intention to contribute these 
changes back to the mainline code, so others can benefit. A patch will be 
incoming.

The basic theme of the new functionality is the ability to accept incoming 
secure connections without requiring them or making them outgoing. Secure mode 
enablement will then be done in two stages.
 * First, all nodes are given configuration to accept secure connections, and 
are gracefully rolling-restarted to adopt this new functionality. I'll be 
adding the new settings to make this stage possible.
 * Second, all nodes are told to require incoming connections be secure, and to 
make secure outgoing connections, and the settings added in the first stage are 
removed. Nodes are again rolling-restarted to adopt this functionality. The 
settings in this final state will look the same as in any secure Hadoop cluster 
today.

I'll include documentation changes explaining how to do this.

  was:
My employer (HubSpot) recently completed transitioning all of the Hadoop 
clusters underlying our HBase databases into secure mode. It was important to 
us that we be able to make this change without impacting the functionality of 
our SaaS product. To accomplish this, we added some new settings to our fork of 
Hadoop, and fixed a latent bug. This ticket is my intention to contribute these 
changes back to the mainline code, so others can benefits. A PR will be 
incoming.

The basic theme of the new functionality is the ability to accept incoming 
secure connections without requiring them or making them outgoing. Secure mode 
enablement will then be done in two stages.
 * First, all nodes are given configuration to accept secure connections, and 
are gracefully rolling-restarted to adopt this new functionality. I'll be 
adding the new settings to make this stage possible.
 * Second, all nodes are told to require incoming connections be secure, and to 
make secure outgoing connections, and the settings added in the first stage are 
removed. Nodes are again rolling-restarted to adopt this functionality. The 
settings in this final state will look the same as in any secure Hadoop cluster 
today.

I'll include documentation changes explaining how to do this.


> Allow secure mode to be enabled with no downtime
> 
>
> Key: HADOOP-18967
> URL: https://issues.apache.org/jira/browse/HADOOP-18967
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Charles Connell
>Priority: Minor
>
> My employer (HubSpot) recently completed transitioning all of the Hadoop 
> clusters underlying our HBase databases into secure mode. It was important to 
> us that we be able to make this change without impacting the functionality of 
> our SaaS product. To accomplish this, we added some new settings to our fork 
> of Hadoop, and fixed a latent bug. This ticket is my intention to contribute 
> these changes back to the mainline code, so others can benefit. A patch will 
> be incoming.
> The basic theme of the new functionality is the ability to accept incoming 
> secure connections without requiring them or making them outgoing. Secure 
> mode enablement will then be done in two stages.
>  * First, all nodes are given configuration to accept secure connections, and 
> are gracefully rolling-restarted to adopt this new functionality. I'll be 
> adding the new settings to make this stage possible.
>  * Second, all nodes are told to require incoming connections be secure, and 
> to make secure outgoing connections, and the settings added in the first 
> stage are removed. Nodes are again rolling-restarted to adopt this 
> functionality. The settings in this final state will look the same as in any 
> secure Hadoop cluster today.
> I'll include documentation changes explaining how to do this.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18967) Allow secure mode to be enabled with no downtime

2023-11-09 Thread Charles Connell (Jira)
Charles Connell created HADOOP-18967:


 Summary: Allow secure mode to be enabled with no downtime
 Key: HADOOP-18967
 URL: https://issues.apache.org/jira/browse/HADOOP-18967
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Charles Connell


My employer (HubSpot) recently completed transitioning all of the Hadoop 
clusters underlying our HBase databases into secure mode. It was important to 
us that we be able to make this change without impacting the functionality of 
our SaaS product. To accomplish this, we added some new settings to our fork of 
Hadoop, and fixed a latent bug. This ticket is my intention to contribute these 
changes back to the mainline code, so others can benefits. A PR will be 
incoming.

The basic theme of the new functionality is the ability to accept incoming 
secure connections without requiring them or making them outgoing. Secure mode 
enablement will then be done in two stages.
 * First, all nodes are given configuration to accept secure connections, and 
are gracefully rolling-restarted to adopt this new functionality. I'll be 
adding the new settings to make this stage possible.
 * Second, all nodes are told to require incoming connections be secure, and to 
make secure outgoing connections, and the settings added in the first stage are 
removed. Nodes are again rolling-restarted to adopt this functionality. The 
settings in this final state will look the same as in any secure Hadoop cluster 
today.

I'll include documentation changes explaining how to do this.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HDFS-17223. Add journalnode maintenance node list [hadoop]

2023-11-09 Thread via GitHub


hadoop-yetus commented on PR #6183:
URL: https://github.com/apache/hadoop/pull/6183#issuecomment-1804197786

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   0m 27s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  1s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  1s |  |  detect-secrets was not available.  
|
   | +0 :ok: |  xmllint  |   0m  1s |  |  xmllint was not available.  |
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 2 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  32m 21s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   0m 49s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  compile  |   0m 47s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  checkstyle  |   0m 46s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   0m 53s |  |  trunk passed  |
   | +1 :green_heart: |  javadoc  |   0m 49s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   1m  8s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   1m 56s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  21m 40s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 40s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   0m 46s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javac  |   0m 46s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   0m 40s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  javac  |   0m 40s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | -0 :warning: |  checkstyle  |   0m 34s | 
[/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/5/artifact/out/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs-project/hadoop-hdfs: The patch generated 10 new + 247 unchanged 
- 0 fixed = 257 total (was 247)  |
   | +1 :green_heart: |  mvnsite  |   0m 43s |  |  the patch passed  |
   | -1 :x: |  javadoc  |   0m 36s | 
[/results-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/5/artifact/out/results-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04.txt)
 |  
hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 
with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 generated 1 new + 98 
unchanged - 1 fixed = 99 total (was 99)  |
   | +1 :green_heart: |  javadoc  |   1m  1s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   1m 52s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  21m 19s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | -1 :x: |  unit  | 194m 23s | 
[/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/5/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs in the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 35s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 286m  5s |  |  |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed junit tests | 
hadoop.hdfs.qjournal.client.TestQuorumJournalManagerUnit |
   |   | hadoop.hdfs.TestDFSUtil |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.43 ServerAPI=1.43 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/5/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/6183 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets xmllint |
   | uname | Linux 89187ef1a3fa 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 
13:30:12 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / fc74e49f909f52160e93d57585e1a1da

Re: [PR] YARN-11610. [Federation] Add WeightedHomePolicyManager. [hadoop]

2023-11-09 Thread via GitHub


hadoop-yetus commented on PR #6256:
URL: https://github.com/apache/hadoop/pull/6256#issuecomment-1804189093

   :confetti_ball: **+1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   0m 27s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  1s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  1s |  |  detect-secrets was not available.  
|
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 1 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  32m 22s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   0m 32s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  compile  |   0m 30s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  checkstyle  |   0m 25s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   0m 33s |  |  trunk passed  |
   | +1 :green_heart: |  javadoc  |   0m 36s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   0m 30s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   1m  1s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  20m 48s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 24s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   0m 25s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javac  |   0m 25s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   0m 22s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  javac  |   0m 22s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | +1 :green_heart: |  checkstyle  |   0m 14s |  |  the patch passed  |
   | +1 :green_heart: |  mvnsite  |   0m 24s |  |  the patch passed  |
   | +1 :green_heart: |  javadoc  |   0m 24s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   0m 22s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   0m 57s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  20m 45s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | +1 :green_heart: |  unit  |   2m 46s |  |  hadoop-yarn-server-common in 
the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 29s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   |  87m 47s |  |  |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.43 ServerAPI=1.43 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6256/3/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/6256 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets |
   | uname | Linux 139bafe44b96 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 
13:30:12 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / 7f1095497b51c23bef48941f9bc3137069a35d0b |
   | Default Java | Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05 |
   | Multi-JDK versions | 
/usr/lib/jvm/java-11-openjdk-amd64:Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 
/usr/lib/jvm/java-8-openjdk-amd64:Private 
Build-1.8.0_382-8u382-ga-1~20.04.1-b05 |
   |  Test Results | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6256/3/testReport/ |
   | Max. process+thread count | 635 (vs. ulimit of 5500) |
   | modules | C: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common |
   | Console output | 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6256/3/console |
   | versions | git=2.25.1 maven=3.6.3 spotbugs=4.2.2 |
   | Powered by | Apache Yetus 0.14.0 https://yetus.apache.org |
   
   
   This message was automatically generated.
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-un

Re: [PR] MAPREDUCE-7461: Fixed assertionComparision failure by resolving xml path for 'name' [hadoop]

2023-11-09 Thread via GitHub


kavvya97 commented on PR #6252:
URL: https://github.com/apache/hadoop/pull/6252#issuecomment-1804132740

   Requesting for review


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18966) s3a file rename does double HEAD or LIST on source file/dir

2023-11-09 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18966:

Summary: s3a file rename does double HEAD or LIST on source file/dir  (was: 
s3a file rename does double HEAD or LIST on source fiile/dir)

> s3a file rename does double HEAD or LIST on source file/dir
> ---
>
> Key: HADOOP-18966
> URL: https://issues.apache.org/jira/browse/HADOOP-18966
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.3.6
>Reporter: Steve Loughran
>Priority: Minor
>
> going to add this as a v2 issue, though really it's been latent.
> when you do a rename() we do a getFileStatus call to get the list/file 
> status..if this is a dir that's a single LIST; a file then it'll be LIST + 
> HEAD.
> when the actual copy is kicked off, 
> * if the source is a dir, then a duplicate LIST is initiated straight 
> afterwards.
> * FIle: there's a second HEAD to get the metadata for the copy
> Proposed.
> * LIST initiates a full list, and if it returns a list of objects, that list 
> iterator is passed in to rename.
> * full result of HEAD preserved and passed down to copyFile()
> will cut out one round trip regardless of source type



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-18956) Zookeeper SSL/TLS support in DelegationTokenAuthenticationFilter

2023-11-09 Thread Zita Dombi (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18956?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zita Dombi reassigned HADOOP-18956:
---

Assignee: István Fajth  (was: Zita Dombi)

> Zookeeper SSL/TLS support in DelegationTokenAuthenticationFilter
> 
>
> Key: HADOOP-18956
> URL: https://issues.apache.org/jira/browse/HADOOP-18956
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Zita Dombi
>Assignee: István Fajth
>Priority: Major
>
> HADOOP-18709 added support for Zookeeper to communicate with SSL/TLS enabled 
> in hadoop-common. With those changes we have the necessary parameters, that 
> we need to set to enable SSL/TLS in a ZK Client. That change also did changes 
> in ZKCuratorManager, so with that it is easy to set the SSL/TLS, for Yarn it 
> was done in YARN-11468.
> In DelegationTokenAuthenticationFilter currently we are using 
> CuratorFrameworkFactory, it'd be good to change it to use ZKCuratorManager 
> and with that we should support SSL/TLS enablement.
> *UPDATE*
> So as I investigated this a bit more, it wouldn't be so easy to move to using 
> ZKCuratorManager. 
> DelegationTokenAuthenticationFilter uses ZK from two places: in 
> ZKDelegationTokenSecretManager and in ZKSignerSecretProvider. In both places 
> it uses CuratorFrameworkFactory, but the attributes and creation 
> differentiates from ZKCuratorManager. 
> In ZKDelegationTokenSecretManager it would be easy to add the new config and 
> based on that create ZK with CuratorFrameworkFactory. But 
> ZKSignerSecretProvider is in hadoop-auth module and with my change it would 
> need hadoop-common, so it would introduce circular dependency between modules 
> 'hadoop-auth' and 'hadoop-common'. I'm still working on a straightforward 
> solution. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] YARN-11577. Improve FederationInterceptorREST Method Result. [hadoop]

2023-11-09 Thread via GitHub


slfan1989 commented on PR #6190:
URL: https://github.com/apache/hadoop/pull/6190#issuecomment-1804026482

   @goiri Can you help review this PR? Thank you very much! 
   
   I've added some comments to the code. There are 3 purposes for this PR
   
   - Fixed some cases where the Router and RM interfaces returned inconsistent 
results, especially when empty results were returned.
   - Fixed the inconsistency between Router and RM interfaces.
   - Added some unit tests for CapacityScheduler and FairScheduler. These unit 
tests are tested using real test clusters.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] YARN-11610. [Federation] Add WeightedHomePolicyManager. [hadoop]

2023-11-09 Thread via GitHub


slfan1989 commented on PR #6256:
URL: https://github.com/apache/hadoop/pull/6256#issuecomment-1804013761

   @goiri Can you help review this PR? Thank you very much!


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18965) ITestS3AHugeFilesEncryption failure

2023-11-09 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17784501#comment-17784501
 ] 

ASF GitHub Bot commented on HADOOP-18965:
-

ahmarsuhail commented on code in PR #6261:
URL: https://github.com/apache/hadoop/pull/6261#discussion_r1388132245


##
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/ITestS3AHugeFilesEncryption.java:
##
@@ -68,13 +70,16 @@ protected boolean isEncrypted(S3AFileSystem fileSystem) {
   @Override
   protected void assertEncrypted(Path hugeFile) throws IOException {
 Configuration c = new Configuration();
-String kmsKey = getS3EncryptionKey(getTestBucketName(c), c);
-if (SSE_KMS.getMethod().equals(c.get(S3_ENCRYPTION_ALGORITHM))) {
+
+final String bucket = getTestBucketName(c);
+String kmsKey = getS3EncryptionKey(bucket, c);

Review Comment:
   nit: consider changing variable name from c to conf





> ITestS3AHugeFilesEncryption failure
> ---
>
> Key: HADOOP-18965
> URL: https://issues.apache.org/jira/browse/HADOOP-18965
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> test failures for me with a test setup of per-bucket encryption of sse-kms.
> suspect (but can't guarantee) HADOOP-18850 may be a factor.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HADOOP-18965. ITestS3AHugeFilesEncryption failure [hadoop]

2023-11-09 Thread via GitHub


ahmarsuhail commented on code in PR #6261:
URL: https://github.com/apache/hadoop/pull/6261#discussion_r1388132245


##
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/scale/ITestS3AHugeFilesEncryption.java:
##
@@ -68,13 +70,16 @@ protected boolean isEncrypted(S3AFileSystem fileSystem) {
   @Override
   protected void assertEncrypted(Path hugeFile) throws IOException {
 Configuration c = new Configuration();
-String kmsKey = getS3EncryptionKey(getTestBucketName(c), c);
-if (SSE_KMS.getMethod().equals(c.get(S3_ENCRYPTION_ALGORITHM))) {
+
+final String bucket = getTestBucketName(c);
+String kmsKey = getS3EncryptionKey(bucket, c);

Review Comment:
   nit: consider changing variable name from c to conf



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] YARN-11483. [Federation] Router AdminCLI Supports Clean Finish Apps. [hadoop]

2023-11-09 Thread via GitHub


slfan1989 commented on PR #6251:
URL: https://github.com/apache/hadoop/pull/6251#issuecomment-1803981261

   @goiri Can you help review this PR? Thank you very much! We have added a new 
command to the router command-line tool for deleting applications. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18966) s3a file rename does double HEAD or LIST on source fiile/dir

2023-11-09 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-18966:
---

 Summary: s3a file rename does double HEAD or LIST on source 
fiile/dir
 Key: HADOOP-18966
 URL: https://issues.apache.org/jira/browse/HADOOP-18966
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.3.6
Reporter: Steve Loughran


going to add this as a v2 issue, though really it's been latent.

when you do a rename() we do a getFileStatus call to get the list/file 
status..if this is a dir that's a single LIST; a file then it'll be LIST + HEAD.

when the actual copy is kicked off, 
* if the source is a dir, then a duplicate LIST is initiated straight 
afterwards.
* FIle: there's a second HEAD to get the metadata for the copy

Proposed.
* LIST initiates a full list, and if it returns a list of objects, that list 
iterator is passed in to rename.
* full result of HEAD preserved and passed down to copyFile()

will cut out one round trip regardless of source type





--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18965) ITestS3AHugeFilesEncryption failure

2023-11-09 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18965?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17784485#comment-17784485
 ] 

ASF GitHub Bot commented on HADOOP-18965:
-

steveloughran commented on PR #6261:
URL: https://github.com/apache/hadoop/pull/6261#issuecomment-1803968401

   could I get a review by another, e.g @ahmarsuhail ?




> ITestS3AHugeFilesEncryption failure
> ---
>
> Key: HADOOP-18965
> URL: https://issues.apache.org/jira/browse/HADOOP-18965
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> test failures for me with a test setup of per-bucket encryption of sse-kms.
> suspect (but can't guarantee) HADOOP-18850 may be a factor.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HADOOP-18965. ITestS3AHugeFilesEncryption failure [hadoop]

2023-11-09 Thread via GitHub


steveloughran commented on PR #6261:
URL: https://github.com/apache/hadoop/pull/6261#issuecomment-1803968401

   could I get a review by another, e.g @ahmarsuhail ?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HDFS-17223. Add journalnode maintenance node list [hadoop]

2023-11-09 Thread via GitHub


hadoop-yetus commented on PR #6183:
URL: https://github.com/apache/hadoop/pull/6183#issuecomment-1803928579

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   0m 52s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  1s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  1s |  |  detect-secrets was not available.  
|
   | +0 :ok: |  xmllint  |   0m  1s |  |  xmllint was not available.  |
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 3 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  48m 25s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   1m 33s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  compile  |   1m 23s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  checkstyle  |   1m 15s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   1m 28s |  |  trunk passed  |
   | +1 :green_heart: |  javadoc  |   1m 12s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   1m 41s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   3m 39s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  41m 44s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   1m 15s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   1m 23s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javac  |   1m 23s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   1m 11s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  javac  |   1m 11s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | -0 :warning: |  checkstyle  |   1m  3s | 
[/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/4/artifact/out/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs-project/hadoop-hdfs: The patch generated 10 new + 248 unchanged 
- 0 fixed = 258 total (was 248)  |
   | +1 :green_heart: |  mvnsite  |   1m 14s |  |  the patch passed  |
   | -1 :x: |  javadoc  |   0m 57s | 
[/results-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/4/artifact/out/results-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04.txt)
 |  
hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 
with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 generated 1 new + 98 
unchanged - 1 fixed = 99 total (was 99)  |
   | +1 :green_heart: |  javadoc  |   1m 34s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   3m 31s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  41m 58s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | -1 :x: |  unit  | 256m 11s | 
[/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/4/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs in the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 47s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 415m  8s |  |  |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed junit tests | hadoop.hdfs.TestRollingUpgrade |
   |   | hadoop.hdfs.qjournal.client.TestQuorumJournalManager |
   |   | hadoop.hdfs.qjournal.client.TestQuorumJournalManagerUnit |
   |   | hadoop.hdfs.TestDFSUtil |
   |   | hadoop.hdfs.server.datanode.TestDirectoryScanner |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.43 ServerAPI=1.43 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/4/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/6183 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets xmllint |
   | uname | Linux 24b2798ef174 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 
13:30:12 UTC 2023 x

Re: [PR] HDFS-17223. Add journalnode maintenance node list [hadoop]

2023-11-09 Thread via GitHub


hadoop-yetus commented on PR #6183:
URL: https://github.com/apache/hadoop/pull/6183#issuecomment-1803910333

   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime |  Logfile | Comment |
   |::|--:|:|::|:---:|
   | +0 :ok: |  reexec  |   0m 48s |  |  Docker mode activated.  |
    _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  |  No case conflicting files 
found.  |
   | +0 :ok: |  codespell  |   0m  1s |  |  codespell was not available.  |
   | +0 :ok: |  detsecrets  |   0m  1s |  |  detect-secrets was not available.  
|
   | +0 :ok: |  xmllint  |   0m  1s |  |  xmllint was not available.  |
   | +1 :green_heart: |  @author  |   0m  0s |  |  The patch does not contain 
any @author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  |  The patch appears to 
include 3 new or modified test files.  |
    _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  48m  8s |  |  trunk passed  |
   | +1 :green_heart: |  compile  |   1m 25s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  compile  |   1m 14s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  checkstyle  |   1m 11s |  |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   1m 25s |  |  trunk passed  |
   | +1 :green_heart: |  javadoc  |   1m 13s |  |  trunk passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javadoc  |   1m 39s |  |  trunk passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   3m 38s |  |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  41m 51s |  |  branch has no errors 
when building and testing our client artifacts.  |
    _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   1m 13s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   1m 17s |  |  the patch passed with JDK 
Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04  |
   | +1 :green_heart: |  javac  |   1m 17s |  |  the patch passed  |
   | +1 :green_heart: |  compile  |   1m  7s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  javac  |   1m  7s |  |  the patch passed  |
   | +1 :green_heart: |  blanks  |   0m  0s |  |  The patch has no blanks 
issues.  |
   | -0 :warning: |  checkstyle  |   1m  2s | 
[/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/3/artifact/out/results-checkstyle-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs-project/hadoop-hdfs: The patch generated 11 new + 248 unchanged 
- 0 fixed = 259 total (was 248)  |
   | +1 :green_heart: |  mvnsite  |   1m 23s |  |  the patch passed  |
   | -1 :x: |  javadoc  |   0m 58s | 
[/results-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/3/artifact/out/results-javadoc-javadoc-hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04.txt)
 |  
hadoop-hdfs-project_hadoop-hdfs-jdkUbuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 
with JDK Ubuntu-11.0.20.1+1-post-Ubuntu-0ubuntu120.04 generated 1 new + 98 
unchanged - 1 fixed = 99 total (was 99)  |
   | +1 :green_heart: |  javadoc  |   1m 30s |  |  the patch passed with JDK 
Private Build-1.8.0_382-8u382-ga-1~20.04.1-b05  |
   | +1 :green_heart: |  spotbugs  |   3m 34s |  |  the patch passed  |
   | +1 :green_heart: |  shadedclient  |  40m 43s |  |  patch has no errors 
when building and testing our client artifacts.  |
    _ Other Tests _ |
   | -1 :x: |  unit  | 258m  9s | 
[/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt](https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/3/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt)
 |  hadoop-hdfs in the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 54s |  |  The patch does not 
generate ASF License warnings.  |
   |  |   | 414m 53s |  |  |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed junit tests | hadoop.hdfs.qjournal.client.TestQuorumJournalManager |
   |   | hadoop.hdfs.TestDFSUtil |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | ClientAPI=1.43 ServerAPI=1.43 base: 
https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-6183/3/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/6183 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient spotbugs checkstyle codespell detsecrets xmllint |
   | uname | Linux 1145ceb826bb 4.15.0-213-generic #224-Ubuntu SMP Mon Jun 19 
13:30:12 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | dev-support/bin/hadoop.sh |
   | git revision | trunk / 8cbb5a74495d6c9ad49384a1f682a91a122e8

Re: [PR] HDFS-17218. NameNode should process time out excess redundancy blocks [hadoop]

2023-11-09 Thread via GitHub


haiyang1987 commented on PR #6176:
URL: https://github.com/apache/hadoop/pull/6176#issuecomment-180365

   Test failures seems unrelated.
   
   However changes in the `contains` method and `remove` method input blockInfo 
and the equals `method` in ExcessBlockInfo will cause some SpotBugs problems.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18954) Filter NaN values from JMX json interface

2023-11-09 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17784333#comment-17784333
 ] 

ASF GitHub Bot commented on HADOOP-18954:
-

Hexiaoqiao commented on PR #6229:
URL: https://github.com/apache/hadoop/pull/6229#issuecomment-1803445430

   Committed to trunk. Thanks @K0K0V0K and @ferdelyi !




> Filter NaN values from JMX json interface
> -
>
> Key: HADOOP-18954
> URL: https://issues.apache.org/jira/browse/HADOOP-18954
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Bence Kosztolnik
>Assignee: Bence Kosztolnik
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> As we can see in this [Yarn 
> documentation|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html]
>  beans can represent Float values as NaN. These values will be represented in 
> the JMX response JSON like:
> {noformat}
> ...
> "GuaranteedCapacity": NaN,
> ...
> {noformat}
> Based on the [JSON doc|https://www.json.org/] NaN is not a valid JSON token ( 
> however some of the parser libs can handle it ), so not every consumer can 
> parse values like these.
> To be able to parse NaN values, a new feature flag should be created.
> The new feature will replace the NaN values with 0.0 values.
> The feature is default turned off. It can be enabled with the 
> *hadoop.http.jmx.nan-filter.enabled* config.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HADOOP-18954. Filter NaN values from JMX json interface [hadoop]

2023-11-09 Thread via GitHub


Hexiaoqiao commented on PR #6229:
URL: https://github.com/apache/hadoop/pull/6229#issuecomment-1803445430

   Committed to trunk. Thanks @K0K0V0K and @ferdelyi !


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18954) Filter NaN values from JMX json interface

2023-11-09 Thread Xiaoqiao He (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoqiao He updated HADOOP-18954:
-
Fix Version/s: 3.4.0
 Hadoop Flags: Reviewed
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

> Filter NaN values from JMX json interface
> -
>
> Key: HADOOP-18954
> URL: https://issues.apache.org/jira/browse/HADOOP-18954
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Bence Kosztolnik
>Assignee: Bence Kosztolnik
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> As we can see in this [Yarn 
> documentation|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html]
>  beans can represent Float values as NaN. These values will be represented in 
> the JMX response JSON like:
> {noformat}
> ...
> "GuaranteedCapacity": NaN,
> ...
> {noformat}
> Based on the [JSON doc|https://www.json.org/] NaN is not a valid JSON token ( 
> however some of the parser libs can handle it ), so not every consumer can 
> parse values like these.
> To be able to parse NaN values, a new feature flag should be created.
> The new feature will replace the NaN values with 0.0 values.
> The feature is default turned off. It can be enabled with the 
> *hadoop.http.jmx.nan-filter.enabled* config.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18954) Filter NaN values from JMX json interface

2023-11-09 Thread ASF GitHub Bot (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17784332#comment-17784332
 ] 

ASF GitHub Bot commented on HADOOP-18954:
-

Hexiaoqiao merged PR #6229:
URL: https://github.com/apache/hadoop/pull/6229




> Filter NaN values from JMX json interface
> -
>
> Key: HADOOP-18954
> URL: https://issues.apache.org/jira/browse/HADOOP-18954
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Bence Kosztolnik
>Assignee: Bence Kosztolnik
>Priority: Major
>  Labels: pull-request-available
>
> As we can see in this [Yarn 
> documentation|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html]
>  beans can represent Float values as NaN. These values will be represented in 
> the JMX response JSON like:
> {noformat}
> ...
> "GuaranteedCapacity": NaN,
> ...
> {noformat}
> Based on the [JSON doc|https://www.json.org/] NaN is not a valid JSON token ( 
> however some of the parser libs can handle it ), so not every consumer can 
> parse values like these.
> To be able to parse NaN values, a new feature flag should be created.
> The new feature will replace the NaN values with 0.0 values.
> The feature is default turned off. It can be enabled with the 
> *hadoop.http.jmx.nan-filter.enabled* config.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



Re: [PR] HADOOP-18954. Filter NaN values from JMX json interface [hadoop]

2023-11-09 Thread via GitHub


Hexiaoqiao merged PR #6229:
URL: https://github.com/apache/hadoop/pull/6229


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org