[ 
https://issues.apache.org/jira/browse/HDFS-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17409322#comment-17409322
 ] 

Ayush Saxena commented on HDFS-15528:
-------------------------------------

CryptoAdmin isn't supported with ViewFs and is being tracked at HDFS-14178

> Not able to list encryption zone with federation
> ------------------------------------------------
>
>                 Key: HDFS-15528
>                 URL: https://issues.apache.org/jira/browse/HDFS-15528
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: encryption, federation
>    Affects Versions: 3.0.0
>            Reporter: Thangamani Murugasamy
>            Priority: Major
>
>  hdfs crypto -listZones
> IllegalArgumentException: 'viewfs://cluster14' is not an HDFS URI.
>  
> ------
> debug log
> 20/08/12 05:53:14 DEBUG util.Shell: setsid exited with exit code 0
> 20/08/12 05:53:14 DEBUG lib.MutableMetricsFactory: field 
> org.apache.hadoop.metrics2.lib.MutableRate 
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with 
> annotation @org.apache.hadoop.metrics2.annotation.Metric(always=false, 
> sampleName=Ops, about=, type=DEFAULT, valueName=Time, value=[Rate of 
> successful kerberos logins and latency (milliseconds)])
> 20/08/12 05:53:14 DEBUG lib.MutableMetricsFactory: field 
> org.apache.hadoop.metrics2.lib.MutableRate 
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with 
> annotation @org.apache.hadoop.metrics2.annotation.Metric(always=false, 
> sampleName=Ops, about=, type=DEFAULT, valueName=Time, value=[Rate of failed 
> kerberos logins and latency (milliseconds)])
> 20/08/12 05:53:14 DEBUG lib.MutableMetricsFactory: field 
> org.apache.hadoop.metrics2.lib.MutableRate 
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with 
> annotation @org.apache.hadoop.metrics2.annotation.Metric(always=false, 
> sampleName=Ops, about=, type=DEFAULT, valueName=Time, value=[GetGroups])
> 20/08/12 05:53:14 DEBUG lib.MutableMetricsFactory: field private 
> org.apache.hadoop.metrics2.lib.MutableGaugeLong 
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.renewalFailuresTotal
>  with annotation @org.apache.hadoop.metrics2.annotation.Metric(always=false, 
> sampleName=Ops, about=, type=DEFAULT, valueName=Time, value=[Renewal failures 
> since startup])
> 20/08/12 05:53:14 DEBUG lib.MutableMetricsFactory: field private 
> org.apache.hadoop.metrics2.lib.MutableGaugeInt 
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.renewalFailures 
> with annotation @org.apache.hadoop.metrics2.annotation.Metric(always=false, 
> sampleName=Ops, about=, type=DEFAULT, valueName=Time, value=[Renewal failures 
> since last successful login])
> 20/08/12 05:53:14 DEBUG impl.MetricsSystemImpl: UgiMetrics, User and group 
> related metrics
> 20/08/12 05:53:14 DEBUG security.SecurityUtil: Setting 
> hadoop.security.token.service.use_ip to true
> 20/08/12 05:53:14 DEBUG security.Groups: Creating new Groups object
> 20/08/12 05:53:14 DEBUG security.Groups: Group mapping 
> impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping; 
> cacheTimeout=300000; warningDeltaMs=5000
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: hadoop login
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: hadoop login commit
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: using kerberos 
> user:h...@corp.epsilon.com
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: Using user: 
> "h...@corp.epsilon.com" with name h...@corp.epsilon.com
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: User entry: 
> "h...@corp.epsilon.com"
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: UGI 
> loginUser:h...@corp.epsilon.com (auth:KERBEROS)
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: Current time is 
> 1597233194735
> 20/08/12 05:53:14 DEBUG security.UserGroupInformation: Next refresh is 
> 1597261977000
> 20/08/12 05:53:14 DEBUG core.Tracer: sampler.classes = ; loaded no samplers
> 20/08/12 05:53:14 DEBUG core.Tracer: span.receiver.classes = ; loaded no span 
> receivers
> 20/08/12 05:53:14 DEBUG fs.FileSystem: Loading filesystems
> 20/08/12 05:53:14 DEBUG fs.FileSystem: s3n:// = class 
> org.apache.hadoop.fs.s3native.NativeS3FileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-aws-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: gs:// = class 
> com.google.cloud.hadoop.fs.gcs.GoogleHadoopFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/lib/hadoop/gcs-connector-hadoop3-1.9.10-cdh6.2.1-shaded.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: file:// = class 
> org.apache.hadoop.fs.LocalFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-common-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: viewfs:// = class 
> org.apache.hadoop.fs.viewfs.ViewFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-common-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: ftp:// = class 
> org.apache.hadoop.fs.ftp.FTPFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-common-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: har:// = class 
> org.apache.hadoop.fs.HarFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-common-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: http:// = class 
> org.apache.hadoop.fs.http.HttpFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-common-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: https:// = class 
> org.apache.hadoop.fs.http.HttpsFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-common-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: hdfs:// = class 
> org.apache.hadoop.hdfs.DistributedFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-hdfs-client-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: webhdfs:// = class 
> org.apache.hadoop.hdfs.web.WebHdfsFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-hdfs-client-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: swebhdfs:// = class 
> org.apache.hadoop.hdfs.web.SWebHdfsFileSystem from 
> /opt/cloudera/parcels/CDH-6.2.1-1.cdh6.2.1.p0.1425774/jars/hadoop-hdfs-client-3.0.0-cdh6.2.1.jar
> 20/08/12 05:53:14 DEBUG fs.FileSystem: Looking for FS supporting viewfs
> 20/08/12 05:53:14 DEBUG fs.FileSystem: looking for configuration option 
> fs.viewfs.impl
> 20/08/12 05:53:14 DEBUG fs.FileSystem: Looking in service filesystems for 
> implementation class
> 20/08/12 05:53:14 DEBUG fs.FileSystem: FS for viewfs is class 
> org.apache.hadoop.fs.viewfs.ViewFileSystem
> 20/08/12 05:53:14 DEBUG fs.FileSystem: Looking for FS supporting hdfs
> 20/08/12 05:53:14 DEBUG fs.FileSystem: looking for configuration option 
> fs.hdfs.impl
> 20/08/12 05:53:14 DEBUG fs.FileSystem: Looking in service filesystems for 
> implementation class
> 20/08/12 05:53:14 DEBUG fs.FileSystem: FS for hdfs is class 
> org.apache.hadoop.hdfs.DistributedFileSystem
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: 
> dfs.client.use.legacy.blockreader.local = false
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: dfs.client.read.shortcircuit = 
> false
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: 
> dfs.client.domain.socket.data.traffic = false
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: dfs.domain.socket.path = 
> /var/run/hdfs-sockets/dn
> 20/08/12 05:53:14 DEBUG hdfs.DFSClient: Sets 
> dfs.client.block.write.replace-datanode-on-failure.min-replication to 0
> 20/08/12 05:53:14 DEBUG hdfs.HAUtilClient: No HA service delegation token 
> found for logical URI hdfs://pii/user
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: 
> dfs.client.use.legacy.blockreader.local = false
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: dfs.client.read.shortcircuit = 
> false
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: 
> dfs.client.domain.socket.data.traffic = false
> 20/08/12 05:53:14 DEBUG impl.DfsClientConf: dfs.domain.socket.path = 
> /var/run/hdfs-sockets/dn
> 20/08/12 05:53:14 DEBUG retry.RetryUtils: multipleLinearRandomRetry = null
> 20/08/12 05:53:14 DEBUG ipc.Server: rpcKind=RPC_PROTOCOL_BUFFER, 
> rpcRequestWrapperClass=class 
> org.apache.hadoop.ipc.ProtobufRpcEngine$RpcProtobufRequest, 
> rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker@291b4bf5
> 20/08/12 05:53:14 DEBUG ipc.Client: ipc.client.bind.wildcard.addr set to 
> true. Will bind client sockets to wildcard address.
> 20/08/12 05:53:14 DEBUG ipc.Client: getting client out of cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG util.NativeCodeLoader: Trying to load the 
> custom-built native-hadoop library...
> 20/08/12 05:53:15 DEBUG util.NativeCodeLoader: Loaded the native-hadoop 
> library
> 20/08/12 05:53:15 DEBUG unix.DomainSocketWatcher: 
> org.apache.hadoop.net.unix.DomainSocketWatcher$2@3310b42: starting with 
> interruptCheckPeriodMs = 60000
> 20/08/12 05:53:15 DEBUG util.PerformanceAdvisory: Both short-circuit local 
> reads and UNIX domain socket are disabled.
> 20/08/12 05:53:15 DEBUG sasl.DataTransferSaslUtil: DataTransferProtocol not 
> using SaslPropertiesResolver, no QOP found in configuration for 
> dfs.data.transfer.protection
> 20/08/12 05:53:15 DEBUG fs.FileSystem: Looking for FS supporting hdfs
> 20/08/12 05:53:15 DEBUG fs.FileSystem: looking for configuration option 
> fs.hdfs.impl
> 20/08/12 05:53:15 DEBUG fs.FileSystem: Looking in service filesystems for 
> implementation class
> 20/08/12 05:53:15 DEBUG fs.FileSystem: FS for hdfs is class 
> org.apache.hadoop.hdfs.DistributedFileSystem
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.use.legacy.blockreader.local = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.client.read.shortcircuit = 
> false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.domain.socket.data.traffic = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.domain.socket.path = 
> /var/run/hdfs-sockets/dn
> 20/08/12 05:53:15 DEBUG hdfs.DFSClient: Sets 
> dfs.client.block.write.replace-datanode-on-failure.min-replication to 0
> 20/08/12 05:53:15 DEBUG hdfs.HAUtilClient: No HA service delegation token 
> found for logical URI hdfs://attribute/attribute
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.use.legacy.blockreader.local = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.client.read.shortcircuit = 
> false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.domain.socket.data.traffic = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.domain.socket.path = 
> /var/run/hdfs-sockets/dn
> 20/08/12 05:53:15 DEBUG retry.RetryUtils: multipleLinearRandomRetry = null
> 20/08/12 05:53:15 DEBUG ipc.Client: getting client out of cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG sasl.DataTransferSaslUtil: DataTransferProtocol not 
> using SaslPropertiesResolver, no QOP found in configuration for 
> dfs.data.transfer.protection
> 20/08/12 05:53:15 DEBUG fs.FileSystem: Looking for FS supporting hdfs
> 20/08/12 05:53:15 DEBUG fs.FileSystem: looking for configuration option 
> fs.hdfs.impl
> 20/08/12 05:53:15 DEBUG fs.FileSystem: Looking in service filesystems for 
> implementation class
> 20/08/12 05:53:15 DEBUG fs.FileSystem: FS for hdfs is class 
> org.apache.hadoop.hdfs.DistributedFileSystem
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.use.legacy.blockreader.local = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.client.read.shortcircuit = 
> false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.domain.socket.data.traffic = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.domain.socket.path = 
> /var/run/hdfs-sockets/dn
> 20/08/12 05:53:15 DEBUG hdfs.DFSClient: Sets 
> dfs.client.block.write.replace-datanode-on-failure.min-replication to 0
> 20/08/12 05:53:15 DEBUG hdfs.HAUtilClient: No HA service delegation token 
> found for logical URI hdfs://mapping/
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.use.legacy.blockreader.local = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.client.read.shortcircuit = 
> false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: 
> dfs.client.domain.socket.data.traffic = false
> 20/08/12 05:53:15 DEBUG impl.DfsClientConf: dfs.domain.socket.path = 
> /var/run/hdfs-sockets/dn
> 20/08/12 05:53:15 DEBUG retry.RetryUtils: multipleLinearRandomRetry = null
> 20/08/12 05:53:15 DEBUG ipc.Client: getting client out of cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG sasl.DataTransferSaslUtil: DataTransferProtocol not 
> using SaslPropertiesResolver, no QOP found in configuration for 
> dfs.data.transfer.protection
> IllegalArgumentException: 'viewfs://cluster14' is not an HDFS URI.
> 20/08/12 05:53:15 DEBUG ipc.Client: stopping client from cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG ipc.Client: stopping client from cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG ipc.Client: stopping client from cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG ipc.Client: removing client from cache: 
> org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG ipc.Client: stopping actual client because no more 
> references remain: org.apache.hadoop.ipc.Client@1ebd319f
> 20/08/12 05:53:15 DEBUG ipc.Client: Stopping client
> 20/08/12 05:53:15 DEBUG util.ShutdownHookManager: Completed shutdown in 0.003 
> seconds; Timeouts: 0
> 20/08/12 05:53:15 DEBUG util.ShutdownHookManager: ShutdownHookManger 
> completed shutdown.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to