Hi Ayush,

Thanks for your reply, yes we followed the same document and applied the 
config. For our setup, we have configured HA and its working fine. But, we are 
getting the following error while executing the the commands.

-ls: java.net.UnknownHostException: hdfs
-mkdir: java.net.UnknownHostException: hdfs


Regards,
Kamaraj
InfoSight - SRE

From: Ayush Saxena<mailto:ayush...@gmail.com>
Sent: Sunday, August 23, 2020 7:17 PM
To: Muthupandiyan, Kamaraj<mailto:kamaraj.muthupandi...@hpe.com>
Cc: Mingliang Liu<mailto:lium...@gmail.com>; 
user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: Re: Error while running "hdfs fs" commands

Hi Kamaraj,
If you are trying to setup an HA cluster, you need couple of more configs as 
well, You can follow this document, this should answer all your doubts :
https://hadoop.apache.org/docs/r3.3.0/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html#Deployment<https://hadoop.apache.org/docs/r3.3.0/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html#Deployment>

-Ayush

On Sun, 23 Aug 2020 at 18:53, Muthupandiyan, Kamaraj 
<kamaraj.muthupandi...@hpe.com<mailto:kamaraj.muthupandi...@hpe.com>> wrote:
Hi Mingliang Liu,

Thank you for your reply.

As I am completely new to setting up the Hadoop cluster I am not sure about the 
value of  fs.defaultFS. I have set the below one as the value, Could you please 
help me with the correct value.

<property>
        <name>fs.defaultFS</name>
        <value>hdfs://hdfs</value>
    </property>


As a additional information, we have two name nodes in our cluster.

Regards,
Kamaraj
InfoSight - SRE

From: Mingliang Liu<mailto:lium...@gmail.com>
Sent: Friday, August 21, 2020 11:14 PM
To: Muthupandiyan, Kamaraj<mailto:kamaraj.muthupandi...@hpe.com>
Cc: user@hadoop.apache.org<mailto:user@hadoop.apache.org>
Subject: Re: Error while running "hdfs fs" commands

Seems that you have a wrong fs.defaultFS configuration in your core-site.xml

For the second command you should follow the user manual. The input is simply 
not valid. Try something like:

hadoop fs -ls 
hdfs://sym-hdfsnn1.lvs.nimblestorage.com/user<http://sym-hdfsnn1.lvs.nimblestorage.com/user>
or
hadoop fs -ls file:///tmp

On Fri, Aug 21, 2020 at 9:07 AM Muthupandiyan, Kamaraj 
<kamaraj.muthupandi...@hpe.com<mailto:kamaraj.muthupandi...@hpe.com>> wrote:
Hi Team,

Whenever I am executing the hdfs fs command and I am getting below error, 
please help me with this.

[hdfs@hdfsnn1 ~]$ hdfs dfs -ls /user
-ls: java.net.UnknownHostException: hdfs


[hdfs@sym-hdfsnn1 ~]$ hadoop fs -ls / file:///%7Chdfs:hdfsnn1:8020/
-bash: 
hdfs:sym-hdfsnn1.lvs.nimblestorage.com:8020/<http://sym-hdfsnn1.lvs.nimblestorage.com:8020/>:
 No such file or directory
-ls: java.net.UnknownHostException: hdfs
Usage: hadoop fs [generic options]
        [-appendToFile <localsrc> ... <dst>]
        [-cat [-ignoreCrc] <src> ...]
        [-checksum <src> ...]
        [-chgrp [-R] GROUP PATH...]
        [-chmod [-R] <MODE[,MODE]... | OCTALMODE> PATH...]
        [-chown [-R] [OWNER][:[GROUP]] PATH...]
        [-copyFromLocal [-f] [-p] [-l] [-d] <localsrc> ... <dst>]
        [-copyToLocal [-f] [-p] [-ignoreCrc] [-crc] <src> ... <localdst>]
        [-count [-q] [-h] [-v] [-t [<storage type>]] [-u] [-x] <path> ...]
        [-cp [-f] [-p | -p[topax]] [-d] <src> ... <dst>]
        [-createSnapshot <snapshotDir> [<snapshotName>]]
        [-deleteSnapshot <snapshotDir> <snapshotName>]
        [-df [-h] [<path> ...]]
        [-du [-s] [-h] [-x] <path> ...]
        [-expunge]
        [-find <path> ... <expression> ...]
        [-get [-f] [-p] [-ignoreCrc] [-crc] <src> ... <localdst>]
        [-getfacl [-R] <path>]
        [-getfattr [-R] {-n name | -d} [-e en] <path>]
        [-getmerge [-nl] [-skip-empty-file] <src> <localdst>]
        [-help [cmd ...]]
        [-ls [-C] [-d] [-h] [-q] [-R] [-t] [-S] [-r] [-u] [<path> ...]]
        [-mkdir [-p] <path> ...]
        [-moveFromLocal <localsrc> ... <dst>]



Regards,
Kamaraj
InfoSight - SRE



--
L


Reply via email to