> Incorrect configuration: namenode address dfs.namenode.servicerpc-address or 
> dfs.namenode.rpc-address is not configured.
> Starting namenodes on []

NameNode/DataNode are part of a HDFS service. It makes no sense to try
and run them over an S3 URL default, which is a distributed filesystem
in itself. The services need fs.defaultFS to be set to a HDFS URI to
be able to start up.

> but unable to get an s3 config started via hadoop

You can run jobs over S3 input and output data by running a regular MR
cluster on HDFS - just pass the right URI as input and output
parameters of the job. Set your S3 properties in core-site.xml but let
the fs.defaultFS be of HDFS type, to do this.

> There is an s3.impl until 1.2.1 release. So does the 2.5.0 release support s3 
> or do i need to do anything else.

In Apache Hadoop 2 we dynamically load the FS classes, so we do not
need the fs.NAME.impl configs anymore as we did in Apache Hadoop 1.

On Wed, Sep 10, 2014 at 1:15 PM, Dhiraj <jar...@gmail.com> wrote:
> Hi,
>
> I have downloaded hadoop-2.5.0 and am trying to get it working for s3
> backend (single-node in a pseudo-distributed mode).
> I have made changes to the core-site.xml according to
> https://wiki.apache.org/hadoop/AmazonS3
>
> I have an backend object store running on my machine that supports S3.
>
> I get the following message when i try to start the daemons
> Incorrect configuration: namenode address dfs.namenode.servicerpc-address or
> dfs.namenode.rpc-address is not configured.
>
>
> root@ubuntu:/build/hadoop/hadoop-2.5.0# ./sbin/start-dfs.sh
> Incorrect configuration: namenode address dfs.namenode.servicerpc-address or
> dfs.namenode.rpc-address is not configured.
> Starting namenodes on []
> localhost: starting namenode, logging to
> /build/hadoop/hadoop-2.5.0/logs/hadoop-root-namenode-ubuntu.out
> localhost: starting datanode, logging to
> /build/hadoop/hadoop-2.5.0/logs/hadoop-root-datanode-ubuntu.out
> Starting secondary namenodes [0.0.0.0]
> 0.0.0.0: starting secondarynamenode, logging to
> /build/hadoop/hadoop-2.5.0/logs/hadoop-root-secondarynamenode-ubuntu.out
> root@ubuntu:/build/hadoop/hadoop-2.5.0#
>
> The deamons dont start after the above.
> i get the same error if i add the property "fs.defaultFS" and set its value
> to the s3 bucket but if i change the defaultFS to hdfs:// it works fine - am
> able to launch the daemons.
>
> my core-site.xml:
> <configuration>
>     <property>
>         <name>fs.defaultFS</name>
>         <value>s3://bucket1</value>
>     </property>
>     <property>
>         <name>fs.s3.awsAccessKeyId</name>
>         <value>abcd</value>
>     </property>
>     <property>
>         <name>fs.s3.awsSecretAccessKey</name>
>         <value>1234</value>
>     </property>
> </configuration>
>
>
> I am able to list the buckets and its contents via s3cmd and boto; but
> unable to get an s3 config started via hadoop
>
> Also from the following core-file.xml listed on the website; i dont see an
> implementation for s3
> http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/core-default.xml
>
> There is an s3.impl until 1.2.1 release. So does the 2.5.0 release support
> s3 or do i need to do anything else.
>
> cheers,
> Dhiraj
>
>
>



-- 
Harsh J

Reply via email to