On May 21, 2013, at 4:00 PM, Andrei Savu <[email protected]> wrote:

> You need sane dns settings (forward and reverse for each machine to make this 
> work). 
> 

Can I try to hack configure_hostname.sh in:

services/cdh/target/classes/functions

Adding some entry in /etc/hosts

Will that be enough ?


> -- Andrei Savu
> 
> On Tue, May 21, 2013 at 10:57 PM, Sebastien Goasguen <[email protected]> wrote:
> 
> On May 21, 2013, at 3:48 PM, Andrew Bayer <[email protected]> wrote:
> 
>> Yeah, DNS is a giant pain. If at all possible, you need to get the hostnames 
>> resolvable from wherever you're spinning the instances up, as well as on the 
>> instances themselves. The DNS that CloudStack's DHCP assigns should do the 
>> trick for that.
> 
> argh…
> 
> These instances have public IPs but not DNS entries.
> 
> @andrei the hadoop-3d5 and other names are setup as the name of the 
> instances. They are used for local 'hostname'. so no not resolvable.
> 
> 
> 
>> 
>> A.
>> 
>> On Tue, May 21, 2013 at 12:46 PM, Sebastien Goasguen <[email protected]> 
>> wrote:
>> Hi,
>> 
>> I installed whirr 0.8.1, I am using it against a CloudStack endpoint.
>> Instances get launched and I am trying to setup cdh.
>> 
>> I believe I am running into a DNS issue as I am running into lots of issues 
>> of this type:
>> 
>> 13/05/21 21:21:28 WARN net.DNS: Unable to determine local hostname -falling 
>> back to "localhost"
>> java.net.UnknownHostException: hadoop-3d5: hadoop-3d5
>> 
>> If I log in to the name node and try to use hadoop I get things like:
>> 
>> $ hadoop fs -mkdir /toto
>> -mkdir: java.net.UnknownHostException: hadoop-3d5
>> 
>> my hadoop-site.xml looks like:
>> 
>> <?xml version="1.0"?>
>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>> <configuration>
>>  <property>
>>    <name>dfs.client.use.legacy.blockreader</name>
>>    <value>true</value>
>>  </property>
>>  <property>
>>    <name>fs.default.name</name>
>>    <value>hdfs://hadoop-3d5:8020/</value>
>>  </property>
>>  <property>
>>    <name>mapred.job.tracker</name>
>>    <value>hadoop-3d5:8021</value>
>>  </property>
>>  <property>
>>    <name>hadoop.job.ugi</name>
>>    <value>root,root</value>
>>  </property>
>>  <property>
>>    <name>hadoop.rpc.socket.factory.class.default</name>
>>    <value>org.apache.hadoop.net.SocksSocketFactory</value>
>>  </property>
>>  <property>
>>    <name>hadoop.socks.server</name>
>>    <value>localhost:6666</value>
>>  </property>
>> </configuration>
>> 
>> my ~/.whirr/hadoop/instances file has all the right IP addresses, but I 
>> don't think the security group rules got created.
>> 
>> Any thoughts ?
>> 
>> thanks,
>> 
>> -sebastien
>> 
>> 
> 
> 

Reply via email to