Hi Susheel,

Tried that - same result. DataNode still not starting.

Thanks,
Charles

On 16 September 2014 11:49, Susheel Kumar Gadalay <skgada...@gmail.com>
wrote:

> The VERSION file has to be same across all the data nodes directories.
>
> So I suggested to copy it as it is using OS command and start data node.
>
> On 9/16/14, Charles Robertson <charles.robert...@gmail.com> wrote:
> > Hi Susheel,
> >
> > Thanks for the reply. I'm not entirely sure what you mean.
> >
> > When I created the new directory on the new volume I simply created an
> > empty directory. I see from the existing data node directory that it has
> a
> > sub-directory called current containing a file called VERSION.
> >
> > Your advice is to create the 'current' sub-directory and copy the VERSION
> > file across to it without changes? I see it has various guids, and so I'm
> > worried about it clashing with the VERSION file in the other data
> > directory.
> >
> > Thanks,
> > Charles
> >
> > On 16 September 2014 10:57, Susheel Kumar Gadalay <skgada...@gmail.com>
> > wrote:
> >
> >> Is it something to do current/VERSION file in data node directory.
> >>
> >> Just copy from the existing directory and start.
> >>
> >> On 9/16/14, Charles Robertson <charles.robert...@gmail.com> wrote:
> >> > Hi all,
> >> >
> >> > I am running out of space on a data node, so added a new volume to the
> >> > host, mounted it and made sure the permissions were set OK. Then I
> >> updated
> >> > the 'DataNode Directories' property in Ambari to include the new path
> >> > (comma separated, i.e. '/hadoop/hdfs/data,/data/hdfs'). Next I
> >> > restarted
> >> > the components with stale configs for that host, but the DataNode
> >> wouldn't
> >> > come back up, reporting 'connection refused'. When I remove the new
> >> > data
> >> > directory path from the property and restart, it starts fine.
> >> >
> >> > What am I doing wrong?
> >> >
> >> > Thanks,
> >> > Charles
> >> >
> >>
> >
>

Reply via email to