Hi,
The following "segmentation fault" still exists.
I re wrote my application to use ant, but when I integrate it with libhdfs
it fails saying "segmentation fault" and "exiting with 139".
Please do help, as I have already spent a lot of time on re writing my
application to use hadoop and this one fails for no reason.

On Wed, Mar 19, 2008 at 12:52 PM, Raghavendra K <[EMAIL PROTECTED]>
wrote:

> I am passing the following arguments
>
> OS_NAME=Linux
> OS_ARCH=i386
> LIBHDFS_BUILD_DIR=/garl/garl-alpha1/home1/raghu/Desktop/hadoop-0.15.3
> /build/libhdfs
> JAVA_HOME=/garl/garl-alpha1/home1/raghu/Desktop/jdk1.5.0_14
> PLATFORM=linux
> SHLIB_VERSION=1
>
> I have commented out the line
> #PLATFORM = $(shell echo $$OS_NAME | tr [A-Z] [a-z])
> and passing PLATFORM=linux
> as the line was not executing if I just type
> make test
> separately.....
>  and also changed the line
>
> $(HDFS_TEST): hdfs_test.c
>         $(CC) $(CPPFLAGS) $< -L$(LIBHDFS_BUILD_DIR) -l$(LIB_NAME)
> $(LDFLAGS) -o $@
>
> (have added LDFLAGS, because when run it was complaining that ljvm.so was
> not found)....
>
> Where am I going wrong? Kindly let me know if I have to provide any other
> information.
>
>
> On Tue, Mar 18, 2008 at 11:41 PM, Arun C Murthy <[EMAIL PROTECTED]>
> wrote:
>
> >
> > On Mar 14, 2008, at 11:48 PM, Raghavendra K wrote:
> >
> > > Hi,
> > >   My apologies for bugging the forum again and again.
> > > I am able to get the sample program for libhdfs working. I followed
> > > these
> > > steps.
> > >
> > > ---> compiled using ant
> > > ---> modified the test-libhdfs.sh to include CLASSPATH, HADOOP_HOME,
> > > HADOOP_CONF_DIR, HADOOP_LOG_DIR, LIBHDFS_BUILD_DIR (since I ran
> > > test-libhdfs.sh individually and dint invoke it from ant)
> > > ---> The program ran succesfully and was able to write, read and all.
> > >
> > > Now I copy the same program to a different directory and use the same
> > > Makefile(used by ant) and modified the variables accordingly. Used
> > > make test
> > > compiled successfully
> > > Used the same test-libhdfs.sh to invoke hdfs_test, but now it fails
> > > saying
> > > Segmentation Fault.
> > > I dont know where it is going wrong.
> > > Cant libhdfs be compiled without using ant? I want to test it and
> > > integrate
> > > libhdfs with my program
> > > Please do reply and help me out as this is driving me crazy.
> >
> > I can only assume there is something wrong with the values you are
> > passing for the requisite environment variables: OS_{NAME|OS_ARCH},
> > SHLIB_VERSION, LIBHDFS_VERSION, HADOOP_{HOME|CONF_DIR|LOG_DIR} since
> > it works when you run 'make test'.
> >
> > Sorry it isn't of much help... could you share the values you are
> > using for these?
> >
> > Arun
> >
> >
> > > Thanks in advance.
> > >
> > > --
> > > Regards,
> > > Raghavendra K
> >
> >
>
>
> --
> Regards,
> Raghavendra K




-- 
Regards,
Raghavendra K

Reply via email to