Your program may be trying to run locally on your windows box instead of talking to the ubuntu micro cluster.
On 10/13/07 11:43 AM, "dcave555" <[EMAIL PROTECTED]> wrote: > > Thx for your reply > > But i think i don't understand something. > i have installed hadoop single node on ubuntu. > my test is running on windows. > > i changed property in hadoop-site.xml in my application to > > <property> > <name>hadoop.tmp.dir</name> > <value>//172.16.50.13/usr/local/hadoop-datastore/hadoop-hadoop</value> > <description>A base for other temporary directories.</description> > </property> > > amd i am getting Login failed. > > If i want to save file from windows to hadoop on ubuntu i need to import ssh > public key? > how can i tell my test application on windows to use this pulic key? > > Or i need to do something else? > > Thx > > > > > > Ted Dunning-3 wrote: >> >> >> Hadoop is having trouble spawning a process to measure how much disk space >> you have available. >> >> It really is much easier to run hadoop on Linux. In case you can't do >> that, >> make sure that you really do have cygwin installed in a standard way on >> each >> hadoop node. >> >> >> On 10/13/07 8:05 AM, "dcave555" <[EMAIL PROTECTED]> wrote: >> >>> >>> Hello all >>> >>> I am new to hadoop . >>> >>> I am trying to write file to single cluster and getting this exception >>> when >>> i am trying to close output stream >>> >>> java.io.IOException: CreateProcess: df -k >>> C:\usr\local\hadoop-datastore\hadoop-hadoop\dfs\tmp error=2 >>> at java.lang.ProcessImpl.create(Native Method) >>> at java.lang.ProcessImpl.<init>(Unknown Source) >>> at java.lang.ProcessImpl.start(Unknown Source) >>> at java.lang.ProcessBuilder.start(Unknown Source) >>> at java.lang.Runtime.exec(Unknown Source) >>> at java.lang.Runtime.exec(Unknown Source) >>> at org.apache.hadoop.fs.DF.doDF(DF.java:60) >>> at org.apache.hadoop.fs.DF.<init>(DF.java:53) >>> at >>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.confChanged(Local >>> Di >>> rAllocator.java:198) >>> at >>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWr >>> it >>> e(LocalDirAllocator.java:235) >>> at >>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.createTmpFileForW >>> ri >>> te(LocalDirAllocator.java:276) >>> at >>> org.apache.hadoop.fs.LocalDirAllocator.createTmpFileForWrite(LocalDirAllocat >>> or >>> .java:155) >>> at >>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.newBackupFile(DFSClient.java >>> :1 >>> 475) >>> at >>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.openBackupStream(DFSClient.j >>> av >>> a:1442) >>> at >>> >> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.writeChunk(DFSClient.java:160 >> 0> >> ) >>> at >>> >> org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:14 >> 0> >> ) >>> at >>> org.apache.hadoop.fs.FSOutputSummer.flushBuffer(FSOutputSummer.java:122) >>> at >>> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1739) >>> at >>> org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStre >>> am >>> .java:49) >>> at >>> org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:64) >>> at Test1.main(Test1.java:23) >>> >>> >>> >>> My test is: >>> >>> Configuration configuration = new >>> Configuration(); >>> FileSystem fileSystem = >>> FileSystem.get(configuration); >>> Path path = new Path("/testfile"); >>> //writing: >>> FSDataOutputStream dataOutputStream = fileSystem.create(path); >>> dataOutputStream.writeUTF("hello world"); >>> dataOutputStream.close(); >>> //reading >>> FSDataInputStream dataInputStream = fileSystem.open(path); >>> System.out.println(dataInputStream.readUTF()); >>> dataInputStream.close(); >>> fileSystem.close(); >>> >>> i added hadoop-site.xml to classpath : >>> >>> >>> <configuration> >>> >>> <property> >>> <name>hadoop.tmp.dir</name> >>> <value>/usr/local/hadoop-datastore/hadoop-hadoop</value> >>> <description>A base for other temporary directories.</description> >>> </property> >>> >>> <property> >>> <name>fs.default.name</name> >>> <value>hdfs://172.16.50.13:54310</value> >>> <description>The name of the default file system. A URI whose >>> scheme and authority determine the FileSystem implementation. The >>> uri's scheme determines the config property (fs.SCHEME.impl) naming >>> the FileSystem implementation class. The uri's authority is used to >>> determine the host, port, etc. for a filesystem.</description> >>> </property> >>> >>> <property> >>> <name>mapred.job.tracker</name> >>> <value>172.16.50.13:54311</value> >>> <description>The host and port that the MapReduce job tracker runs >>> at. If "local", then jobs are run in-process as a single map >>> and reduce task. >>> </description> >>> </property> >>> >>> <property> >>> <name>dfs.replication</name> >>> <value>1</value> >>> <description>Default block replication. >>> The actual number of replications can be specified when the file is >>> created. >>> The default is used if replication is not specified in create time. >>> </description> >>> </property> >>> </configuration> >>> >>> >>> >>> Please help me >>> thx >> >> >>