[ https://issues.apache.org/jira/browse/WHIRR-413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13149429#comment-13149429 ]
Jongwook Woo commented on WHIRR-413: ------------------------------------ 1) core-site.xml jongwook@domU-12-31-39-12-91-A6:/tmp$ more core-site.xml <configuration> <property> <name>hadoop.tmp.dir</name> <value>/data/tmp/hadoop-${user.name}</value> </property> <property> <name>io.file.buffer.size</name> <value>65536</value> </property> <property> <name>hadoop.rpc.socket.factory.class.default</name> <value>org.apache.hadoop.net.StandardSocketFactory</value> <final>true</final> </property> <property> <name>hadoop.rpc.socket.factory.class.ClientProtocol</name> <value></value> </property> <property> <name>hadoop.rpc.socket.factory.class.JobSubmissionProtocol</name> <value></value> </property> <property> <name>fs.trash.interval</name> <value>1440</value> </property> <property> <name>fs.default.name</name> <value>hdfs://ec2-184-72-193-67.compute-1.amazonaws.com:8020/</value> </property> </configuration> 2) hbase-site.xml jongwook@domU-12-31-39-12-91-A6:/tmp$ more hbase-site.xml <configuration> <property> <name>hbase.cluster.distributed</name> <value>true</value> </property> <property> <name>hbase.regionserver.handler.count</name> <value>100</value> </property> <property> <name>dfs.replication</name> <value>3</value> </property> <property> <name>zookeeper.session.timeout</name> <value>60000</value> </property> <property> <name>hbase.tmp.dir</name> <value>/data/tmp/hbase-${user.name}</value> </property> <property> <name>hbase.client.retries.number</name> <value>100</value> </property> <property> <name>hbase.zookeeper.recoverable.waittime</name> <value>600000</value> </property> <property> <name>hbase.rootdir</name> <value>hdfs://ec2-184-72-193-67.compute-1.amazonaws.com:8020/hbase</value> </property> <property> <name>hbase.zookeeper.quorum</name> <value>ec2-184-72-193-67.compute-1.amazonaws.com:2181</value> </property> </configuration> 3) hdfs-site.xml jongwook@domU-12-31-39-12-91-A6:/tmp$ more hdfs-site.xml <configuration> <property> <name>dfs.block.size</name> <value>134217728</value> </property> <property> <name>dfs.data.dir</name> <value>/data/hadoop/hdfs/data</value> </property> <property> <name>dfs.datanode.du.reserved</name> <value>1073741824</value> </property> <property> <name>dfs.name.dir</name> <value>/data/hadoop/hdfs/name</value> </property> <property> <name>fs.checkpoint.dir</name> <value>/data/hadoop/hdfs/secondary</value> </property> </configuration> 4) mapred-site.xml jongwook@domU-12-31-39-12-91-A6:/tmp$ more mapred-site.xml <configuration> <property> <name>mapred.local.dir</name> <value>/data/hadoop/mapred/local</value> </property> <property> <name>mapred.map.tasks.speculative.execution</name> <value>true</value> </property> <property> <name>mapred.reduce.tasks.speculative.execution</name> <value>false</value> </property> <property> <name>mapred.system.dir</name> <value>/hadoop/system/mapred</value> </property> <property> <name>mapreduce.jobtracker.staging.root.dir</name> <value>/user</value> </property> <property> <name>mapred.compress.map.output</name> <value>true</value> </property> <property> <name>mapred.output.compression.type</name> <value>BLOCK</value> </property> <property> <name>mapred.child.java.opts</name> <value>-Xmx550m</value> </property> <property> <name>mapred.child.ulimit</name> <value>1126400</value> </property> <property> <name>mapred.tasktracker.map.tasks.maximum</name> <value>2</value> </property> <property> <name>mapred.tasktracker.reduce.tasks.maximum</name> <value>2</value> </property> <property> <name>mapred.reduce.tasks</name> <value>10</value> </property> <property> <name>mapred.job.tracker</name> <value>ec2-184-72-193-67.compute-1.amazonaws.com:8021</value> </property> </configuration> > jobcache file is stored at /tmp/ folder so that it has out of storage error > --------------------------------------------------------------------------- > > Key: WHIRR-413 > URL: https://issues.apache.org/jira/browse/WHIRR-413 > Project: Whirr > Issue Type: Bug > Components: build, service/hadoop > Affects Versions: 0.6.0, 0.7.0 > Environment: - Ubuntu-11.10 > - java version "1.6.0_23" > OpenJDK Runtime Environment (IcedTea6 1.11pre) (6b23~pre10-0ubuntu5) > OpenJDK Client VM (build 20.0-b11, mixed mode, sharing) > - ruby 1.8.7 (2011-06-30 patchlevel 352) [i686-linux] > - Apache Maven 3.0.3 (r1075438; 2011-02-28 09:31:09-0800) > Maven home: /home/jongwook/apache/apache-maven-3.0.3 > Java version: 1.6.0_23, vendor: Sun Microsystems Inc. > Java home: /usr/lib/jvm/java-6-openjdk/jre > Default locale: en_US, platform encoding: UTF-8 > OS name: "linux", version: "3.0.0-12-generic", arch: "i386", family: "unix" > Reporter: Jongwook Woo > Priority: Critical > Labels: build > Fix For: 0.6.0, 0.7.0 > > Original Estimate: 48h > Remaining Estimate: 48h > > when I run Hadoop to read/write data from/to HBase, I got the following error > because of the less storage space at /tmp/. > I guess whirr is supposed to use /data/tmp/ to store jobcache file such as > taskTracker/jobcache/job_local_0001/attempt_local_0001_m_0000xx_0/output/file.out > because /data/tmp/ has 335GB. However, it is stored at /tmp/ that has only > 9.9G. Thus, some configuration xml file seems not correct. It generates > errors both at 0.6.0 and 0.7.0 > -----Storage space check --------------------------------------- > jongwook@ip-10-245-174-15:/tmp/hadoop-jongwook/mapred/local/taskTracker/jobcache/job_local_0001$ > cd /tmp > jongwook@ip-10-245-174-15:/tmp$ df -h . > Filesystem Size Used Avail Use% Mounted on > /dev/sda1 9.9G 9.1G 274M 98% / > jongwook@ip-10-245-174-15:/tmp$ df -h > Filesystem Size Used Avail Use% Mounted on > /dev/sda1 9.9G 9.1G 274M 98% / > none 846M 116K 846M 1% /dev > none 879M 0 879M 0% /dev/shm > none 879M 68K 878M 1% /var/run > none 879M 0 879M 0% /var/lock > none 879M 0 879M 0% /lib/init/rw > /dev/sda2 335G 199M 318G 1% /mnt > -----Error msg at the end of hadoop/hbase code > ------------------------------------------------------- > 11/10/27 03:33:09 INFO mapred.MapTask: Finished spill 61 > 11/10/27 03:33:09 WARN mapred.LocalJobRunner: job_local_0001 > org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any > valid local directory for > taskTracker/jobcache/job_local_0001/attempt_local_0001_m_000016_0/output/file.out > at > org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:343) > at > org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124) > at > org.apache.hadoop.mapred.MapOutputFile.getOutputFileForWrite(MapOutputFile.java:61) > at > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.mergeParts(MapTask.java:1469) > at > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:1154) > at > org.apache.hadoop.mapred.MapTask$NewOutputCollector.close(MapTask.java:549) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:623) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:305) > at > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:177) > 11/10/27 03:33:09 INFO mapred.JobClient: Job complete: job_local_0001 > 11/10/27 03:33:09 INFO mapred.JobClient: Counters: 8 > 11/10/27 03:33:09 INFO mapred.JobClient: FileSystemCounters > 11/10/27 03:33:09 INFO mapred.JobClient: FILE_BYTES_READ=103074405254 > 11/10/27 03:33:09 INFO mapred.JobClient: FILE_BYTES_WRITTEN=156390149579 > 11/10/27 03:33:09 INFO mapred.JobClient: Map-Reduce Framework > 11/10/27 03:33:09 INFO mapred.JobClient: Combine output records=0 > 11/10/27 03:33:09 INFO mapred.JobClient: Map input records=13248198 > 11/10/27 03:33:09 INFO mapred.JobClient: Spilled Records=788109966 > 11/10/27 03:33:09 INFO mapred.JobClient: Map output bytes=5347057080 > 11/10/27 03:33:09 INFO mapred.JobClient: Combine input records=0 > 11/10/27 03:33:09 INFO mapred.JobClient: Map output records=278212138 > It takes: 1966141 msec > 11/10/27 03:33:10 INFO zookeeper.ZooKeeper: Session: 0x13341a966cb000d closed -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira