Sebastian You've made a minor typo in the configuration setting which means you haven't actually enabled out of core graph mode.
You have giraph.useOutOfCoreGiraph when it should be giraph.useOutOfCoreGraph note that the last word is Graph not Giraph Rob From: Sebastian Stipkovic <sebastian.stipko...@gmail.com> Reply-To: <user@giraph.apache.org> Date: Thursday, 5 December 2013 20:39 To: <user@giraph.apache.org> Subject: out of core option > Hello, > > I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1 on a single > node cluster. It computes a tiny graph successful. But if the > input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector) > exception, although I had turned on the out-of-memory-option. The job with > out-of-memory-option works only well with a tiny graph (0.9 GB). What is > Wrong? Does I have to do furthermore configurations? > > My Configurations are as follows: > > > namevalue > fs.s3n.implorg.apache.hadoop.fs.s3native.NativeS3FileSystem > mapred.task.cache.levels2 > giraph.vertexOutputFormatClassorg.apache.giraph.examples.MyShortestPaths$MyOut > putFormat > hadoop.tmp.dir/app/hadoop/tmp > hadoop.native.libtrue > map.sort.classorg.apache.hadoop.util.QuickSort > dfs.namenode.decommission.nodes.per.interval5 > dfs.https.need.client.authfalse > ipc.client.idlethreshold4000 > dfs.datanode.data.dir.perm755 > mapred.system.dir${hadoop.tmp.dir}/mapred/system > mapred.job.tracker.persist.jobstatus.hours0 > dfs.datanode.address0.0.0.0:50010 <http://0.0.0.0:50010> > dfs.namenode.logging.levelinfo > dfs.block.access.token.enablefalse > io.skip.checksum.errorsfalse > fs.default.name <http://fs.default.name> hdfs://localhost:54310 > mapred.cluster.reduce.memory.mb-1 > mapred.child.tmp./tmp > fs.har.impl.disable.cachetrue > dfs.safemode.threshold.pct0.999f > mapred.skip.reduce.max.skip.groups0 > dfs.namenode.handler.count10 > dfs.blockreport.initialDelay0 > mapred.heartbeats.in.second100 > mapred.tasktracker.dns.nameserverdefault > io.sort.factor10 > mapred.task.timeout600000 > giraph.maxWorkers1 > mapred.max.tracker.failures4 > hadoop.rpc.socket.factory.class.defaultorg.apache.hadoop.net.StandardSocketFac > tory > mapred.job.tracker.jobhistory.lru.cache.size5 > fs.hdfs.implorg.apache.hadoop.hdfs.DistributedFileSystem > mapred.queue.default.acl-administer-jobs* > dfs.block.access.key.update.interval600 > mapred.skip.map.auto.incr.proc.counttrue > mapreduce.job.complete.cancel.delegation.tokenstrue > io.mapfile.bloom.size1048576 > mapreduce.reduce.shuffle.connect.timeout180000 > dfs.safemode.extension30000 > mapred.jobtracker.blacklist.fault-timeout-window180 > tasktracker.http.threads40 > mapred.job.shuffle.merge.percent0.66 > mapreduce.inputformat.classorg.apache.giraph.bsp.BspInputFormat > fs.ftp.implorg.apache.hadoop.fs.ftp.FTPFileSystem > user.name <http://user.name> hduser > mapred.output.compressfalse > io.bytes.per.checksum512 > giraph.isStaticGraphtrue > mapred.healthChecker.script.timeout600000 > topology.node.switch.mapping.implorg.apache.hadoop.net.ScriptBasedMapping > dfs.https.server.keystore.resourcessl-server.xml > mapred.reduce.slowstart.completed.maps0.05 > mapred.reduce.max.attempts4 > fs.ramfs.implorg.apache.hadoop.fs.InMemoryFileSystem > dfs.block.access.token.lifetime600 > dfs.name.edits.dir${dfs.name.dir} > mapred.skip.map.max.skip.records0 > mapred.cluster.map.memory.mb-1 > hadoop.security.group.mappingorg.apache.hadoop.security.ShellBasedUnixGroupsMa > pping > mapred.job.tracker.persist.jobstatus.dir/jobtracker/jobsInfo > mapred.jarhdfs://localhost:54310/app/hadoop/tmp/mapred/staging/hduser/.staging > /job_201312051827_0001/job.jar > dfs.block.size67108864 > fs.s3.buffer.dir${hadoop.tmp.dir}/s3 > job.end.retry.attempts0 > fs.file.implorg.apache.hadoop.fs.LocalFileSystem > mapred.local.dir.minspacestart0 > mapred.output.compression.typeRECORD > dfs.datanode.ipc.address0.0.0.0:50020 <http://0.0.0.0:50020> > dfs.permissionstrue > topology.script.number.args100 > io.mapfile.bloom.error.rate0.005 > mapred.cluster.max.reduce.memory.mb-1 > mapred.max.tracker.blacklists4 > mapred.task.profile.maps0-2 > dfs.datanode.https.address0.0.0.0:50475 <http://0.0.0.0:50475> > mapred.userlog.retain.hours24 > dfs.secondary.http.address0.0.0.0:50090 <http://0.0.0.0:50090> > dfs.replication.max512 > mapred.job.tracker.persist.jobstatus.activefalse > hadoop.security.authorizationfalse > local.cache.size10737418240 > dfs.namenode.delegation.token.renew-interval86400000 > mapred.min.split.size0 > mapred.map.tasks2 > mapred.child.java.opts-Xmx4000m > mapreduce.job.counters.limit120 > dfs.https.client.keystore.resourcessl-client.xml > mapred.job.queue.name <http://mapred.job.queue.name> default > dfs.https.address0.0.0.0:50470 <http://0.0.0.0:50470> > mapred.job.tracker.retiredjobs.cache.size1000 > dfs.balance.bandwidthPerSec1048576 > ipc.server.listen.queue.size128 > mapred.inmem.merge.threshold1000 > job.end.retry.interval30000 > mapred.skip.attempts.to.start.skipping2 > fs.checkpoint.dir${hadoop.tmp.dir}/dfs/namesecondary > mapred.reduce.tasks0 > mapred.merge.recordsBeforeProgress10000 > mapred.userlog.limit.kb0 > mapred.job.reduce.memory.mb-1 > dfs.max.objects0 > webinterface.private.actionsfalse > io.sort.spill.percent0.80 > mapred.job.shuffle.input.buffer.percent0.70 > mapred.job.name <http://mapred.job.name> Giraph: > org.apache.giraph.examples.MyShortestPaths > dfs.datanode.dns.nameserverdefault > mapred.map.tasks.speculative.executionfalse > hadoop.util.hash.typemurmur > dfs.blockreport.intervalMsec3600000 > mapred.map.max.attempts0 > mapreduce.job.acl-view-job > dfs.client.block.write.retries3 > mapred.job.tracker.handler.count10 > mapreduce.reduce.shuffle.read.timeout180000 > mapred.tasktracker.expiry.interval600000 > dfs.https.enablefalse > mapred.jobtracker.maxtasks.per.job-1 > mapred.jobtracker.job.history.block.size3145728 > giraph.useOutOfCoreGiraphtrue > keep.failed.task.filesfalse > mapreduce.outputformat.classorg.apache.giraph.bsp.BspOutputFormat > dfs.datanode.failed.volumes.tolerated0 > ipc.client.tcpnodelayfalse > mapred.task.profile.reduces0-2 > mapred.output.compression.codecorg.apache.hadoop.io.compress.DefaultCodec > io.map.index.skip0 > mapred.working.dirhdfs://localhost:54310/user/hduser > ipc.server.tcpnodelayfalse > mapred.jobtracker.blacklist.fault-bucket-width15 > dfs.namenode.delegation.key.update-interval86400000 > mapred.used.genericoptionsparsertrue > mapred.mapper.new-apitrue > mapred.job.map.memory.mb-1 > giraph.vertex.input.dirhdfs://localhost:54310/user/hduser/output > dfs.default.chunk.view.size32768 > hadoop.logfile.size10000000 > mapred.reduce.tasks.speculative.executiontrue > mapreduce.job.dirhdfs://localhost:54310/app/hadoop/tmp/mapred/staging/hduser/. > staging/job_201312051827_0001 > mapreduce.tasktracker.outofband.heartbeatfalse > mapreduce.reduce.input.limit-1 > dfs.datanode.du.reserved0 > hadoop.security.authenticationsimple > fs.checkpoint.period3600 > dfs.web.ugiwebuser,webgroup > mapred.job.reuse.jvm.num.tasks1 > mapred.jobtracker.completeuserjobs.maximum100 > dfs.df.interval60000 > dfs.data.dir${hadoop.tmp.dir}/dfs/data > mapred.task.tracker.task-controllerorg.apache.hadoop.mapred.DefaultTaskControl > ler > giraph.minWorkers1 > fs.s3.maxRetries4 > dfs.datanode.dns.interfacedefault > mapred.cluster.max.map.memory.mb-1 > dfs.support.appendfalse > mapreduce.job.acl-modify-job > dfs.permissions.supergroupsupergroup > mapred.local.dir${hadoop.tmp.dir}/mapred/local > fs.hftp.implorg.apache.hadoop.hdfs.HftpFileSystem > fs.trash.interval0 > fs.s3.sleepTimeSeconds10 > dfs.replication.min1 > mapred.submit.replication10 > fs.har.implorg.apache.hadoop.fs.HarFileSystem > mapred.map.output.compression.codecorg.apache.hadoop.io.compress.DefaultCodec > mapred.tasktracker.dns.interfacedefault > dfs.namenode.decommission.interval30 > dfs.http.address0.0.0.0:50070 <http://0.0.0.0:50070> > dfs.heartbeat.interval3 > mapred.job.trackerlocalhost:54311 > mapreduce.job.submithosthduser > io.seqfile.sorter.recordlimit1000000 > giraph.vertexInputFormatClassorg.apache.giraph.examples.MyShortestPaths$MyInpu > tFormat > dfs.name.dir${hadoop.tmp.dir}/dfs/name > mapred.line.input.format.linespermap1 > mapred.jobtracker.taskSchedulerorg.apache.hadoop.mapred.JobQueueTaskScheduler > dfs.datanode.http.address0.0.0.0:50075 <http://0.0.0.0:50075> > mapred.local.dir.minspacekill0 > dfs.replication.interval3 > io.sort.record.percent0.05 > fs.kfs.implorg.apache.hadoop.fs.kfs.KosmosFileSystem > mapred.temp.dir${hadoop.tmp.dir}/mapred/temp > mapred.tasktracker.reduce.tasks.maximum2 > mapreduce.job.user.classpath.firsttrue > dfs.replication1 > fs.checkpoint.edits.dir${fs.checkpoint.dir} > giraph.computationClassorg.apache.giraph.examples.MyShortestPaths > mapred.tasktracker.tasks.sleeptime-before-sigkill5000 > mapred.job.reduce.input.buffer.percent0.0 > mapred.tasktracker.indexcache.mb10 > mapreduce.job.split.metainfo.maxsize10000000 > hadoop.logfile.count10 > mapred.skip.reduce.auto.incr.proc.counttrue > mapreduce.job.submithostaddress127.0.1.1 > io.seqfile.compress.blocksize1000000 > fs.s3.block.size67108864 > mapred.tasktracker.taskmemorymanager.monitoring-interval5000 > giraph.minPercentResponded100.0 > mapred.queue.default.stateRUNNING > mapred.acls.enabledfalse > mapreduce.jobtracker.staging.root.dir${hadoop.tmp.dir}/mapred/staging > mapred.queue.namesdefault > dfs.access.time.precision3600000 > fs.hsftp.implorg.apache.hadoop.hdfs.HsftpFileSystem > mapred.task.tracker.http.address0.0.0.0:50060 <http://0.0.0.0:50060> > mapred.reduce.parallel.copies5 > io.seqfile.lazydecompresstrue > mapred.output.dir/user/hduser/output/shortestpaths > io.sort.mb100 > ipc.client.connection.maxidletime10000 > mapred.compress.map.outputfalse > hadoop.security.uid.cache.secs14400 > mapred.task.tracker.report.address127.0.0.1:0 <http://127.0.0.1:0> > mapred.healthChecker.interval60000 > ipc.client.kill.max10 > ipc.client.connect.max.retries10 > ipc.ping.interval300000 > mapreduce.user.classpath.firsttrue > mapreduce.map.classorg.apache.giraph.graph.GraphMapper > fs.s3.implorg.apache.hadoop.fs.s3.S3FileSystem > mapred.user.jobconf.limit5242880 > mapred.job.tracker.http.address0.0.0.0:50030 <http://0.0.0.0:50030> > io.file.buffer.size4096 > mapred.jobtracker.restart.recoverfalse > io.serializationsorg.apache.hadoop.io.serializer.WritableSerialization > dfs.datanode.handler.count3 > mapred.reduce.copy.backoff300 > mapred.task.profilefalse > dfs.replication.considerLoadtrue > jobclient.output.filterFAILED > dfs.namenode.delegation.token.max-lifetime604800000 > mapred.tasktracker.map.tasks.maximum4 > io.compression.codecsorg.apache.hadoop.io.compress.DefaultCodec,org.apache.had > oop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec > fs.checkpoint.size67108864 > > Additionally, if I have more than one worker I get an Exception, too? Are my > configurations wrong? > > > best regards, > Sebastian >