Sebastian

You've made a minor typo in the configuration setting which means you
haven't actually enabled out of core graph mode.

You have giraph.useOutOfCoreGiraph when it should be
giraph.useOutOfCoreGraph ­ note that the last word is Graph not Giraph

Rob

From:  Sebastian Stipkovic <sebastian.stipko...@gmail.com>
Reply-To:  <user@giraph.apache.org>
Date:  Thursday, 5 December 2013 20:39
To:  <user@giraph.apache.org>
Subject:  out of core option

> Hello,
> 
> I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
> node cluster. It computes a tiny graph successful. But if the
> input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
> exception, although I had turned on the out-of-memory-option. The job with
> out-of-memory-option works only well with a tiny graph (0.9 GB).  What is
> Wrong? Does I have to do furthermore configurations?
> 
> My Configurations are as follows:
> 
> 
> namevalue
> fs.s3n.implorg.apache.hadoop.fs.s3native.NativeS3FileSystem
> mapred.task.cache.levels2
> giraph.vertexOutputFormatClassorg.apache.giraph.examples.MyShortestPaths$MyOut
> putFormat
> hadoop.tmp.dir/app/hadoop/tmp
> hadoop.native.libtrue
> map.sort.classorg.apache.hadoop.util.QuickSort
> dfs.namenode.decommission.nodes.per.interval5
> dfs.https.need.client.authfalse
> ipc.client.idlethreshold4000
> dfs.datanode.data.dir.perm755
> mapred.system.dir${hadoop.tmp.dir}/mapred/system
> mapred.job.tracker.persist.jobstatus.hours0
> dfs.datanode.address0.0.0.0:50010 <http://0.0.0.0:50010>
> dfs.namenode.logging.levelinfo
> dfs.block.access.token.enablefalse
> io.skip.checksum.errorsfalse
> fs.default.name <http://fs.default.name> hdfs://localhost:54310
> mapred.cluster.reduce.memory.mb-1
> mapred.child.tmp./tmp
> fs.har.impl.disable.cachetrue
> dfs.safemode.threshold.pct0.999f
> mapred.skip.reduce.max.skip.groups0
> dfs.namenode.handler.count10
> dfs.blockreport.initialDelay0
> mapred.heartbeats.in.second100
> mapred.tasktracker.dns.nameserverdefault
> io.sort.factor10
> mapred.task.timeout600000
> giraph.maxWorkers1
> mapred.max.tracker.failures4
> hadoop.rpc.socket.factory.class.defaultorg.apache.hadoop.net.StandardSocketFac
> tory
> mapred.job.tracker.jobhistory.lru.cache.size5
> fs.hdfs.implorg.apache.hadoop.hdfs.DistributedFileSystem
> mapred.queue.default.acl-administer-jobs*
> dfs.block.access.key.update.interval600
> mapred.skip.map.auto.incr.proc.counttrue
> mapreduce.job.complete.cancel.delegation.tokenstrue
> io.mapfile.bloom.size1048576
> mapreduce.reduce.shuffle.connect.timeout180000
> dfs.safemode.extension30000
> mapred.jobtracker.blacklist.fault-timeout-window180
> tasktracker.http.threads40
> mapred.job.shuffle.merge.percent0.66
> mapreduce.inputformat.classorg.apache.giraph.bsp.BspInputFormat
> fs.ftp.implorg.apache.hadoop.fs.ftp.FTPFileSystem
> user.name <http://user.name> hduser
> mapred.output.compressfalse
> io.bytes.per.checksum512
> giraph.isStaticGraphtrue
> mapred.healthChecker.script.timeout600000
> topology.node.switch.mapping.implorg.apache.hadoop.net.ScriptBasedMapping
> dfs.https.server.keystore.resourcessl-server.xml
> mapred.reduce.slowstart.completed.maps0.05
> mapred.reduce.max.attempts4
> fs.ramfs.implorg.apache.hadoop.fs.InMemoryFileSystem
> dfs.block.access.token.lifetime600
> dfs.name.edits.dir${dfs.name.dir}
> mapred.skip.map.max.skip.records0
> mapred.cluster.map.memory.mb-1
> hadoop.security.group.mappingorg.apache.hadoop.security.ShellBasedUnixGroupsMa
> pping
> mapred.job.tracker.persist.jobstatus.dir/jobtracker/jobsInfo
> mapred.jarhdfs://localhost:54310/app/hadoop/tmp/mapred/staging/hduser/.staging
> /job_201312051827_0001/job.jar
> dfs.block.size67108864
> fs.s3.buffer.dir${hadoop.tmp.dir}/s3
> job.end.retry.attempts0
> fs.file.implorg.apache.hadoop.fs.LocalFileSystem
> mapred.local.dir.minspacestart0
> mapred.output.compression.typeRECORD
> dfs.datanode.ipc.address0.0.0.0:50020 <http://0.0.0.0:50020>
> dfs.permissionstrue
> topology.script.number.args100
> io.mapfile.bloom.error.rate0.005
> mapred.cluster.max.reduce.memory.mb-1
> mapred.max.tracker.blacklists4
> mapred.task.profile.maps0-2
> dfs.datanode.https.address0.0.0.0:50475 <http://0.0.0.0:50475>
> mapred.userlog.retain.hours24
> dfs.secondary.http.address0.0.0.0:50090 <http://0.0.0.0:50090>
> dfs.replication.max512
> mapred.job.tracker.persist.jobstatus.activefalse
> hadoop.security.authorizationfalse
> local.cache.size10737418240
> dfs.namenode.delegation.token.renew-interval86400000
> mapred.min.split.size0
> mapred.map.tasks2
> mapred.child.java.opts-Xmx4000m
> mapreduce.job.counters.limit120
> dfs.https.client.keystore.resourcessl-client.xml
> mapred.job.queue.name <http://mapred.job.queue.name> default
> dfs.https.address0.0.0.0:50470 <http://0.0.0.0:50470>
> mapred.job.tracker.retiredjobs.cache.size1000
> dfs.balance.bandwidthPerSec1048576
> ipc.server.listen.queue.size128
> mapred.inmem.merge.threshold1000
> job.end.retry.interval30000
> mapred.skip.attempts.to.start.skipping2
> fs.checkpoint.dir${hadoop.tmp.dir}/dfs/namesecondary
> mapred.reduce.tasks0
> mapred.merge.recordsBeforeProgress10000
> mapred.userlog.limit.kb0
> mapred.job.reduce.memory.mb-1
> dfs.max.objects0
> webinterface.private.actionsfalse
> io.sort.spill.percent0.80
> mapred.job.shuffle.input.buffer.percent0.70
> mapred.job.name <http://mapred.job.name> Giraph:
> org.apache.giraph.examples.MyShortestPaths
> dfs.datanode.dns.nameserverdefault
> mapred.map.tasks.speculative.executionfalse
> hadoop.util.hash.typemurmur
> dfs.blockreport.intervalMsec3600000
> mapred.map.max.attempts0
> mapreduce.job.acl-view-job
> dfs.client.block.write.retries3
> mapred.job.tracker.handler.count10
> mapreduce.reduce.shuffle.read.timeout180000
> mapred.tasktracker.expiry.interval600000
> dfs.https.enablefalse
> mapred.jobtracker.maxtasks.per.job-1
> mapred.jobtracker.job.history.block.size3145728
> giraph.useOutOfCoreGiraphtrue
> keep.failed.task.filesfalse
> mapreduce.outputformat.classorg.apache.giraph.bsp.BspOutputFormat
> dfs.datanode.failed.volumes.tolerated0
> ipc.client.tcpnodelayfalse
> mapred.task.profile.reduces0-2
> mapred.output.compression.codecorg.apache.hadoop.io.compress.DefaultCodec
> io.map.index.skip0
> mapred.working.dirhdfs://localhost:54310/user/hduser
> ipc.server.tcpnodelayfalse
> mapred.jobtracker.blacklist.fault-bucket-width15
> dfs.namenode.delegation.key.update-interval86400000
> mapred.used.genericoptionsparsertrue
> mapred.mapper.new-apitrue
> mapred.job.map.memory.mb-1
> giraph.vertex.input.dirhdfs://localhost:54310/user/hduser/output
> dfs.default.chunk.view.size32768
> hadoop.logfile.size10000000
> mapred.reduce.tasks.speculative.executiontrue
> mapreduce.job.dirhdfs://localhost:54310/app/hadoop/tmp/mapred/staging/hduser/.
> staging/job_201312051827_0001
> mapreduce.tasktracker.outofband.heartbeatfalse
> mapreduce.reduce.input.limit-1
> dfs.datanode.du.reserved0
> hadoop.security.authenticationsimple
> fs.checkpoint.period3600
> dfs.web.ugiwebuser,webgroup
> mapred.job.reuse.jvm.num.tasks1
> mapred.jobtracker.completeuserjobs.maximum100
> dfs.df.interval60000
> dfs.data.dir${hadoop.tmp.dir}/dfs/data
> mapred.task.tracker.task-controllerorg.apache.hadoop.mapred.DefaultTaskControl
> ler
> giraph.minWorkers1
> fs.s3.maxRetries4
> dfs.datanode.dns.interfacedefault
> mapred.cluster.max.map.memory.mb-1
> dfs.support.appendfalse
> mapreduce.job.acl-modify-job
> dfs.permissions.supergroupsupergroup
> mapred.local.dir${hadoop.tmp.dir}/mapred/local
> fs.hftp.implorg.apache.hadoop.hdfs.HftpFileSystem
> fs.trash.interval0
> fs.s3.sleepTimeSeconds10
> dfs.replication.min1
> mapred.submit.replication10
> fs.har.implorg.apache.hadoop.fs.HarFileSystem
> mapred.map.output.compression.codecorg.apache.hadoop.io.compress.DefaultCodec
> mapred.tasktracker.dns.interfacedefault
> dfs.namenode.decommission.interval30
> dfs.http.address0.0.0.0:50070 <http://0.0.0.0:50070>
> dfs.heartbeat.interval3
> mapred.job.trackerlocalhost:54311
> mapreduce.job.submithosthduser
> io.seqfile.sorter.recordlimit1000000
> giraph.vertexInputFormatClassorg.apache.giraph.examples.MyShortestPaths$MyInpu
> tFormat
> dfs.name.dir${hadoop.tmp.dir}/dfs/name
> mapred.line.input.format.linespermap1
> mapred.jobtracker.taskSchedulerorg.apache.hadoop.mapred.JobQueueTaskScheduler
> dfs.datanode.http.address0.0.0.0:50075 <http://0.0.0.0:50075>
> mapred.local.dir.minspacekill0
> dfs.replication.interval3
> io.sort.record.percent0.05
> fs.kfs.implorg.apache.hadoop.fs.kfs.KosmosFileSystem
> mapred.temp.dir${hadoop.tmp.dir}/mapred/temp
> mapred.tasktracker.reduce.tasks.maximum2
> mapreduce.job.user.classpath.firsttrue
> dfs.replication1
> fs.checkpoint.edits.dir${fs.checkpoint.dir}
> giraph.computationClassorg.apache.giraph.examples.MyShortestPaths
> mapred.tasktracker.tasks.sleeptime-before-sigkill5000
> mapred.job.reduce.input.buffer.percent0.0
> mapred.tasktracker.indexcache.mb10
> mapreduce.job.split.metainfo.maxsize10000000
> hadoop.logfile.count10
> mapred.skip.reduce.auto.incr.proc.counttrue
> mapreduce.job.submithostaddress127.0.1.1
> io.seqfile.compress.blocksize1000000
> fs.s3.block.size67108864
> mapred.tasktracker.taskmemorymanager.monitoring-interval5000
> giraph.minPercentResponded100.0
> mapred.queue.default.stateRUNNING
> mapred.acls.enabledfalse
> mapreduce.jobtracker.staging.root.dir${hadoop.tmp.dir}/mapred/staging
> mapred.queue.namesdefault
> dfs.access.time.precision3600000
> fs.hsftp.implorg.apache.hadoop.hdfs.HsftpFileSystem
> mapred.task.tracker.http.address0.0.0.0:50060 <http://0.0.0.0:50060>
> mapred.reduce.parallel.copies5
> io.seqfile.lazydecompresstrue
> mapred.output.dir/user/hduser/output/shortestpaths
> io.sort.mb100
> ipc.client.connection.maxidletime10000
> mapred.compress.map.outputfalse
> hadoop.security.uid.cache.secs14400
> mapred.task.tracker.report.address127.0.0.1:0 <http://127.0.0.1:0>
> mapred.healthChecker.interval60000
> ipc.client.kill.max10
> ipc.client.connect.max.retries10
> ipc.ping.interval300000
> mapreduce.user.classpath.firsttrue
> mapreduce.map.classorg.apache.giraph.graph.GraphMapper
> fs.s3.implorg.apache.hadoop.fs.s3.S3FileSystem
> mapred.user.jobconf.limit5242880
> mapred.job.tracker.http.address0.0.0.0:50030 <http://0.0.0.0:50030>
> io.file.buffer.size4096
> mapred.jobtracker.restart.recoverfalse
> io.serializationsorg.apache.hadoop.io.serializer.WritableSerialization
> dfs.datanode.handler.count3
> mapred.reduce.copy.backoff300
> mapred.task.profilefalse
> dfs.replication.considerLoadtrue
> jobclient.output.filterFAILED
> dfs.namenode.delegation.token.max-lifetime604800000
> mapred.tasktracker.map.tasks.maximum4
> io.compression.codecsorg.apache.hadoop.io.compress.DefaultCodec,org.apache.had
> oop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec
> fs.checkpoint.size67108864
> 
> Additionally, if I have more than one worker I get an Exception, too? Are my
> configurations wrong?
> 
> 
> best regards,
> Sebastian
> 

Reply via email to