Re: out of core option

2014-01-23 Thread Claudio Martella
(DiskBackedPartitionStore.java:228)
 ... 13 more Caused by: java.lang.NullPointerException at
 org.apache.giraph.partition.DiskBackedPartitionStore$GetPartition.call(DiskBackedPartitionStore.java:692)
 at
 org.apache.giraph.partition.DiskBackedPartitionStore$GetPartition.call(DiskBackedPartitionStore.java:658)
 at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at
 java.util.concurrent.FutureTask.run(FutureTask.java:166) at
 org.apache.giraph.partition.DiskBackedPartitionStore$DirectExecutorService.execute(DiskBackedPartitionStore.java:972)
 at
 java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:132)
 ... 14 more


 Thanks,
 Sebastian


 2013/12/5 Rob Vesse rve...@dotnetrdf.org

 Sebastian

 You've made a minor typo in the configuration setting which means you
 haven't actually enabled out of core graph mode.

 You have *giraph.useOutOfCoreGiraph *when it should be 
 *giraph.useOutOfCoreGraph
 *– note that the last word is Graph not Giraph

 Rob

 From: Sebastian Stipkovic sebastian.stipko...@gmail.com
 Reply-To: user@giraph.apache.org
 Date: Thursday, 5 December 2013 20:39
 To: user@giraph.apache.org
 Subject: out of core option

 Hello,

 I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
 node cluster. It computes a tiny graph successful. But if the
 input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
 exception, although I had turned on the out-of-memory-option. The job
 with out-of-memory-option works only well with a tiny graph (0.9 GB).  What
 is Wrong? Does I have to do furthermore configurations?

 My Configurations are as follows:


 namevalue*fs.s3n.impl*org.apache.hadoop.fs.s3native.NativeS3FileSystem
 *mapred.task.cache.levels*2*giraph.vertexOutputFormatClass*
 org.apache.giraph.examples.MyShortestPaths$MyOutputFormat
 *hadoop.tmp.dir*/app/hadoop/tmp*hadoop.native.lib*true*map.sort.class*org.apache.hadoop.util.QuickSort
 *dfs.namenode.decommission.nodes.per.interval*5
 *dfs.https.need.client.auth*false *ipc.client.idlethreshold*4000
 *dfs.datanode.data.dir.perm*755*mapred.system.dir*
 ${hadoop.tmp.dir}/mapred/system
 *mapred.job.tracker.persist.jobstatus.hours*0*dfs.datanode.address*
 0.0.0.0:50010*dfs.namenode.logging.level*info
 *dfs.block.access.token.enable* 
 false*io.skip.checksum.errors*false*fs.default.name
 http://fs.default.name* hdfs://localhost:54310
 *mapred.cluster.reduce.memory.mb*-1*mapred.child.tmp* ./tmp
 *fs.har.impl.disable.cache*true*dfs.safemode.threshold.pct*0.999f
 *mapred.skip.reduce.max.skip.groups*0*dfs.namenode.handler.count*10
 *dfs.blockreport.initialDelay* 0*mapred.heartbeats.in.second*100
 *mapred.tasktracker.dns.nameserver*default*io.sort.factor* 10
 *mapred.task.timeout*60*giraph.maxWorkers*1
 *mapred.max.tracker.failures* 4
 *hadoop.rpc.socket.factory.class.default*
 org.apache.hadoop.net.StandardSocketFactory
 *mapred.job.tracker.jobhistory.lru.cache.size* 5*fs.hdfs.impl*
 org.apache.hadoop.hdfs.DistributedFileSystem
 *mapred.queue.default.acl-administer-jobs* *
 *dfs.block.access.key.update.interval*600
 *mapred.skip.map.auto.incr.proc.count*true
 *mapreduce.job.complete.cancel.delegation.tokens*true
 *io.mapfile.bloom.size*1048576
 *mapreduce.reduce.shuffle.connect.timeout* 18
 *dfs.safemode.extension*3
 *mapred.jobtracker.blacklist.fault-timeout-window*180
 *tasktracker.http.threads*40*mapred.job.shuffle.merge.percent*0.66
 *mapreduce.inputformat.class* org.apache.giraph.bsp.BspInputFormat
 *fs.ftp.impl*org.apache.hadoop.fs.ftp.FTPFileSystem*user.name
 http://user.name* hduser*mapred.output.compress*false
 *io.bytes.per.checksum*512*giraph.isStaticGraph* true
 *mapred.healthChecker.script.timeout*60
 *topology.node.switch.mapping.impl*
 org.apache.hadoop.net.ScriptBasedMapping
 *dfs.https.server.keystore.resource*ssl-server.xml
 *mapred.reduce.slowstart.completed.maps*0.05
 *mapred.reduce.max.attempts*4*fs.ramfs.impl*
 org.apache.hadoop.fs.InMemoryFileSystem
 *dfs.block.access.token.lifetime* 600*dfs.name.edits.dir*
 ${dfs.name.dir}*mapred.skip.map.max.skip.records*0
 *mapred.cluster.map.memory.mb*-1*hadoop.security.group.mapping*
 org.apache.hadoop.security.ShellBasedUnixGroupsMapping
 *mapred.job.tracker.persist.jobstatus.dir*/jobtracker/jobsInfo
 *mapred.jar*hdfs://localhost:54310
 /app/hadoop/tmp/mapred/staging/hduser/.staging/job_201312051827_0001/job.jar
 *dfs.block.size*67108864*fs.s3.buffer.dir*${hadoop.tmp.dir}/s3
 *job.end.retry.attempts* 0*fs.file.impl*
 org.apache.hadoop.fs.LocalFileSystem*mapred.local.dir.minspacestart*0
 *mapred.output.compression.type*RECORD*dfs.datanode.ipc.address*
 0.0.0.0:50020 *dfs.permissions*true*topology.script.number.args*100
 *io.mapfile.bloom.error.rate* 0.005
 *mapred.cluster.max.reduce.memory.mb*-1*mapred.max.tracker.blacklists*4
 *mapred.task.profile.maps*0-2*dfs.datanode.https.address*0.0.0.0:50475
 *mapred.userlog.retain.hours*24*dfs.secondary.http.address*
 0.0.0.0:50090 *dfs.replication.max*512

Re: out of core option

2014-01-22 Thread Yingyi Bu
(DiskBackedPartitionStore.java:692)
 at
 org.apache.giraph.partition.DiskBackedPartitionStore$GetPartition.call(DiskBackedPartitionStore.java:658)
 at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at
 java.util.concurrent.FutureTask.run(FutureTask.java:166) at
 org.apache.giraph.partition.DiskBackedPartitionStore$DirectExecutorService.execute(DiskBackedPartitionStore.java:972)
 at
 java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:132)
 ... 14 more


 Thanks,
 Sebastian


 2013/12/5 Rob Vesse rve...@dotnetrdf.org

 Sebastian

 You've made a minor typo in the configuration setting which means you
 haven't actually enabled out of core graph mode.

 You have *giraph.useOutOfCoreGiraph *when it should be 
 *giraph.useOutOfCoreGraph
 *– note that the last word is Graph not Giraph

 Rob

 From: Sebastian Stipkovic sebastian.stipko...@gmail.com
 Reply-To: user@giraph.apache.org
 Date: Thursday, 5 December 2013 20:39
 To: user@giraph.apache.org
 Subject: out of core option

 Hello,

 I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
 node cluster. It computes a tiny graph successful. But if the
 input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
 exception, although I had turned on the out-of-memory-option. The job
 with out-of-memory-option works only well with a tiny graph (0.9 GB).  What
 is Wrong? Does I have to do furthermore configurations?

 My Configurations are as follows:


 namevalue*fs.s3n.impl*org.apache.hadoop.fs.s3native.NativeS3FileSystem
 *mapred.task.cache.levels*2*giraph.vertexOutputFormatClass*
 org.apache.giraph.examples.MyShortestPaths$MyOutputFormat
 *hadoop.tmp.dir*/app/hadoop/tmp*hadoop.native.lib*true*map.sort.class*org.apache.hadoop.util.QuickSort
 *dfs.namenode.decommission.nodes.per.interval*5
 *dfs.https.need.client.auth*false *ipc.client.idlethreshold*4000
 *dfs.datanode.data.dir.perm*755*mapred.system.dir*
 ${hadoop.tmp.dir}/mapred/system
 *mapred.job.tracker.persist.jobstatus.hours*0*dfs.datanode.address*
 0.0.0.0:50010*dfs.namenode.logging.level*info
 *dfs.block.access.token.enable* 
 false*io.skip.checksum.errors*false*fs.default.name
 http://fs.default.name* hdfs://localhost:54310
 *mapred.cluster.reduce.memory.mb*-1*mapred.child.tmp* ./tmp
 *fs.har.impl.disable.cache*true*dfs.safemode.threshold.pct*0.999f
 *mapred.skip.reduce.max.skip.groups*0*dfs.namenode.handler.count*10
 *dfs.blockreport.initialDelay* 0*mapred.heartbeats.in.second*100
 *mapred.tasktracker.dns.nameserver*default*io.sort.factor* 10
 *mapred.task.timeout*60*giraph.maxWorkers*1
 *mapred.max.tracker.failures* 4*hadoop.rpc.socket.factory.class.default*
 org.apache.hadoop.net.StandardSocketFactory
 *mapred.job.tracker.jobhistory.lru.cache.size* 5*fs.hdfs.impl*
 org.apache.hadoop.hdfs.DistributedFileSystem
 *mapred.queue.default.acl-administer-jobs* *
 *dfs.block.access.key.update.interval*600
 *mapred.skip.map.auto.incr.proc.count*true
 *mapreduce.job.complete.cancel.delegation.tokens*true
 *io.mapfile.bloom.size*1048576*mapreduce.reduce.shuffle.connect.timeout*
 18*dfs.safemode.extension*3
 *mapred.jobtracker.blacklist.fault-timeout-window*180
 *tasktracker.http.threads*40*mapred.job.shuffle.merge.percent*0.66
 *mapreduce.inputformat.class* org.apache.giraph.bsp.BspInputFormat
 *fs.ftp.impl*org.apache.hadoop.fs.ftp.FTPFileSystem*user.name
 http://user.name* hduser*mapred.output.compress*false
 *io.bytes.per.checksum*512*giraph.isStaticGraph* true
 *mapred.healthChecker.script.timeout*60
 *topology.node.switch.mapping.impl*
 org.apache.hadoop.net.ScriptBasedMapping
 *dfs.https.server.keystore.resource*ssl-server.xml
 *mapred.reduce.slowstart.completed.maps*0.05
 *mapred.reduce.max.attempts*4*fs.ramfs.impl*
 org.apache.hadoop.fs.InMemoryFileSystem*dfs.block.access.token.lifetime*
 600*dfs.name.edits.dir*${dfs.name.dir}*mapred.skip.map.max.skip.records*
 0 *mapred.cluster.map.memory.mb*-1*hadoop.security.group.mapping*
 org.apache.hadoop.security.ShellBasedUnixGroupsMapping
 *mapred.job.tracker.persist.jobstatus.dir*/jobtracker/jobsInfo
 *mapred.jar*hdfs://localhost:54310
 /app/hadoop/tmp/mapred/staging/hduser/.staging/job_201312051827_0001/job.jar
 *dfs.block.size*67108864*fs.s3.buffer.dir*${hadoop.tmp.dir}/s3
 *job.end.retry.attempts* 0*fs.file.impl*
 org.apache.hadoop.fs.LocalFileSystem*mapred.local.dir.minspacestart*0
 *mapred.output.compression.type*RECORD*dfs.datanode.ipc.address*
 0.0.0.0:50020 *dfs.permissions*true*topology.script.number.args*100
 *io.mapfile.bloom.error.rate* 0.005*mapred.cluster.max.reduce.memory.mb*
 -1*mapred.max.tracker.blacklists*4 *mapred.task.profile.maps*0-2
 *dfs.datanode.https.address*0.0.0.0:50475 *mapred.userlog.retain.hours*
 24*dfs.secondary.http.address*0.0.0.0:50090 *dfs.replication.max*512
 *mapred.job.tracker.persist.jobstatus.active*false
 *hadoop.security.authorization* false*local.cache.size*10737418240
 *dfs.namenode.delegation.token.renew-interval*8640
 *mapred.min.split.size*0

Re: out of core option

2013-12-06 Thread Sebastian Stipkovic
) at
 java.util.concurrent.FutureTask.run(FutureTask.java:166) at
 org.apache.giraph.partition.DiskBackedPartitionStore$DirectExecutorService.execute(DiskBackedPartitionStore.java:972)
 at
 java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:132)
 ... 14 more


 Thanks,
 Sebastian


 2013/12/5 Rob Vesse rve...@dotnetrdf.org

 Sebastian

 You've made a minor typo in the configuration setting which means you
 haven't actually enabled out of core graph mode.

 You have *giraph.useOutOfCoreGiraph *when it should be 
 *giraph.useOutOfCoreGraph
 *– note that the last word is Graph not Giraph

 Rob

 From: Sebastian Stipkovic sebastian.stipko...@gmail.com
 Reply-To: user@giraph.apache.org
 Date: Thursday, 5 December 2013 20:39
 To: user@giraph.apache.org
 Subject: out of core option

 Hello,

 I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
 node cluster. It computes a tiny graph successful. But if the
 input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
 exception, although I had turned on the out-of-memory-option. The job
 with out-of-memory-option works only well with a tiny graph (0.9 GB).  What
 is Wrong? Does I have to do furthermore configurations?

 My Configurations are as follows:


 namevalue*fs.s3n.impl*org.apache.hadoop.fs.s3native.NativeS3FileSystem
 *mapred.task.cache.levels*2*giraph.vertexOutputFormatClass*
 org.apache.giraph.examples.MyShortestPaths$MyOutputFormat
 *hadoop.tmp.dir*/app/hadoop/tmp*hadoop.native.lib*true*map.sort.class*org.apache.hadoop.util.QuickSort
 *dfs.namenode.decommission.nodes.per.interval*5
 *dfs.https.need.client.auth*false *ipc.client.idlethreshold*4000
 *dfs.datanode.data.dir.perm*755*mapred.system.dir*
 ${hadoop.tmp.dir}/mapred/system
 *mapred.job.tracker.persist.jobstatus.hours*0*dfs.datanode.address*
 0.0.0.0:50010*dfs.namenode.logging.level*info
 *dfs.block.access.token.enable* 
 false*io.skip.checksum.errors*false*fs.default.name
 http://fs.default.name* hdfs://localhost:54310
 *mapred.cluster.reduce.memory.mb*-1*mapred.child.tmp* ./tmp
 *fs.har.impl.disable.cache*true*dfs.safemode.threshold.pct*0.999f
 *mapred.skip.reduce.max.skip.groups*0*dfs.namenode.handler.count*10
 *dfs.blockreport.initialDelay* 0*mapred.heartbeats.in.second*100
 *mapred.tasktracker.dns.nameserver*default*io.sort.factor* 10
 *mapred.task.timeout*60*giraph.maxWorkers*1
 *mapred.max.tracker.failures* 4*hadoop.rpc.socket.factory.class.default*
 org.apache.hadoop.net.StandardSocketFactory
 *mapred.job.tracker.jobhistory.lru.cache.size* 5*fs.hdfs.impl*
 org.apache.hadoop.hdfs.DistributedFileSystem
 *mapred.queue.default.acl-administer-jobs* *
 *dfs.block.access.key.update.interval*600
 *mapred.skip.map.auto.incr.proc.count*true
 *mapreduce.job.complete.cancel.delegation.tokens*true
 *io.mapfile.bloom.size*1048576*mapreduce.reduce.shuffle.connect.timeout*
 18*dfs.safemode.extension*3
 *mapred.jobtracker.blacklist.fault-timeout-window*180
 *tasktracker.http.threads*40*mapred.job.shuffle.merge.percent*0.66
 *mapreduce.inputformat.class* org.apache.giraph.bsp.BspInputFormat
 *fs.ftp.impl*org.apache.hadoop.fs.ftp.FTPFileSystem*user.name
 http://user.name* hduser*mapred.output.compress*false
 *io.bytes.per.checksum*512*giraph.isStaticGraph* true
 *mapred.healthChecker.script.timeout*60
 *topology.node.switch.mapping.impl*
 org.apache.hadoop.net.ScriptBasedMapping
 *dfs.https.server.keystore.resource*ssl-server.xml
 *mapred.reduce.slowstart.completed.maps*0.05 *mapred.reduce.max.attempts*
 4*fs.ramfs.impl*org.apache.hadoop.fs.InMemoryFileSystem
 *dfs.block.access.token.lifetime* 600*dfs.name.edits.dir*${dfs.name.dir}
 *mapred.skip.map.max.skip.records*0 *mapred.cluster.map.memory.mb*-1
 *hadoop.security.group.mapping*
 org.apache.hadoop.security.ShellBasedUnixGroupsMapping
 *mapred.job.tracker.persist.jobstatus.dir*/jobtracker/jobsInfo
 *mapred.jar*hdfs://localhost:54310
 /app/hadoop/tmp/mapred/staging/hduser/.staging/job_201312051827_0001/job.jar
 *dfs.block.size*67108864*fs.s3.buffer.dir*${hadoop.tmp.dir}/s3
 *job.end.retry.attempts* 0*fs.file.impl*
 org.apache.hadoop.fs.LocalFileSystem*mapred.local.dir.minspacestart*0
 *mapred.output.compression.type*RECORD*dfs.datanode.ipc.address*
 0.0.0.0:50020 *dfs.permissions*true*topology.script.number.args*100
 *io.mapfile.bloom.error.rate* 0.005*mapred.cluster.max.reduce.memory.mb*
 -1*mapred.max.tracker.blacklists*4 *mapred.task.profile.maps*0-2
 *dfs.datanode.https.address*0.0.0.0:50475 *mapred.userlog.retain.hours*24
 *dfs.secondary.http.address*0.0.0.0:50090 *dfs.replication.max*512
 *mapred.job.tracker.persist.jobstatus.active*false
 *hadoop.security.authorization* false*local.cache.size*10737418240
 *dfs.namenode.delegation.token.renew-interval*8640
 *mapred.min.split.size*0*mapred.map.tasks*2*mapred.child.java.opts*-Xmx4000m
 *mapreduce.job.counters.limit*120*dfs.https.client.keystore.resource*
 ssl-client.xml *mapred.job.queue.name http://mapred.job.queue.name*
 default*dfs.https.address

Re: out of core option

2013-12-05 Thread Ameya Vilankar
Each worker is allocated *mapred.child.java.opts *memory, which in your
case is 4000M. Check if your server doesn't have enough memory for 2
Mappers. Also the out of memory option is available in two forms.
1. Out of core graph
2. Out of core messages.

Currently you are setting only the out of core graph and not the out of
core messages. Enable both of them. More information about options can be
found here: http://giraph.apache.org/options.html
set -D giraph.useOutOfCoreGraph=true -D giraph.useOutOfCoreMessages=true
while passing options to GiraphRunner.

Thanks,
Ameya


On Thu, Dec 5, 2013 at 12:39 PM, Sebastian Stipkovic 
sebastian.stipko...@gmail.com wrote:

 Hello,

 I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
 node cluster. It computes a tiny graph successful. But if the
 input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
 exception, although I had turned on the out-of-memory-option. The job with
 out-of-memory-option works only well with a tiny graph (0.9 GB).  What is
 Wrong? Does I have to do furthermore configurations?

 My Configurations are as follows:


 namevalue  *fs.s3n.impl*org.apache.hadoop.fs.s3native.NativeS3FileSystem
 *mapred.task.cache.levels*2  *giraph.vertexOutputFormatClass*
 org.apache.giraph.examples.MyShortestPaths$MyOutputFormat
 *hadoop.tmp.dir*/app/hadoop/tmp  *hadoop.native.lib*true  *map.sort.class*
 org.apache.hadoop.util.QuickSort
 *dfs.namenode.decommission.nodes.per.interval*5
 *dfs.https.need.client.auth*false  *ipc.client.idlethreshold*4000
 *dfs.datanode.data.dir.perm*755  *mapred.system.dir*
 ${hadoop.tmp.dir}/mapred/system
 *mapred.job.tracker.persist.jobstatus.hours*0  *dfs.datanode.address*
 0.0.0.0:50010  *dfs.namenode.logging.level*info
 *dfs.block.access.token.enable*false  *io.skip.checksum.errors*false  
 *fs.default.name
 http://fs.default.name*hdfs://localhost:54310
 *mapred.cluster.reduce.memory.mb*-1  *mapred.child.tmp*./tmp
 *fs.har.impl.disable.cache*true  *dfs.safemode.threshold.pct*0.999f
 *mapred.skip.reduce.max.skip.groups*0  *dfs.namenode.handler.count*10
 *dfs.blockreport.initialDelay*0  *mapred.heartbeats.in.second*100
 *mapred.tasktracker.dns.nameserver*default  *io.sort.factor*10
 *mapred.task.timeout*60  *giraph.maxWorkers*1
 *mapred.max.tracker.failures*4  *hadoop.rpc.socket.factory.class.default*
 org.apache.hadoop.net.StandardSocketFactory
 *mapred.job.tracker.jobhistory.lru.cache.size*5  *fs.hdfs.impl*
 org.apache.hadoop.hdfs.DistributedFileSystem
 *mapred.queue.default.acl-administer-jobs**
 *dfs.block.access.key.update.interval*600
 *mapred.skip.map.auto.incr.proc.count*true
 *mapreduce.job.complete.cancel.delegation.tokens*true
 *io.mapfile.bloom.size*1048576  *mapreduce.reduce.shuffle.connect.timeout*
 18  *dfs.safemode.extension*3
 *mapred.jobtracker.blacklist.fault-timeout-window*180
 *tasktracker.http.threads*40  *mapred.job.shuffle.merge.percent*0.66
 *mapreduce.inputformat.class*org.apache.giraph.bsp.BspInputFormat
 *fs.ftp.impl*org.apache.hadoop.fs.ftp.FTPFileSystem  *user.name
 http://user.name*hduser  *mapred.output.compress*false
 *io.bytes.per.checksum*512  *giraph.isStaticGraph*true
 *mapred.healthChecker.script.timeout*60
 *topology.node.switch.mapping.impl*
 org.apache.hadoop.net.ScriptBasedMapping
 *dfs.https.server.keystore.resource*ssl-server.xml
 *mapred.reduce.slowstart.completed.maps*0.05  *mapred.reduce.max.attempts*
 4  *fs.ramfs.impl*org.apache.hadoop.fs.InMemoryFileSystem
 *dfs.block.access.token.lifetime*600  *dfs.name.edits.dir*${dfs.name.dir}
 *mapred.skip.map.max.skip.records*0  *mapred.cluster.map.memory.mb*-1
 *hadoop.security.group.mapping*
 org.apache.hadoop.security.ShellBasedUnixGroupsMapping
 *mapred.job.tracker.persist.jobstatus.dir*/jobtracker/jobsInfo
 *mapred.jar*
 hdfs://localhost:54310/app/hadoop/tmp/mapred/staging/hduser/.staging/job_201312051827_0001/job.jar
 *dfs.block.size*67108864  *fs.s3.buffer.dir*${hadoop.tmp.dir}/s3
 *job.end.retry.attempts*0  *fs.file.impl*
 org.apache.hadoop.fs.LocalFileSystem  *mapred.local.dir.minspacestart*0
 *mapred.output.compression.type*RECORD  *dfs.datanode.ipc.address*
 0.0.0.0:50020  *dfs.permissions*true  *topology.script.number.args*100
 *io.mapfile.bloom.error.rate*0.005  *mapred.cluster.max.reduce.memory.mb*
 -1  *mapred.max.tracker.blacklists*4  *mapred.task.profile.maps*0-2
 *dfs.datanode.https.address*0.0.0.0:50475  *mapred.userlog.retain.hours*24
 *dfs.secondary.http.address*0.0.0.0:50090  *dfs.replication.max*512
 *mapred.job.tracker.persist.jobstatus.active*false
 *hadoop.security.authorization*false  *local.cache.size*10737418240
 *dfs.namenode.delegation.token.renew-interval*8640
 *mapred.min.split.size*0  *mapred.map.tasks*2  *mapred.child.java.opts*
 -Xmx4000m  *mapreduce.job.counters.limit*120
 *dfs.https.client.keystore.resource*ssl-client.xml  *mapred.job.queue.name
 http://mapred.job.queue.name*default  *dfs.https.address*0.0.0.0:50470
 

Re: out of core option

2013-12-05 Thread Rob Vesse
Sebastian

You've made a minor typo in the configuration setting which means you
haven't actually enabled out of core graph mode.

You have giraph.useOutOfCoreGiraph when it should be
giraph.useOutOfCoreGraph ­ note that the last word is Graph not Giraph

Rob

From:  Sebastian Stipkovic sebastian.stipko...@gmail.com
Reply-To:  user@giraph.apache.org
Date:  Thursday, 5 December 2013 20:39
To:  user@giraph.apache.org
Subject:  out of core option

 Hello,
 
 I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
 node cluster. It computes a tiny graph successful. But if the
 input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
 exception, although I had turned on the out-of-memory-option. The job with
 out-of-memory-option works only well with a tiny graph (0.9 GB).  What is
 Wrong? Does I have to do furthermore configurations?
 
 My Configurations are as follows:
 
 
 namevalue
 fs.s3n.implorg.apache.hadoop.fs.s3native.NativeS3FileSystem
 mapred.task.cache.levels2
 giraph.vertexOutputFormatClassorg.apache.giraph.examples.MyShortestPaths$MyOut
 putFormat
 hadoop.tmp.dir/app/hadoop/tmp
 hadoop.native.libtrue
 map.sort.classorg.apache.hadoop.util.QuickSort
 dfs.namenode.decommission.nodes.per.interval5
 dfs.https.need.client.authfalse
 ipc.client.idlethreshold4000
 dfs.datanode.data.dir.perm755
 mapred.system.dir${hadoop.tmp.dir}/mapred/system
 mapred.job.tracker.persist.jobstatus.hours0
 dfs.datanode.address0.0.0.0:50010 http://0.0.0.0:50010
 dfs.namenode.logging.levelinfo
 dfs.block.access.token.enablefalse
 io.skip.checksum.errorsfalse
 fs.default.name http://fs.default.name hdfs://localhost:54310
 mapred.cluster.reduce.memory.mb-1
 mapred.child.tmp./tmp
 fs.har.impl.disable.cachetrue
 dfs.safemode.threshold.pct0.999f
 mapred.skip.reduce.max.skip.groups0
 dfs.namenode.handler.count10
 dfs.blockreport.initialDelay0
 mapred.heartbeats.in.second100
 mapred.tasktracker.dns.nameserverdefault
 io.sort.factor10
 mapred.task.timeout60
 giraph.maxWorkers1
 mapred.max.tracker.failures4
 hadoop.rpc.socket.factory.class.defaultorg.apache.hadoop.net.StandardSocketFac
 tory
 mapred.job.tracker.jobhistory.lru.cache.size5
 fs.hdfs.implorg.apache.hadoop.hdfs.DistributedFileSystem
 mapred.queue.default.acl-administer-jobs*
 dfs.block.access.key.update.interval600
 mapred.skip.map.auto.incr.proc.counttrue
 mapreduce.job.complete.cancel.delegation.tokenstrue
 io.mapfile.bloom.size1048576
 mapreduce.reduce.shuffle.connect.timeout18
 dfs.safemode.extension3
 mapred.jobtracker.blacklist.fault-timeout-window180
 tasktracker.http.threads40
 mapred.job.shuffle.merge.percent0.66
 mapreduce.inputformat.classorg.apache.giraph.bsp.BspInputFormat
 fs.ftp.implorg.apache.hadoop.fs.ftp.FTPFileSystem
 user.name http://user.name hduser
 mapred.output.compressfalse
 io.bytes.per.checksum512
 giraph.isStaticGraphtrue
 mapred.healthChecker.script.timeout60
 topology.node.switch.mapping.implorg.apache.hadoop.net.ScriptBasedMapping
 dfs.https.server.keystore.resourcessl-server.xml
 mapred.reduce.slowstart.completed.maps0.05
 mapred.reduce.max.attempts4
 fs.ramfs.implorg.apache.hadoop.fs.InMemoryFileSystem
 dfs.block.access.token.lifetime600
 dfs.name.edits.dir${dfs.name.dir}
 mapred.skip.map.max.skip.records0
 mapred.cluster.map.memory.mb-1
 hadoop.security.group.mappingorg.apache.hadoop.security.ShellBasedUnixGroupsMa
 pping
 mapred.job.tracker.persist.jobstatus.dir/jobtracker/jobsInfo
 mapred.jarhdfs://localhost:54310/app/hadoop/tmp/mapred/staging/hduser/.staging
 /job_201312051827_0001/job.jar
 dfs.block.size67108864
 fs.s3.buffer.dir${hadoop.tmp.dir}/s3
 job.end.retry.attempts0
 fs.file.implorg.apache.hadoop.fs.LocalFileSystem
 mapred.local.dir.minspacestart0
 mapred.output.compression.typeRECORD
 dfs.datanode.ipc.address0.0.0.0:50020 http://0.0.0.0:50020
 dfs.permissionstrue
 topology.script.number.args100
 io.mapfile.bloom.error.rate0.005
 mapred.cluster.max.reduce.memory.mb-1
 mapred.max.tracker.blacklists4
 mapred.task.profile.maps0-2
 dfs.datanode.https.address0.0.0.0:50475 http://0.0.0.0:50475
 mapred.userlog.retain.hours24
 dfs.secondary.http.address0.0.0.0:50090 http://0.0.0.0:50090
 dfs.replication.max512
 mapred.job.tracker.persist.jobstatus.activefalse
 hadoop.security.authorizationfalse
 local.cache.size10737418240
 dfs.namenode.delegation.token.renew-interval8640
 mapred.min.split.size0
 mapred.map.tasks2
 mapred.child.java.opts-Xmx4000m
 mapreduce.job.counters.limit120
 dfs.https.client.keystore.resourcessl-client.xml
 mapred.job.queue.name http://mapred.job.queue.name default
 dfs.https.address0.0.0.0:50470 http://0.0.0.0:50470
 mapred.job.tracker.retiredjobs.cache.size1000
 dfs.balance.bandwidthPerSec1048576
 ipc.server.listen.queue.size128
 mapred.inmem.merge.threshold1000
 job.end.retry.interval3
 mapred.skip.attempts.to.start.skipping2
 fs.checkpoint.dir${hadoop.tmp.dir}/dfs/namesecondary
 mapred.reduce.tasks0
 mapred.merge.recordsBeforeProgress1
 mapred.userlog.limit.kb0

Re: out of core option

2013-12-05 Thread Sebastian Stipkovic
Hi Ameya,

thanks for the answer. My allocated memory was too high. My server has
altogether 4000M. I have turned the memory down to 2000M for each Mapper.

Now I have set both out of core options and get the following exception:

 2013-12-05 23:10:18,568 INFO org.apache.hadoop.mapred.JobTracker: Adding
task (MAP) 'attempt_201312052304_0001_m_01_0' to tip
task_201312052304_0001_m_01, for tracker 'tracker_hduser:localhost/
127.0.0.1:39793' 2013-12-05 23:10:27,645 INFO
org.apache.hadoop.mapred.TaskInProgress: Error from
attempt_201312052304_0001_m_01_0: java.lang.IllegalStateException: run:
Caught an unrecoverable exception waitFor: ExecutionException occurred
while waiting for
org.apache.giraph.utils.ProgressableUtils$FutureWaitable@62bf5822 at
org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:101) at
org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:763) at
org.apache.hadoop.mapred.MapTask.run(MapTask.java:369) at
org.apache.hadoop.mapred.Child$4.run(Child.java:259) at
java.security.AccessController.doPrivileged(Native Method) at
javax.security.auth.Subject.doAs(Subject.java:415) at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059)
at org.apache.hadoop.mapred.Child.main(Child.java:253) Caused by:
java.lang.IllegalStateException: waitFor: ExecutionException occurred while
waiting for
org.apache.giraph.utils.ProgressableUtils$FutureWaitable@62bf5822 at
org.apache.giraph.utils.ProgressableUtils.waitFor(ProgressableUtils.java:181)
at
org.apache.giraph.utils.ProgressableUtils.waitForever(ProgressableUtils.java:139)
at
org.apache.giraph.utils.ProgressableUtils.waitForever(ProgressableUtils.java:124)
at
org.apache.giraph.utils.ProgressableUtils.getFutureResult(ProgressableUtils.java:87)
at
org.apache.giraph.utils.ProgressableUtils.getResultsWithNCallables(ProgressableUtils.java:221)
at
org.apache.giraph.worker.BspServiceWorker.loadInputSplits(BspServiceWorker.java:281)
at
org.apache.giraph.worker.BspServiceWorker.loadVertices(BspServiceWorker.java:325)
at
org.apache.giraph.worker.BspServiceWorker.setup(BspServiceWorker.java:506)
at
org.apache.giraph.graph.GraphTaskManager.execute(GraphTaskManager.java:244)
at org.apache.giraph.graph.GraphMapper.run(GraphMapper.java:91) ... 7 more
Caused by: java.util.concurrent.ExecutionException:
java.lang.IllegalStateException: getOrCreatePartition: cannot retrieve
partition 0 at
java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:262) at
java.util.concurrent.FutureTask.get(FutureTask.java:119) at
org.apache.giraph.utils.ProgressableUtils$FutureWaitable.waitFor(ProgressableUtils.java:300)
at
org.apache.giraph.utils.ProgressableUtils.waitFor(ProgressableUtils.java:173)
... 16 more Caused by: java.lang.IllegalStateException:
getOrCreatePartition: cannot retrieve partition 0 at
org.apache.giraph.partition.DiskBackedPartitionStore.getOrCreatePartition(DiskBackedPartitionStore.java:243)
at
org.apache.giraph.comm.requests.SendWorkerVerticesRequest.doRequest(SendWorkerVerticesRequest.java:110)
at
org.apache.giraph.comm.netty.NettyWorkerClientRequestProcessor.doRequest(NettyWorkerClientRequestProcessor.java:482)
at
org.apache.giraph.comm.netty.NettyWorkerClientRequestProcessor.sendVertexRequest(NettyWorkerClientRequestProcessor.java:276)
at
org.apache.giraph.worker.VertexInputSplitsCallable.readInputSplit(VertexInputSplitsCallable.java:172)
at
org.apache.giraph.worker.InputSplitsCallable.loadInputSplit(InputSplitsCallable.java:267)
at
org.apache.giraph.worker.InputSplitsCallable.call(InputSplitsCallable.java:211)
at
org.apache.giraph.worker.InputSplitsCallable.call(InputSplitsCallable.java:60)
at
org.apache.giraph.utils.LogStacktraceCallable.call(LogStacktraceCallable.java:51)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at
java.util.concurrent.FutureTask.run(FutureTask.java:166) at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724) Caused by:
java.util.concurrent.ExecutionException: java.lang.NullPointerException at
java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:252) at
java.util.concurrent.FutureTask.get(FutureTask.java:111) at
org.apache.giraph.partition.DiskBackedPartitionStore.getOrCreatePartition(DiskBackedPartitionStore.java:228)
... 13 more Caused by: java.lang.NullPointerException at
org.apache.giraph.partition.DiskBackedPartitionStore$GetPartition.call(DiskBackedPartitionStore.java:692)
at
org.apache.giraph.partition.DiskBackedPartitionStore$GetPartition.call(DiskBackedPartitionStore.java:658)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at
java.util.concurrent.FutureTask.run(FutureTask.java:166) at
org.apache.giraph.partition.DiskBackedPartitionStore$DirectExecutorService.execute(DiskBackedPartitionStore.java:972)
at

Re: out of core option

2013-12-05 Thread Sebastian Stipkovic
 in the configuration setting which means you
 haven't actually enabled out of core graph mode.

 You have *giraph.useOutOfCoreGiraph *when it should be 
 *giraph.useOutOfCoreGraph
 *– note that the last word is Graph not Giraph

 Rob

 From: Sebastian Stipkovic sebastian.stipko...@gmail.com
 Reply-To: user@giraph.apache.org
 Date: Thursday, 5 December 2013 20:39
 To: user@giraph.apache.org
 Subject: out of core option

 Hello,

 I had setup giraph 1.1.0 with hadoop-0.20.203.0rc1  on a single
 node cluster. It computes a tiny graph successful. But if the
 input graph is huge (5 GB), I get an OutOfMemory(Garbage Collector)
 exception, although I had turned on the out-of-memory-option. The job with
 out-of-memory-option works only well with a tiny graph (0.9 GB).  What is
 Wrong? Does I have to do furthermore configurations?

 My Configurations are as follows:


 namevalue*fs.s3n.impl*org.apache.hadoop.fs.s3native.NativeS3FileSystem
 *mapred.task.cache.levels*2*giraph.vertexOutputFormatClass*
 org.apache.giraph.examples.MyShortestPaths$MyOutputFormat*hadoop.tmp.dir*
 /app/hadoop/tmp*hadoop.native.lib*true*map.sort.class*
 org.apache.hadoop.util.QuickSort
 *dfs.namenode.decommission.nodes.per.interval*5
 *dfs.https.need.client.auth*false*ipc.client.idlethreshold*4000
 *dfs.datanode.data.dir.perm*755*mapred.system.dir*
 ${hadoop.tmp.dir}/mapred/system
 *mapred.job.tracker.persist.jobstatus.hours*0*dfs.datanode.address*
 0.0.0.0:50010*dfs.namenode.logging.level*info
 *dfs.block.access.token.enable*false*io.skip.checksum.errors*false*fs.default.name
 http://fs.default.name*hdfs://localhost:54310
 *mapred.cluster.reduce.memory.mb*-1*mapred.child.tmp*./tmp
 *fs.har.impl.disable.cache*true*dfs.safemode.threshold.pct*0.999f
 *mapred.skip.reduce.max.skip.groups*0*dfs.namenode.handler.count*10
 *dfs.blockreport.initialDelay*0*mapred.heartbeats.in.second*100
 *mapred.tasktracker.dns.nameserver*default*io.sort.factor*10
 *mapred.task.timeout*60*giraph.maxWorkers*1
 *mapred.max.tracker.failures*4*hadoop.rpc.socket.factory.class.default*
 org.apache.hadoop.net.StandardSocketFactory
 *mapred.job.tracker.jobhistory.lru.cache.size*5*fs.hdfs.impl*
 org.apache.hadoop.hdfs.DistributedFileSystem
 *mapred.queue.default.acl-administer-jobs**
 *dfs.block.access.key.update.interval*600
 *mapred.skip.map.auto.incr.proc.count*true
 *mapreduce.job.complete.cancel.delegation.tokens*true
 *io.mapfile.bloom.size*1048576*mapreduce.reduce.shuffle.connect.timeout*
 18*dfs.safemode.extension*3
 *mapred.jobtracker.blacklist.fault-timeout-window*180
 *tasktracker.http.threads*40*mapred.job.shuffle.merge.percent*0.66
 *mapreduce.inputformat.class*org.apache.giraph.bsp.BspInputFormat
 *fs.ftp.impl*org.apache.hadoop.fs.ftp.FTPFileSystem*user.name
 http://user.name*hduser*mapred.output.compress*false
 *io.bytes.per.checksum*512*giraph.isStaticGraph*true
 *mapred.healthChecker.script.timeout*60
 *topology.node.switch.mapping.impl*
 org.apache.hadoop.net.ScriptBasedMapping
 *dfs.https.server.keystore.resource*ssl-server.xml
 *mapred.reduce.slowstart.completed.maps*0.05*mapred.reduce.max.attempts*4
 *fs.ramfs.impl*org.apache.hadoop.fs.InMemoryFileSystem
 *dfs.block.access.token.lifetime*600*dfs.name.edits.dir*${dfs.name.dir}
 *mapred.skip.map.max.skip.records*0*mapred.cluster.map.memory.mb*-1
 *hadoop.security.group.mapping*
 org.apache.hadoop.security.ShellBasedUnixGroupsMapping
 *mapred.job.tracker.persist.jobstatus.dir*/jobtracker/jobsInfo*mapred.jar*
 hdfs://localhost:54310
 /app/hadoop/tmp/mapred/staging/hduser/.staging/job_201312051827_0001/job.jar
 *dfs.block.size*67108864*fs.s3.buffer.dir*${hadoop.tmp.dir}/s3
 *job.end.retry.attempts*0*fs.file.impl*
 org.apache.hadoop.fs.LocalFileSystem*mapred.local.dir.minspacestart*0
 *mapred.output.compression.type*RECORD*dfs.datanode.ipc.address*
 0.0.0.0:50020*dfs.permissions*true*topology.script.number.args*100
 *io.mapfile.bloom.error.rate*0.005*mapred.cluster.max.reduce.memory.mb*-1
 *mapred.max.tracker.blacklists*4*mapred.task.profile.maps*0-2
 *dfs.datanode.https.address*0.0.0.0:50475*mapred.userlog.retain.hours*24
 *dfs.secondary.http.address*0.0.0.0:50090*dfs.replication.max*512
 *mapred.job.tracker.persist.jobstatus.active*false
 *hadoop.security.authorization*false*local.cache.size*10737418240
 *dfs.namenode.delegation.token.renew-interval*8640
 *mapred.min.split.size*0*mapred.map.tasks*2*mapred.child.java.opts*
 -Xmx4000m*mapreduce.job.counters.limit*120
 *dfs.https.client.keystore.resource*ssl-client.xml*mapred.job.queue.name
 http://mapred.job.queue.name*default*dfs.https.address*0.0.0.0:50470
 *mapred.job.tracker.retiredjobs.cache.size*1000
 *dfs.balance.bandwidthPerSec*1048576*ipc.server.listen.queue.size*128
 *mapred.inmem.merge.threshold*1000*job.end.retry.interval*3
 *mapred.skip.attempts.to.start.skipping*2*fs.checkpoint.dir*
 ${hadoop.tmp.dir}/dfs/namesecondary*mapred.reduce.tasks*0
 *mapred.merge.recordsBeforeProgress*1*mapred.userlog.limit.kb*0