Re: java.io.NotSerializableException
Which class is not Serializable? I run shark0.9 has a similarity exception: java.io.NotSerializableException (java.io.NotSerializableException: shark.execution.ReduceKeyReduceSide) java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1183) java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1547) java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1508) java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1431) java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1177) java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:347) org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:28) org.apache.spark.storage.DiskBlockObjectWriter.write(BlockObjectWriter.scala:176) org.apache.spark.util.collection.ExternalAppendOnlyMap.spill(ExternalAppendOnlyMap.scala:191) org.apache.spark.util.collection.ExternalAppendOnlyMap.insert(ExternalAppendOnlyMap.scala:141) org.apache.spark.Aggregator.combineValuesByKey(Aggregator.scala:59) org.apache.hadoop.hive.ql.exec.GroupByPostShuffleOperator$$anonfun$7.apply(GroupByPostShuffleOperator.scala:225) org.apache.hadoop.hive.ql.exec.GroupByPostShuffleOperator$$anonfun$7.apply(GroupByPostShuffleOperator.scala:225) org.apache.spark.rdd.RDD$$anonfun$3.apply(RDD.scala:471) org.apache.spark.rdd.RDD$$anonfun$3.apply(RDD.scala:471) org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:34) org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:241) org.apache.spark.rdd.RDD.iterator(RDD.scala:232) org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:34) org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:241) org.apache.spark.rdd.RDD.iterator(RDD.scala:232) org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:34) org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:241) org.apache.spark.rdd.RDD.iterator(RDD.scala:232) org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109) org.apache.spark.scheduler.Task.run(Task.scala:53) org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mcV$sp(Executor.scala:213) org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.scala:49) org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:744) leosand...@gmail.com From: yaoxin Date: 2014-02-24 19:18 To: user Subject: java.io.NotSerializableException I got a error org.apache.spark.SparkException: Job aborted: Task not serializable: java.io.NotSerializableException: But the class it complains is a java lib class that I dependents on, that I can't change it to Serializable. Is there any method to work this around? I am using Spark 0.9, spark master using local[2] mode. -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/java-io-NotSerializableException-tp1973.html Sent from the Apache Spark User List mailing list archive at Nabble.com.
How could I set spark.scheduler.pool in the shark cli ?
Hi all How could I set the shark cli use different pools ? I set spark.scheduler.mode = fair and scheduler.file=xxx . and I see my scheduler pools in the appilication . but when I set spark.scheduler.pool=mypool in the shark cli , these stage still run in the default pool . THX leosand...@gmail.com
Can you help me ?
HI, I run a hql in hive set these params : set hive.exec.parallel=true; set hive.exec.dynamic.partition.mode=nonstrict; set hive.input.format=org.apache.hadoop.hive.ql.io.CombineHiveInputFormat; set mapred.max.split.size=1; set mapred.min.split.size.per.node=1; set mapred.min.split.size.per.rack=1; set tcl.name=cr_24hourdM.tcl; set mapred.queue.name=tcl2; set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat; set mapred.min.split.size=536870912; set hive.exec.reducers.max=239; set hive.exec.reducers.bytes.per.reducer=8000; Could I also set the same params in the cli when I run the hql in shark ? or I should set shark.exec.mode=hive ? What's the diff between the two mode ? If I just run a hql without catch table, could I change spark.storage.memoryFraction=0.1 or smaller ? THX ! leosand...@gmail.com
Re: squestion on using spark parallelism vs using num partitions in spark api
I think the parallelism param just control how many tasks could be run together in each work. it could't control how many tasks should be split . leosand...@gmail.com From: hussam_jar...@dell.com Date: 2014-01-14 09:17 To: user@spark.incubator.apache.org Subject: squestion on using spark parallelism vs using num partitions in spark api Hi, Using spark 0.8.1 … jave code running on 8 CPU with 16GRAM single node It’s looks like upon setting spark parallelism using System.setProperty(spark.default.parallelism, 24) before creating my spark context as described in http://spark.incubator.apache.org/docs/latest/tuning.html#level-of-parallelism has no effect on the default number of partitions that spark uses in its api’s like saveAsTextFile() . For example if I set spark.default.parallelism to 24, I was expecting 24 tasks to be invoked upon calling saveAsTextFile() but it’s not the case as I am seeing only 1 task get invoked If I set my RDD parallelize() to 2 as dataSetRDD = SparkDriver.getSparkContext().parallelize(mydata,2); then invoke dataSetRDD.saveAsTextFile(JavaRddFilePath); I am seeing 2 tasks get invoked even my spark.default.parallelism was set to 24 Can someone explain the above behavior? Thanks, Hussam
转发: some problems about shark on spark
leosand...@gmail.com 发件人: leosand...@gmail.com 发送时间: 2014-01-10 22:29 收件人: user; shark-users 主题: some problems about shark on spark HI ALL, How could I set the param MEMORY_ONLY_SER 、Spark.kryoserializer.buffer.mb 、 Spark.default.parallelism and Spark.worker.timeout when I run a shark query ? May I set other params in spark-env.sh or hive-site.xml instead ? or set name=value in the shark cli ? I have a shark query test : table a 38b ; table b 23b ; sql: select a.* , b.* from a join b on a.id = b.id ; it build three stages : stage1 has tow tasks: task1: rdd.HadoopRDD : input split table a 0+19 ; task2: rdd.HadoopRDD : input split table a 19+19; stage2 has two tasks: task1: rdd.HadoopRDD : input split table b 0+11 ; task2: rdd.HadoopRDD : input split table b 11+12; stage3 has one task: task1: just fetch map outputs for shuffle and write to hdfs path . Why these tables so small , but build two tasks to read it ? How could I control the reduce task nums in shark ? It seems compute by the biggest father RDD's partitions ? THX ! leosand...@gmail.com
some problems
Hi, I'm runing a shark sql, but I don't know How spark build these stages and what work it does in each stage . I waited the job 1 hour , it seems that the task 504 has some problem. there are some logs in the machine which runs task 504 : 14/01/09 20:22:55 INFO executor.Executor: Finished task ID 287 14/01/09 20:23:00 INFO executor.StandaloneExecutorBackend: Got assigned task 504 14/01/09 20:23:00 INFO executor.Executor: Running task ID 504 14/01/09 20:23:00 INFO executor.Executor: Its epoch is 2 14/01/09 20:23:00 INFO spark.MapOutputTracker: Updating epoch to 2 and clearing cache 14/01/09 20:23:00 INFO spark.MapOutputTracker: Don't have map outputs for shuffle 2, fetching them 14/01/09 20:23:00 INFO spark.MapOutputTracker: Doing the fetch; tracker actor = Actor[akka://spark@OCDC-DD-002:53976/user/MapOutputTracker] 14/01/09 20:23:00 INFO spark.MapOutputTracker: Got the output locations 14/01/09 20:23:00 INFO storage.BlockFetcherIterator$BasicBlockFetcherIterator: maxBytesInFlight: 50331648, minRequest: 10066329 14/01/09 20:23:00 INFO storage.BlockFetcherIterator$BasicBlockFetcherIterator: maxBytesInFlight: 50331648, minRequest: 10066329 14/01/09 20:23:00 INFO storage.BlockFetcherIterator$BasicBlockFetcherIterator: maxBytesInFlight: 50331648, minRequest: 10066329 14/01/09 20:23:00 INFO storage.BlockFetcherIterator$BasicBlockFetcherIterator: maxBytesInFlight: 50331648, minRequest: 10066329 14/01/09 20:23:00 INFO storage.BlockFetcherIterator$BasicBlockFetcherIterator: Getting 24 non-zero-bytes blocks out of 264 blocks 14/01/09 20:23:00 INFO storage.BlockFetcherIterator$BasicBlockFetcherIterator: Started 1 remote gets in 74 ms 14/01/09 20:23:00 INFO network.SendingConnection: Initiating connection to [OCDC-DATA-013/172.17.254.33:54930] 14/01/09 20:23:00 INFO network.SendingConnection: Connected to [OCDC-DATA-013/172.17.254.33:54930], 1 messages pending 14/01/09 20:23:00 INFO network.ConnectionManager: Accepted connection from [OCDC-DATA-013/172.17.254.33] 14/01/09 20:23:18 INFO network.SendingConnection: Initiating connection to [OCDC-DATA-012/172.17.254.32:45718] 14/01/09 20:23:18 INFO network.SendingConnection: Connected to [OCDC-DATA-012/172.17.254.32:45718], 1 messages pending 14/01/09 20:23:18 INFO network.ConnectionManager: Accepted connection from [OCDC-DATA-012/172.17.254.32] 14/01/09 20:23:46 INFO network.SendingConnection: Initiating connection to [OCDC-DATA-011/172.17.254.31:51067] 14/01/09 20:23:46 INFO network.SendingConnection: Connected to [OCDC-DATA-011/172.17.254.31:51067], 1 messages pending 14/01/09 20:23:46 INFO network.ConnectionManager: Accepted connection from [OCDC-DATA-011/172.17.254.31] 14/01/09 20:56:48 INFO storage.BlockManager: BlockManager reregistering with master 14/01/09 20:56:48 INFO storage.BlockManagerMaster: Trying to register BlockManager 14/01/09 20:56:48 INFO storage.BlockManagerMaster: Registered BlockManager 14/01/09 20:56:48 INFO storage.BlockManager: Reporting 117 blocks to the master. 14/01/09 21:04:43 INFO storage.BlockManager: BlockManager reregistering with master 14/01/09 21:04:43 INFO storage.BlockManagerMaster: Trying to register BlockManager 14/01/09 21:04:43 INFO storage.BlockManagerMaster: Registered BlockManager 14/01/09 21:04:43 INFO storage.BlockManager: Reporting 117 blocks to the master. 14/01/09 21:06:43 INFO storage.BlockManager: BlockManager reregistering with master 14/01/09 21:06:43 INFO storage.BlockManagerMaster: Trying to register BlockManager 14/01/09 21:06:43 INFO storage.BlockManagerMaster: Registered BlockManager 14/01/09 21:06:43 INFO storage.BlockManager: Reporting 117 blocks to the master. 14/01/09 21:11:48 INFO storage.BlockManager: BlockManager reregistering with master 14/01/09 21:11:48 INFO storage.BlockManagerMaster: Trying to register BlockManager 14/01/09 21:11:48 INFO storage.BlockManagerMaster: Registered BlockManager 14/01/09 21:11:48 INFO storage.BlockManager: Reporting 117 blocks to the master. 14/01/09 21:12:48 INFO storage.BlockManager: BlockManager reregistering with master 14/01/09 21:12:48 INFO storage.BlockManagerMaster: Trying to register BlockManager 14/01/09 21:12:48 INFO storage.BlockManagerMaster: Registered BlockManager 14/01/09 21:12:48 INFO storage.BlockManager: Reporting 117 blocks to the master. 14/01/09 21:13:51 INFO storage.BlockManager: BlockManager reregistering with master 14/01/09 21:13:51 INFO storage.BlockManagerMaster: Trying to register BlockManager can anyone give me the hint thank you ! leosand...@gmail.com
native-lzo / gpl lib
HI, I do a query from shark , it read a compress data from hdfs . but spark could't find the native-lzo lib . 14/01/08 22:58:21 ERROR executor.Executor: Exception in task ID 286 java.lang.RuntimeException: native-lzo library not available at com.hadoop.compression.lzo.LzoCodec.getDecompressorType(LzoCodec.java:175) at org.apache.hadoop.hive.ql.io.CodecPool.getDecompressor(CodecPool.java:122) at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1299) at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1139) at org.apache.hadoop.hive.ql.io.RCFile$Reader.init(RCFile.java:1118) at org.apache.hadoop.hive.ql.io.RCFileRecordReader.init(RCFileRecordReader.java:52) at org.apache.hadoop.hive.ql.io.RCFileInputFormat.getRecordReader(RCFileInputFormat.java:57) at org.apache.spark.rdd.HadoopRDD$$anon$1.init(HadoopRDD.scala:93) at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:83) at org.apache.spark.rdd.HadoopRDD.compute(HadoopRDD.scala:51) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:237) at org.apache.spark.rdd.RDD.iterator(RDD.scala:226) at org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:29) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:237) at org.apache.spark.rdd.RDD.iterator(RDD.scala:226) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:36) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:237) at org.apache.spark.rdd.RDD.iterator(RDD.scala:226) at org.apache.spark.rdd.UnionPartition.iterator(UnionRDD.scala:29) at org.apache.spark.rdd.UnionRDD.compute(UnionRDD.scala:69) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:237) at org.apache.spark.rdd.RDD.iterator(RDD.scala:226) at org.apache.spark.rdd.MapPartitionsWithIndexRDD.compute(MapPartitionsWithIndexRDD.scala:40) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:237) at org.apache.spark.rdd.RDD.iterator(RDD.scala:226) at org.apache.spark.rdd.MapPartitionsWithIndexRDD.compute(MapPartitionsWithIndexRDD.scala:40) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:237) at org.apache.spark.rdd.RDD.iterator(RDD.scala:226) at org.apache.spark.scheduler.ResultTask.run(ResultTask.scala:99) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:158) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) at java.lang.Thread.run(Thread.java:662) can anyone give me the hint thank you ! leosand...@gmail.com
Shark compile
Hi I want to compile Shark , Could I just run $ sbt/sbt package under shark_home ,without compile spark and install hive ? tharks ! leosand...@gmail.com
Re: RE: Errors with spark-0.8.1 hadoop-yarn 2.2.0
What is your classpath ? Had you builded your spark when you changed the new version and with yarn? Have you find your jar under the $SPARK_HOME/assembly/target/scala-2.9.3 ? or there is not just only one ? leosand...@gmail.com From: Liu, Raymond Date: 2013-12-30 08:36 To: user@spark.incubator.apache.org Subject: RE: Errors with spark-0.8.1 hadoop-yarn 2.2.0 Hi Izhar Is that the exact command you are running? Say with 0.8.0 instead of 0.8.1 in the cmd? Raymond Liu From: Izhar ul Hassan [mailto:ezh...@gmail.com] Sent: Friday, December 27, 2013 9:40 PM To: user@spark.incubator.apache.org Subject: Errors with spark-0.8.1 hadoop-yarn 2.2.0 Hi, I have a 3 node installation of hadoop 2.2.0 with yarn. I have installed spark-0.8.1 with support for spark enabled. I get the following errors when trying to run the examples: SPARK_JAR=./assembly/target/scala-2.9.3/spark-assembly-0.8.0-incubating-hadoop2.0.5-alpha.jar \ ./spark-class org.apache.spark.deploy.yarn.Client \ --jar examples/target/scala-2.9.3/spark-examples-assembly-0.8.0-incubating.jar \ --class org.apache.spark.examples.SparkPi \ --args yarn-standalone \ --num-workers 3 \ --master-memory 4g \ --worker-memory 2g \ --worker-cores 1 Exception in thread main java.lang.NoClassDefFoundError: org/apache/spark/deploy/yarn/Client Caused by: java.lang.ClassNotFoundException: org.apache.spark.deploy.yarn.Client at java.net.URLClassLoader$1.run(URLClassLoader.java:217) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:205) at java.lang.ClassLoader.loadClass(ClassLoader.java:323) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:294) at java.lang.ClassLoader.loadClass(ClassLoader.java:268) Could not find the main class: org.apache.spark.deploy.yarn.Client. Program will exit. spark-0.8.0 with hadooop 2.0.5-alpha works fine. -- /Izhar
Re: Re: endless job and slant tasks
No , just standalone cluster leosand...@gmail.com From: Azuryy Yu Date: 2013-12-25 19:21 To: user@spark.incubator.apache.org Subject: Re: endless job and slant tasks Hi Leo, Did you run Spark on Yarn or mesos? On Wed, Dec 25, 2013 at 6:58 PM, leosand...@gmail.com leosand...@gmail.com wrote: hi all : I run an example three times , it just read data from hdfs then do map and reduce then write to hdfs . the first time and second time it works well , read almost 7G data and finished in 15 minutes , but there have a problem when I run it the third time . one machine in my cluster lack of hard disk . The job begin at 17:11:15 , but it has been unable to end . I wait it for 1 hour then kill it . there are the logs : LogA (from the sick machine): . 13/12/25 17:13:56 INFO Executor: Its epoch is 0 13/12/25 17:13:56 ERROR Executor: Exception in task ID 26 java.lang.NullPointerException at org.apache.spark.storage.DiskStore$DiskBlockObjectWriter.revertPartialWrites(DiskStore.scala:99) at org.apache.spark.scheduler.ShuffleMapTask$$anonfun$run$2.apply(ShuffleMapTask.scala:175) at org.apache.spark.scheduler.ShuffleMapTask$$anonfun$run$2.apply(ShuffleMapTask.scala:175) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34) at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:38) at org.apache.spark.scheduler.ShuffleMapTask.run(ShuffleMapTask.scala:175) at org.apache.spark.scheduler.ShuffleMapTask.run(ShuffleMapTask.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:158) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) END LogB (a healthy machine): .. 13/12/25 17:12:35 INFO Executor: Serialized size of result for 54 is 817 13/12/25 17:12:35 INFO Executor: Finished task ID 54 [GC 4007686K-7782K(15073280K), 0.0287070 secs] END LogC(the master and a worker): ... [GC 4907439K-110393K(15457280K), 0.0533800 secs] 13/12/25 17:13:23 INFO Executor: Serialized size of result for 203 is 817 13/12/25 17:13:23 INFO Executor: Finished task ID 203 13/12/25 17:13:24 INFO Executor: Serialized size of result for 202 is 817 13/12/25 17:13:24 INFO Executor: Finished task ID 202 END I don't know why the job doesn't shut down ? the log message doesn't been writen when the job runs 2 minuts . why one machine assigned tasks so many more than others ? How could I get the job and task' status when I run a big job ? it looks like a black box ... leosand...@gmail.com
Re: Re: endless job and slant tasks
Yes , disk space is full of the whole machine . leosand...@gmail.com From: Matei Zaharia Date: 2013-12-26 01:50 To: user Subject: Re: endless job and slant tasks Does that machine maybe have a full disk drive, or no space in /tmp (where Spark stores local files by default)? On Dec 25, 2013, at 7:50 AM, leosand...@gmail.com wrote: No , just standalone cluster leosand...@gmail.com From: Azuryy Yu Date: 2013-12-25 19:21 To: user@spark.incubator.apache.org Subject: Re: endless job and slant tasks Hi Leo, Did you run Spark on Yarn or mesos? On Wed, Dec 25, 2013 at 6:58 PM, leosand...@gmail.com leosand...@gmail.com wrote: hi all : I run an example three times , it just read data from hdfs then do map and reduce then write to hdfs . the first time and second time it works well , read almost 7G data and finished in 15 minutes , but there have a problem when I run it the third time . one machine in my cluster lack of hard disk . The job begin at 17:11:15 , but it has been unable to end . I wait it for 1 hour then kill it . there are the logs : LogA (from the sick machine): . 13/12/25 17:13:56 INFO Executor: Its epoch is 0 13/12/25 17:13:56 ERROR Executor: Exception in task ID 26 java.lang.NullPointerException at org.apache.spark.storage.DiskStore$DiskBlockObjectWriter.revertPartialWrites(DiskStore.scala:99) at org.apache.spark.scheduler.ShuffleMapTask$$anonfun$run$2.apply(ShuffleMapTask.scala:175) at org.apache.spark.scheduler.ShuffleMapTask$$anonfun$run$2.apply(ShuffleMapTask.scala:175) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34) at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:38) at org.apache.spark.scheduler.ShuffleMapTask.run(ShuffleMapTask.scala:175) at org.apache.spark.scheduler.ShuffleMapTask.run(ShuffleMapTask.scala:88) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:158) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) END LogB (a healthy machine): .. 13/12/25 17:12:35 INFO Executor: Serialized size of result for 54 is 817 13/12/25 17:12:35 INFO Executor: Finished task ID 54 [GC 4007686K-7782K(15073280K), 0.0287070 secs] END LogC(the master and a worker): ... [GC 4907439K-110393K(15457280K), 0.0533800 secs] 13/12/25 17:13:23 INFO Executor: Serialized size of result for 203 is 817 13/12/25 17:13:23 INFO Executor: Finished task ID 203 13/12/25 17:13:24 INFO Executor: Serialized size of result for 202 is 817 13/12/25 17:13:24 INFO Executor: Finished task ID 202 END I don't know why the job doesn't shut down ? the log message doesn't been writen when the job runs 2 minuts . why one machine assigned tasks so many more than others ? How could I get the job and task' status when I run a big job ? it looks like a black box ... leosand...@gmail.com
How to set Akka frame size
Hi, everyone I have a question about the arg spark.akka.frameSize , it default value is 10m . I execute the JavaWordCount read data from hdfs , there is a 7G file . there is a oom error caused by some task result exceeded Akka frame size . but when I modify the arg 1G ,2G , 10G , it show me ERROR ClusterScheduler: Lost executor 1 on ocnosql84: remote Akka client shutdown 13/12/24 19:41:14 ERROR StandaloneExecutorBackend: Driver terminated or disconnected! Shutting down. Sometimes it show me different error info : [lh1@ocnosql84 src]$ java MyWordCount spark://ocnosql84:7077 hdfs://ocnosql76:8030/user/lh1/cdr_ismp_20130218 15000 1g 120 13/12/24 19:20:33 ERROR Client$ClientActor: Failed to connect to master org.jboss.netty.channel.ChannelPipelineException: Failed to initialize a pipeline. at org.jboss.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:209) at org.jboss.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:183) at akka.remote.netty.ActiveRemoteClient$$anonfun$connect$1.apply$mcV$sp(Client.scala:173) at akka.util.Switch.liftedTree1$1(LockUtil.scala:33) at akka.util.Switch.transcend(LockUtil.scala:32) at akka.util.Switch.switchOn(LockUtil.scala:55) at akka.remote.netty.ActiveRemoteClient.connect(Client.scala:158) at akka.remote.netty.NettyRemoteTransport.send(NettyRemoteSupport.scala:153) at akka.remote.RemoteActorRef.$bang(RemoteActorRefProvider.scala:247) at org.apache.spark.deploy.client.Client$ClientActor.preStart(Client.scala:61) at akka.actor.ActorCell.create$1(ActorCell.scala:508) at akka.actor.ActorCell.systemInvoke(ActorCell.scala:600) at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:209) at akka.dispatch.Mailbox.run(Mailbox.scala:178) at akka.dispatch.ForkJoinExecutorConfigurator$MailboxExecutionTask.exec(AbstractDispatcher.scala:516) at akka.jsr166y.ForkJoinTask.doExec(ForkJoinTask.java:259) at akka.jsr166y.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:975) at akka.jsr166y.ForkJoinPool.runWorker(ForkJoinPool.java:1479) at akka.jsr166y.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:104) Caused by: java.lang.IllegalArgumentException: maxFrameLength must be a positive integer: -1451229184 at org.jboss.netty.handler.codec.frame.LengthFieldBasedFrameDecoder.init(LengthFieldBasedFrameDecoder.java:270) at org.jboss.netty.handler.codec.frame.LengthFieldBasedFrameDecoder.init(LengthFieldBasedFrameDecoder.java:236) at akka.remote.netty.ActiveRemoteClientPipelineFactory.getPipeline(Client.scala:340) at org.jboss.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:207) ... 18 more 13/12/24 19:20:33 ERROR SparkDeploySchedulerBackend: Disconnected from Spark cluster! 13/12/24 19:20:33 ERROR ClusterScheduler: Exiting due to error from cluster scheduler: Disconnected from Spark cluster It seems caused by LengthFieldBasedFrameDecoder lenDec = new LengthFieldBasedFrameDecoder(this.client.netty().settings().MessageFrameSize(), 0, 4, 0, 4); I don't know what's the value of this.client.netty().settings().MessageFrameSize() and how to calculate this value . my spark args : export SPARK_DAEMON_MEMORY=4000m export SPARK_MEM=1000m export SPARK_WORKER_MEMORY=8g spark.akka.frameSize = 1000 / 2000 / 5000 / 1 / 15000 spark.executor.memory = 1g spark.akka.askTimeout = 120 Any help or reply is very appriciated ! Thanks very much leosand...@gmail.com
AbstractMethodError
I write a example MyWordCount , just set spark.akka.frameSize larger than default . but when I run this jar , there is a problem : 13/12/19 18:53:48 INFO ClusterTaskSetManager: Lost TID 0 (task 0.0:0) 13/12/19 18:53:48 INFO ClusterTaskSetManager: Loss was due to java.lang.AbstractMethodError java.lang.AbstractMethodError: org.apache.spark.api.java.function.WrappedFunction1.call(Ljava/lang/Object;)Ljava/lang/Object; at org.apache.spark.api.java.function.WrappedFunction1.apply(WrappedFunction1.scala:31) at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:90) at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:90) at scala.collection.Iterator$$anon$21.hasNext(Iterator.scala:440) at scala.collection.Iterator$class.foreach(Iterator.scala:772) at scala.collection.Iterator$$anon$21.foreach(Iterator.scala:437) at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:102) at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:250) at scala.collection.Iterator$$anon$21.toBuffer(Iterator.scala:437) at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:237) at scala.collection.Iterator$$anon$21.toArray(Iterator.scala:437) at org.apache.spark.rdd.RDD$$anonfun$1.apply(RDD.scala:560) at org.apache.spark.rdd.RDD$$anonfun$1.apply(RDD.scala:560) at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:758) at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:758) it caused by this code : JavaRDDString words = lines.flatMap(new FlatMapFunctionString, String() { public IterableString call(String s) { return Arrays.asList(s.split( )); } }); there is the parent class: private[spark] abstract class WrappedFunction1[T, R] extends AbstractFunction1[T, R] { @throws(classOf[Exception]) def call(t: T): R final def apply(t: T): R = call(t) } my code is same as the JavaWordCount , I don't know what's the error . Thanks Leo leosand...@gmail.com
AbstractMethodError
I write a example MyWordCount , just set spark.akka.frameSize larger than default . but when I run this jar , there is a problem : 13/12/19 18:53:48 INFO ClusterTaskSetManager: Lost TID 0 (task 0.0:0) 13/12/19 18:53:48 INFO ClusterTaskSetManager: Loss was due to java.lang.AbstractMethodError java.lang.AbstractMethodError: org.apache.spark.api.java.function.WrappedFunction1.call(Ljava/lang/Object;)Ljava/lang/Object; at org.apache.spark.api.java.function.WrappedFunction1.apply(WrappedFunction1.scala:31) at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:90) at org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:90) at scala.collection.Iterator$$anon$21.hasNext(Iterator.scala:440) at scala.collection.Iterator$class.foreach(Iterator.scala:772) at scala.collection.Iterator$$anon$21.foreach(Iterator.scala:437) at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:102) at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:250) at scala.collection.Iterator$$anon$21.toBuffer(Iterator.scala:437) at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:237) at scala.collection.Iterator$$anon$21.toArray(Iterator.scala:437) at org.apache.spark.rdd.RDD$$anonfun$1.apply(RDD.scala:560) at org.apache.spark.rdd.RDD$$anonfun$1.apply(RDD.scala:560) at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:758) at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:758) it caused by this code : JavaRDDString words = lines.flatMap(new FlatMapFunctionString, String() { public IterableString call(String s) { return Arrays.asList(s.split( )); } }); there is the parent class: private[spark] abstract class WrappedFunction1[T, R] extends AbstractFunction1[T, R] { @throws(classOf[Exception]) def call(t: T): R final def apply(t: T): R = call(t) } the code is same as the JavaWordCount , I don't know what's the error . Thanks Leo leosand...@gmail.com
resultset exceed Akka frame size
Hi, everyone I have a problem when I run the WordCount example. I read 6G data from hdfs , when I run collect(), the executer had died . there is the exception : 13/12/18 13:19:39 INFO ClusterTaskSetManager: Lost TID 55 (task 0.0:3) 13/12/18 13:19:39 INFO ClusterTaskSetManager: Loss was due to task 55 result exceeding Akka frame size; aborting job 13/12/18 13:19:39 INFO ClusterScheduler: Remove TaskSet 0.0 from pool 13/12/18 13:19:39 INFO DAGScheduler: Failed to run collect at JavaWordCount.java:60 Exception in thread main org.apache.spark.SparkException: Job failed: Task 55 result exceeded Akka frame size at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:760) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:758) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:758) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:379) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:441) at org.apache.spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:149) I saw there are some issues about this question in the github , it seems that if the middle resultset is larger than Akka frame size , the job will fail . I want to know if I can change some params to solve the problem ? Thanks Leo leosand...@gmail.com
Re: Re: resultset exceed Akka frame size
Thank you ! leosand...@gmail.com From: Azuryy Yu Date: 2013-12-18 17:29 To: user Subject: Re: resultset exceed Akka frame size Hi Leo, Akka is used to transfer the data back to the master, and there is a setting in Akka for the max message size, which is default to 10 MB here, you can find it at: core/src/main/scala/org/apache/spark/util/AkkaUtils.scala So just increase spark.akka.frameSize to a larger number. On Wed, Dec 18, 2013 at 4:49 PM, leosand...@gmail.com leosand...@gmail.com wrote: Hi, everyone I have a problem when I run the WordCount example. I read 6G data from hdfs , when I run collect(), the executer had died . there is the exception : 13/12/18 13:19:39 INFO ClusterTaskSetManager: Lost TID 55 (task 0.0:3) 13/12/18 13:19:39 INFO ClusterTaskSetManager: Loss was due to task 55 result exceeding Akka frame size; aborting job 13/12/18 13:19:39 INFO ClusterScheduler: Remove TaskSet 0.0 from pool 13/12/18 13:19:39 INFO DAGScheduler: Failed to run collect at JavaWordCount.java:60 Exception in thread main org.apache.spark.SparkException: Job failed: Task 55 result exceeded Akka frame size at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:760) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:758) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:758) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:379) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:441) at org.apache.spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:149) I saw there are some issues about this question in the github , it seems that if the middle resultset is larger than Akka frame size , the job will fail . I want to know if I can change some params to solve the problem ? Thanks Leo leosand...@gmail.com
转发: OOM
leosand...@gmail.com 发件人: leosand...@gmail.com 发送时间: 2013-12-16 20:01 收件人: user-subscribe 主题: OOM hello everyone, I have a problem when I run the wordcount example. I read data from hdfs , its almost 7G. I haven't seen the info from the web ui or sparkhome/work . This is the console info : . 13/12/16 19:48:02 INFO LocalTaskSetManager: Size of task 52 is 1834 bytes 13/12/16 19:48:02 INFO LocalScheduler: Running 52 13/12/16 19:48:02 INFO BlockFetcherIterator$BasicBlockFetcherIterator: Getting 52 non-zero-bytes blocks out of 52 blocks 13/12/16 19:48:02 INFO BlockFetcherIterator$BasicBlockFetcherIterator: Started 0 remote gets in 7 ms 13/12/16 19:48:09 INFO LocalTaskSetManager: Loss was due to java.lang.OutOfMemoryError java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOf(Arrays.java:2271) at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113) at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93) at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:140) at java.io.ObjectOutputStream$BlockDataOutputStream.drain(ObjectOutputStream.java:1857) at java.io.ObjectOutputStream$BlockDataOutputStream.setBlockDataMode(ObjectOutputStream.java:1766) at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1185) at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:346) at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:27) at org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:47) at org.apache.spark.scheduler.local.LocalScheduler.runTask(LocalScheduler.scala:204) at org.apache.spark.scheduler.local.LocalActor$$anonfun$launchTask$1$$anon$1.run(LocalScheduler.scala:68) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) 13/12/16 19:48:09 INFO LocalScheduler: Remove TaskSet 0.0 from pool 13/12/16 19:48:09 INFO DAGScheduler: Failed to run collect at console:17 org.apache.spark.SparkException: Job failed: Task 0.0:0 failed more than 4 times; aborting job java.lang.OutOfMemoryError: Java heap space at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:760) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:758) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:758) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:379) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:441) at org.apache.spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:149) this is my spark-env.sh : export SPARK_HOME=/home/lh1/spark_hadoopapp/spark-0.8.0-hadoop2.0.0-cdh4.2.1 export JAVA_HOME=/home/lh1/app/jdk1.7.0 export SCALA_HOME=/home/lh1/sparkapp/scala-2.9.3 export SPARK_WORKER_CORES=2 export SPARK_WORKER_MEMORY=1024m export SPARK_WORKER_INSTANCES=2 export SPARK_DAEMON_JAVA_OPTS=9000m I just started to use Spark , so can you give me some suggestions ? Thanks . Leo leosand...@gmail.com
OOM, help
hello everyone, I have a problem when I run the wordcount example. I read data from hdfs , its almost 7G. I haven't seen the info from the web ui or sparkhome/work . This is the console info : . 13/12/16 19:48:02 INFO LocalTaskSetManager: Size of task 52 is 1834 bytes 13/12/16 19:48:02 INFO LocalScheduler: Running 52 13/12/16 19:48:02 INFO BlockFetcherIterator$BasicBlockFetcherIterator: Getting 52 non-zero-bytes blocks out of 52 blocks 13/12/16 19:48:02 INFO BlockFetcherIterator$BasicBlockFetcherIterator: Started 0 remote gets in 7 ms 13/12/16 19:48:09 INFO LocalTaskSetManager: Loss was due to java.lang.OutOfMemoryError java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOf(Arrays.java:2271) at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:113) at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93) at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:140) at java.io.ObjectOutputStream$BlockDataOutputStream.drain(ObjectOutputStream.java:1857) at java.io.ObjectOutputStream$BlockDataOutputStream.setBlockDataMode(ObjectOutputStream.java:1766) at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1185) at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:346) at org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:27) at org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:47) at org.apache.spark.scheduler.local.LocalScheduler.runTask(LocalScheduler.scala:204) at org.apache.spark.scheduler.local.LocalActor$$anonfun$launchTask$1$$anon$1.run(LocalScheduler.scala:68) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) at java.lang.Thread.run(Thread.java:722) 13/12/16 19:48:09 INFO LocalScheduler: Remove TaskSet 0.0 from pool 13/12/16 19:48:09 INFO DAGScheduler: Failed to run collect at console:17 org.apache.spark.SparkException: Job failed: Task 0.0:0 failed more than 4 times; aborting job java.lang.OutOfMemoryError: Java heap space at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:760) at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:758) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:758) at org.apache.spark.scheduler.DAGScheduler.processEvent(DAGScheduler.scala:379) at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$run(DAGScheduler.scala:441) at org.apache.spark.scheduler.DAGScheduler$$anon$1.run(DAGScheduler.scala:149) this is my spark-env.sh : export SPARK_HOME=/home/lh1/spark_hadoopapp/spark-0.8.0-hadoop2.0.0-cdh4.2.1 export JAVA_HOME=/home/lh1/app/jdk1.7.0 export SCALA_HOME=/home/lh1/sparkapp/scala-2.9.3 export SPARK_WORKER_CORES=2 export SPARK_WORKER_MEMORY=1024m export SPARK_WORKER_INSTANCES=2 export SPARK_DAEMON_JAVA_OPTS=9000m I just started to use Spark , so can you give me some suggestions ? Thanks . Leo leosand...@gmail.com
Re: Re: I need some help
What will be cleaned if I compile Spark with sbt/sbt clean assembly? Actually I find there is a problem in my product's url sparkhome/assembly/target/scala-2.9.3 , there are two jars named spark-assembly-0.8.0-incubating-hadoop2.0.0-cdh4.2.1.jar and spark-assembly_2.9.3-0.8.0-incubating-hadoop2.0.0-mr1-cdh4.2.0.jar , and the compute-classpath.sh make the classpath with ASSEMBLY_JAR=`ls $FWDIR/assembly/target/scala-$SCALA_VERSION/spark-assembly*hadoop*.jar` , so the two jars splited with a space in the classpath . Somebody encounter the same problem ? leosand...@gmail.com From: Paco Nathan Date: 2013-12-12 04:25 To: user Subject: Re: I need some help did you try using: sbt/sbt clean assembly On Tue, Dec 10, 2013 at 10:23 PM, leosand...@gmail.com leosand...@gmail.com wrote: I have deployed two Spark clusters . The first is a simple standalone cluster which is working well . ( sbt/sbt assembly ) But I built Spark against Hadoop2.0.0-cdh4.2.1 in the second cluster, there seems to be a problem when I start the master ! ( SPARK_HADOOP_VERSION=2.0.0-cdh4.2.1 sbt/sbt assembly ) [lh1@ocnosql84 bin]$ start-master.sh starting org.apache.spark.deploy.master.Master, logging to /home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/bin/../logs/spark-lh1-org.apache.spark.deploy.master.Master-1-ocnosql84.out failed to launch org.apache.spark.deploy.master.Master: [Loaded java.lang.Shutdown from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] [Loaded java.lang.Shutdown$Lock from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] full log in /home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/bin/../logs/spark-lh1-org.apache.spark.deploy.master.Master-1-ocnosql84.out It seems that I failed to load the Master , but the class org.apache.spark.deploy.master.Master exists spark-0.8.0-incubating-bin-cdh4/assembly/target/scala-2.9.3/spark-assembly-0.8.0-incubating-hadoop2.0.0-cdh4.2.1.jar . I set the parameter SPARK_DAEMON_JAVA_OPTS = -verbose:class , there are some logs : Spark Command: /home/lh1/app/jdk1.7.0/bin/java -cp :/home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/conf:/home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/assembly/target/scala-2.9.3/spark-assembly-0.8.0-incubating-hadoop2.0.0-cdh4.2.1.jar /home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/assembly/target/scala-2.9.3/spark-assembly_2.9.3-0.8.0-incubating-hadoop2.0.0-mr1-cdh4.2.0.jar -verbose:class -Djava.library.path= -Xms512m -Xmx512m org.apache.spark.deploy.master.Master --ip ocnosql84 --port 7077 --webui-port 8080 .. [Loaded java.text.Format$Field from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] [Loaded java.text.MessageFormat$Field from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] Error: Could not find or load main class org.apache.spark.deploy.master.Master [Loaded java.lang.Shutdown from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] [Loaded java.lang.Shutdown$Lock from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] Thanks ! leosand...@gmail.com
Re: RE: I need some help
NO, I build with sbt leosand...@gmail.com From: Liu, Raymond Date: 2013-12-12 14:12 To: user@spark.incubator.apache.org Subject: RE: Re: I need some help The latter one sound to me like been built by mvn? Best Regards, Raymond Liu From: leosand...@gmail.com [mailto:leosand...@gmail.com] Sent: Thursday, December 12, 2013 2:02 PM To: user Subject: Re: Re: I need some help What will be cleaned if I compile Spark with sbt/sbt clean assembly? Actually I find there is a problem in my product's url sparkhome/assembly/target/scala-2.9.3 , there are two jars named spark-assembly-0.8.0-incubating-hadoop2.0.0-cdh4.2.1.jar and spark-assembly_2.9.3-0.8.0-incubating-hadoop2.0.0-mr1-cdh4.2.0.jar , and the compute-classpath.sh make the classpath with ASSEMBLY_JAR=`ls $FWDIR/assembly/target/scala-$SCALA_VERSION/spark-assembly*hadoop*.jar` , so the two jars splited with a space in the classpath . Somebody encounter the same problem ? leosand...@gmail.com From: Paco Nathan Date: 2013-12-12 04:25 To: user Subject: Re: I need some help did you try using: sbt/sbt clean assembly On Tue, Dec 10, 2013 at 10:23 PM, leosand...@gmail.com leosand...@gmail.com wrote: I have deployed two Spark clusters . The first is a simple standalone cluster which is working well . ( sbt/sbt assembly ) But I built Spark against Hadoop2.0.0-cdh4.2.1 in the second cluster, there seems to be a problem when I start the master ! ( SPARK_HADOOP_VERSION=2.0.0-cdh4.2.1 sbt/sbt assembly ) [lh1@ocnosql84 bin]$ start-master.sh starting org.apache.spark.deploy.master.Master, logging to /home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/bin/../logs/spark-lh1-org.apache.spark.deploy.master.Master-1-ocnosql84.out failed to launch org.apache.spark.deploy.master.Master: [Loaded java.lang.Shutdown from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] [Loaded java.lang.Shutdown$Lock from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] full log in /home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/bin/../logs/spark-lh1-org.apache.spark.deploy.master.Master-1-ocnosql84.out It seems that I failed to load the Master , but the class org.apache.spark.deploy.master.Master exists spark-0.8.0-incubating-bin-cdh4/assembly/target/scala-2.9.3/spark-assembly-0.8.0-incubating-hadoop2.0.0-cdh4.2.1.jar . I set the parameter SPARK_DAEMON_JAVA_OPTS = -verbose:class , there are some logs : Spark Command: /home/lh1/app/jdk1.7.0/bin/java -cp :/home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/conf:/home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/assembly/target/scala-2.9.3/spark-assembly-0.8.0-incubating-hadoop2.0.0-cdh4.2.1.jar /home/lh1/spark_hadoopapp/spark-0.8.0-incubating-bin-cdh4/assembly/target/scala-2.9.3/spark-assembly_2.9.3-0.8.0-incubating-hadoop2.0.0-mr1-cdh4.2.0.jar -verbose:class -Djava.library.path= -Xms512m -Xmx512m org.apache.spark.deploy.master.Master --ip ocnosql84 --port 7077 --webui-port 8080 .. [Loaded java.text.Format$Field from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] [Loaded java.text.MessageFormat$Field from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] Error: Could not find or load main class org.apache.spark.deploy.master.Master [Loaded java.lang.Shutdown from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] [Loaded java.lang.Shutdown$Lock from /home/lh1/app/jdk1.7.0/jre/lib/rt.jar] Thanks ! leosand...@gmail.com