In your spark-env, did you set LD_LIBRARY_PATH ? Cheers
On Wed, Nov 25, 2015 at 7:32 AM, Oriol López Massaguer < oriol.lo...@gmail.com> wrote: > Hello; > > I'm trying to use a native library in Spark. > > I was using a simple standalone cluster with one master and worker. > > According to the documentation I edited the spark-defautls.conf by setting: > > spark.driver.extraClassPath=/opt/eTOX_spark/lib/org.RDKit.jar > spark.driver.extraLibraryPath=/opt/eTOX_spark/lib/ > spark.executor.extraLibraryPath=/opt/eTOX_spark/lib/ > > In the path /opt/eTOX_spark/lib/ there are 3 so files wich are wrapped in > org.RDKit.jar. > > But when I try so submit a job that uses the native library I get: > > Exception in thread "main" java.lang.UnsatisfiedLinkError: > org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Ljava/lang/String;)J > at org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Native Method) > at org.RDKit.RWMol.MolFromSmiles(RWMol.java:426) > at models.spark.sources.eTOX_DB$.main(eTOX.scala:54) > at models.spark.sources.eTOX_DB.main(eTOX.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:727) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > I use the submit.sh with the following parameters: > > /opt/spark/bin/spark-submit --verbose --class > "models.spark.sources.eTOX_DB" --master > spark://localhost.localdomain:7077 > target/scala-2.10/etox_spark_2.10-1.0.jar > > the full output is: > > Using properties file: /opt/spark/conf/spark-defaults.conf > Adding default property: spark.driver.extraLibraryPath=/opt/eTOX_spark/lib/ > Adding default property: > spark.driver.extraClassPath=/opt/eTOX_spark/lib/org.RDKit.jar > Adding default property: > spark.executor.extraLibraryPath=/opt/eTOX_spark/lib/ > Parsed arguments: > master spark://localhost.localdomain:7077 > deployMode null > executorMemory null > executorCores null > totalExecutorCores null > propertiesFile /opt/spark/conf/spark-defaults.conf > driverMemory null > driverCores null > driverExtraClassPath /opt/eTOX_spark/lib/org.RDKit.jar > driverExtraLibraryPath /opt/eTOX_spark/lib/ > driverExtraJavaOptions null > supervise false > queue null > numExecutors null > files null > pyFiles null > archives null > mainClass models.spark.sources.eTOX_DB > primaryResource > file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar > name models.spark.sources.eTOX_DB > childArgs [] > jars null > packages null > packagesExclusions null > repositories null > verbose true > > Spark properties used, including those specified through > --conf and those from the properties file > /opt/spark/conf/spark-defaults.conf: > spark.executor.extraLibraryPath -> /opt/eTOX_spark/lib/ > spark.driver.extraLibraryPath -> /opt/eTOX_spark/lib/ > spark.driver.extraClassPath -> /opt/eTOX_spark/lib/org.RDKit.jar > > > Main class: > models.spark.sources.eTOX_DB > Arguments: > > System properties: > spark.executor.extraLibraryPath -> /opt/eTOX_spark/lib/ > spark.driver.extraLibraryPath -> /opt/eTOX_spark/lib/ > SPARK_SUBMIT -> true > spark.app.name -> models.spark.sources.eTOX_DB > spark.jars -> > file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar > spark.submit.deployMode -> client > spark.master -> spark://localhost.localdomain:7077 > spark.driver.extraClassPath -> /opt/eTOX_spark/lib/org.RDKit.jar > Classpath elements: > file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar > > > Buffer(/opt/jdk1.8.0_45/jre/lib/amd64/libzip.so) > Loading libraries > Buffer(/opt/jdk1.8.0_45/jre/lib/amd64/libzip.so, /opt/eTOX_spark/lib/ > libboost_thread.1.48.0.so, /opt/eTOX_spark/lib/libboost_system.1.48.0.so, > /opt/eTOX_spark/lib/libGraphMolWrap.so) > Loading libraries > Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties > 15/11/25 16:27:32 INFO SparkContext: Running Spark version 1.6.0-SNAPSHOT > 15/11/25 16:27:33 WARN NativeCodeLoader: Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > 15/11/25 16:27:33 WARN Utils: Your hostname, localhost.localdomain > resolves to a loopback address: 127.0.0.1; using 10.0.2.15 instead (on > interface enp0s3) > 15/11/25 16:27:33 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to > another address > 15/11/25 16:27:33 INFO SecurityManager: Changing view acls to: user > 15/11/25 16:27:33 INFO SecurityManager: Changing modify acls to: user > 15/11/25 16:27:33 INFO SecurityManager: SecurityManager: authentication > disabled; ui acls disabled; users with view permissions: Set(user); users > with modify permissions: Set(user) > 15/11/25 16:27:34 INFO Utils: Successfully started service 'sparkDriver' > on port 35799. > 15/11/25 16:27:34 INFO Slf4jLogger: Slf4jLogger started > 15/11/25 16:27:34 INFO Remoting: Starting remoting > 15/11/25 16:27:34 INFO Remoting: Remoting started; listening on addresses > :[akka.tcp://sparkDriverActorSystem@10.0.2.15:48344] > 15/11/25 16:27:34 INFO Utils: Successfully started service > 'sparkDriverActorSystem' on port 48344. > 15/11/25 16:27:34 INFO SparkEnv: Registering MapOutputTracker > 15/11/25 16:27:34 INFO SparkEnv: Registering BlockManagerMaster > 15/11/25 16:27:34 INFO DiskBlockManager: Created local directory at > /tmp/blockmgr-ffbf1759-2e79-4ecd-a0d8-5d7f28d3c132 > 15/11/25 16:27:34 INFO MemoryStore: MemoryStore started with capacity > 736.1 MB > 15/11/25 16:27:34 INFO HttpFileServer: HTTP File server directory is > /tmp/spark-3f269496-c3aa-4bbc-8955-1ab88a179420/httpd-4ea1a9f9-2ce6-41b3-82ef-96e9094f5aa4 > 15/11/25 16:27:34 INFO HttpServer: Starting HTTP Server > 15/11/25 16:27:35 INFO Utils: Successfully started service 'HTTP file > server' on port 50803. > 15/11/25 16:27:35 INFO SparkEnv: Registering OutputCommitCoordinator > 15/11/25 16:27:35 INFO Utils: Successfully started service 'SparkUI' on > port 4040. > 15/11/25 16:27:35 INFO SparkUI: Started SparkUI at http://10.0.2.15:4040 > 15/11/25 16:27:35 INFO SparkContext: Added JAR > file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar at > http://10.0.2.15:50803/jars/etox_spark_2.10-1.0.jar with timestamp > 1448465255275 > 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Connecting to master > spark://localhost.localdomain:7077... > 15/11/25 16:27:35 INFO SparkDeploySchedulerBackend: Connected to Spark > cluster with app ID app-20151125162735-0003 > 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Executor added: > app-20151125162735-0003/0 on worker-20151125161159-10.0.2.15-48408 ( > 10.0.2.15:48408) with 4 cores > 15/11/25 16:27:35 INFO SparkDeploySchedulerBackend: Granted executor ID > app-20151125162735-0003/0 on hostPort 10.0.2.15:48408 with 4 cores, > 1024.0 MB RAM > 15/11/25 16:27:35 INFO Utils: Successfully started service > 'org.apache.spark.network.netty.NettyBlockTransferService' on port 35769. > 15/11/25 16:27:35 INFO NettyBlockTransferService: Server created on 35769 > 15/11/25 16:27:35 INFO BlockManagerMaster: Trying to register BlockManager > 15/11/25 16:27:35 INFO BlockManagerMasterEndpoint: Registering block > manager 10.0.2.15:35769 with 736.1 MB RAM, BlockManagerId(driver, > 10.0.2.15, 35769) > 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Executor updated: > app-20151125162735-0003/0 is now LOADING > 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Executor updated: > app-20151125162735-0003/0 is now RUNNING > 15/11/25 16:27:35 INFO BlockManagerMaster: Registered BlockManager > 15/11/25 16:27:35 INFO SparkDeploySchedulerBackend: SchedulerBackend is > ready for scheduling beginning after reached minRegisteredResourcesRatio: > 0.0 > 15/11/25 16:27:36 INFO ParquetRelation: Listing > file:/opt/etox_reports_2015_1/spark_data/etox/allfindings.parquet on driver > 15/11/25 16:27:37 INFO SparkContext: Starting job: load at eTOX.scala:52 > 15/11/25 16:27:37 INFO DAGScheduler: Got job 0 (load at eTOX.scala:52) > with 2 output partitions > 15/11/25 16:27:37 INFO DAGScheduler: Final stage: ResultStage 0 (load at > eTOX.scala:52) > 15/11/25 16:27:37 INFO DAGScheduler: Parents of final stage: List() > 15/11/25 16:27:37 INFO DAGScheduler: Missing parents: List() > 15/11/25 16:27:37 INFO DAGScheduler: Submitting ResultStage 0 > (MapPartitionsRDD[1] at load at eTOX.scala:52), which has no missing parents > 15/11/25 16:27:37 INFO MemoryStore: Ensuring 1048576 bytes of free space > for block broadcast_0(free: 771883008, max: 771883008) > 15/11/25 16:27:37 INFO MemoryStore: Ensuring 63000 bytes of free space for > block broadcast_0(free: 771883008, max: 771883008) > 15/11/25 16:27:37 INFO MemoryStore: Block broadcast_0 stored as values in > memory (estimated size 61.5 KB, free 61.5 KB) > 15/11/25 16:27:37 INFO MemoryStore: Ensuring 21090 bytes of free space for > block broadcast_0_piece0(free: 771820008, max: 771883008) > 15/11/25 16:27:37 INFO MemoryStore: Block broadcast_0_piece0 stored as > bytes in memory (estimated size 20.6 KB, free 82.1 KB) > 15/11/25 16:27:37 INFO BlockManagerInfo: Added broadcast_0_piece0 in > memory on 10.0.2.15:35769 (size: 20.6 KB, free: 736.1 MB) > 15/11/25 16:27:37 INFO SparkContext: Created broadcast 0 from broadcast at > DAGScheduler.scala:1003 > 15/11/25 16:27:37 INFO DAGScheduler: Submitting 2 missing tasks from > ResultStage 0 (MapPartitionsRDD[1] at load at eTOX.scala:52) > 15/11/25 16:27:37 INFO TaskSchedulerImpl: Adding task set 0.0 with 2 tasks > 15/11/25 16:27:39 INFO SparkDeploySchedulerBackend: Registered executor > NettyRpcEndpointRef(null) (10.0.2.15:50472) with ID 0 > 15/11/25 16:27:39 INFO TaskSetManager: Starting task 0.0 in stage 0.0 (TID > 0, 10.0.2.15, partition 0,PROCESS_LOCAL, 2137 bytes) > 15/11/25 16:27:39 INFO TaskSetManager: Starting task 1.0 in stage 0.0 (TID > 1, 10.0.2.15, partition 1,PROCESS_LOCAL, 2285 bytes) > 15/11/25 16:27:39 INFO BlockManagerMasterEndpoint: Registering block > manager 10.0.2.15:39373 with 736.1 MB RAM, BlockManagerId(0, 10.0.2.15, > 39373) > 15/11/25 16:27:39 INFO BlockManagerInfo: Added broadcast_0_piece0 in > memory on 10.0.2.15:39373 (size: 20.6 KB, free: 736.1 MB) > 15/11/25 16:27:40 INFO TaskSetManager: Finished task 0.0 in stage 0.0 (TID > 0) in 1097 ms on 10.0.2.15 (1/2) > 15/11/25 16:27:42 INFO TaskSetManager: Finished task 1.0 in stage 0.0 (TID > 1) in 3050 ms on 10.0.2.15 (2/2) > 15/11/25 16:27:42 INFO DAGScheduler: ResultStage 0 (load at eTOX.scala:52) > finished in 4.739 s > 15/11/25 16:27:42 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose tasks > have all completed, from pool > 15/11/25 16:27:42 INFO DAGScheduler: Job 0 finished: load at > eTOX.scala:52, took 5.157662 s > 15/11/25 16:27:42 INFO MemoryStore: Ensuring 1048576 bytes of free space > for block broadcast_1(free: 771798918, max: 771883008) > 15/11/25 16:27:42 INFO MemoryStore: Ensuring 63280 bytes of free space for > block broadcast_1(free: 771798918, max: 771883008) > 15/11/25 16:27:42 INFO MemoryStore: Block broadcast_1 stored as values in > memory (estimated size 61.8 KB, free 143.9 KB) > 15/11/25 16:27:42 INFO MemoryStore: Ensuring 19788 bytes of free space for > block broadcast_1_piece0(free: 771735638, max: 771883008) > 15/11/25 16:27:42 INFO MemoryStore: Block broadcast_1_piece0 stored as > bytes in memory (estimated size 19.3 KB, free 163.2 KB) > 15/11/25 16:27:42 INFO BlockManagerInfo: Added broadcast_1_piece0 in > memory on 10.0.2.15:35769 (size: 19.3 KB, free: 736.1 MB) > 15/11/25 16:27:42 INFO SparkContext: Created broadcast 1 from count at > eTOX.scala:53 > 15/11/25 16:27:42 INFO deprecation: mapred.min.split.size is deprecated. > Instead, use mapreduce.input.fileinputformat.split.minsize > 15/11/25 16:27:43 INFO ParquetRelation: Reading Parquet file(s) from > file:/opt/etox_reports_2015_1/spark_data/etox/allfindings.parquet/part-r-00000-a0018ffe-9aaf-4446-a718-79097835d08d.gz.parquet > 15/11/25 16:27:43 INFO SparkContext: Starting job: count at eTOX.scala:53 > 15/11/25 16:27:43 INFO DAGScheduler: Registering RDD 4 (count at > eTOX.scala:53) > 15/11/25 16:27:43 INFO DAGScheduler: Got job 1 (count at eTOX.scala:53) > with 1 output partitions > 15/11/25 16:27:43 INFO DAGScheduler: Final stage: ResultStage 2 (count at > eTOX.scala:53) > 15/11/25 16:27:43 INFO DAGScheduler: Parents of final stage: > List(ShuffleMapStage 1) > 15/11/25 16:27:43 INFO DAGScheduler: Missing parents: List(ShuffleMapStage > 1) > 15/11/25 16:27:43 INFO DAGScheduler: Submitting ShuffleMapStage 1 > (MapPartitionsRDD[4] at count at eTOX.scala:53), which has no missing > parents > 15/11/25 16:27:43 INFO MemoryStore: Ensuring 1048576 bytes of free space > for block broadcast_2(free: 771715850, max: 771883008) > 15/11/25 16:27:43 INFO MemoryStore: Ensuring 12840 bytes of free space for > block broadcast_2(free: 771715850, max: 771883008) > 15/11/25 16:27:43 INFO MemoryStore: Block broadcast_2 stored as values in > memory (estimated size 12.5 KB, free 175.8 KB) > 15/11/25 16:27:43 INFO MemoryStore: Ensuring 6175 bytes of free space for > block broadcast_2_piece0(free: 771703010, max: 771883008) > 15/11/25 16:27:43 INFO MemoryStore: Block broadcast_2_piece0 stored as > bytes in memory (estimated size 6.0 KB, free 181.8 KB) > 15/11/25 16:27:43 INFO BlockManagerInfo: Added broadcast_2_piece0 in > memory on 10.0.2.15:35769 (size: 6.0 KB, free: 736.1 MB) > 15/11/25 16:27:43 INFO SparkContext: Created broadcast 2 from broadcast at > DAGScheduler.scala:1003 > 15/11/25 16:27:43 INFO DAGScheduler: Submitting 1 missing tasks from > ShuffleMapStage 1 (MapPartitionsRDD[4] at count at eTOX.scala:53) > 15/11/25 16:27:43 INFO TaskSchedulerImpl: Adding task set 1.0 with 1 tasks > 15/11/25 16:27:43 INFO TaskSetManager: Starting task 0.0 in stage 1.0 (TID > 2, 10.0.2.15, partition 0,PROCESS_LOCAL, 2300 bytes) > 15/11/25 16:27:43 INFO BlockManagerInfo: Added broadcast_2_piece0 in > memory on 10.0.2.15:39373 (size: 6.0 KB, free: 736.1 MB) > 15/11/25 16:27:43 INFO BlockManagerInfo: Added broadcast_1_piece0 in > memory on 10.0.2.15:39373 (size: 19.3 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO TaskSetManager: Finished task 0.0 in stage 1.0 (TID > 2) in 1000 ms on 10.0.2.15 (1/1) > 15/11/25 16:27:44 INFO TaskSchedulerImpl: Removed TaskSet 1.0, whose tasks > have all completed, from pool > 15/11/25 16:27:44 INFO DAGScheduler: ShuffleMapStage 1 (count at > eTOX.scala:53) finished in 1.003 s > 15/11/25 16:27:44 INFO DAGScheduler: looking for newly runnable stages > 15/11/25 16:27:44 INFO DAGScheduler: running: Set() > 15/11/25 16:27:44 INFO DAGScheduler: waiting: Set(ResultStage 2) > 15/11/25 16:27:44 INFO DAGScheduler: failed: Set() > 15/11/25 16:27:44 INFO DAGScheduler: Submitting ResultStage 2 > (MapPartitionsRDD[7] at count at eTOX.scala:53), which has no missing > parents > 15/11/25 16:27:44 INFO MemoryStore: Ensuring 1048576 bytes of free space > for block broadcast_3(free: 771696835, max: 771883008) > 15/11/25 16:27:44 INFO MemoryStore: Ensuring 14032 bytes of free space for > block broadcast_3(free: 771696835, max: 771883008) > 15/11/25 16:27:44 INFO MemoryStore: Block broadcast_3 stored as values in > memory (estimated size 13.7 KB, free 195.5 KB) > 15/11/25 16:27:44 INFO MemoryStore: Ensuring 6698 bytes of free space for > block broadcast_3_piece0(free: 771682803, max: 771883008) > 15/11/25 16:27:44 INFO MemoryStore: Block broadcast_3_piece0 stored as > bytes in memory (estimated size 6.5 KB, free 202.1 KB) > 15/11/25 16:27:44 INFO BlockManagerInfo: Added broadcast_3_piece0 in > memory on 10.0.2.15:35769 (size: 6.5 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO SparkContext: Created broadcast 3 from broadcast at > DAGScheduler.scala:1003 > 15/11/25 16:27:44 INFO DAGScheduler: Submitting 1 missing tasks from > ResultStage 2 (MapPartitionsRDD[7] at count at eTOX.scala:53) > 15/11/25 16:27:44 INFO TaskSchedulerImpl: Adding task set 2.0 with 1 tasks > 15/11/25 16:27:44 INFO TaskSetManager: Starting task 0.0 in stage 2.0 (TID > 3, 10.0.2.15, partition 0,NODE_LOCAL, 2060 bytes) > 15/11/25 16:27:44 INFO BlockManagerInfo: Added broadcast_3_piece0 in > memory on 10.0.2.15:39373 (size: 6.5 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO MapOutputTrackerMasterEndpoint: Asked to send map > output locations for shuffle 0 to 10.0.2.15:50472 > 15/11/25 16:27:44 INFO MapOutputTrackerMaster: Size of output statuses for > shuffle 0 is 138 bytes > 15/11/25 16:27:44 INFO TaskSetManager: Finished task 0.0 in stage 2.0 (TID > 3) in 190 ms on 10.0.2.15 (1/1) > 15/11/25 16:27:44 INFO TaskSchedulerImpl: Removed TaskSet 2.0, whose tasks > have all completed, from pool > 15/11/25 16:27:44 INFO DAGScheduler: ResultStage 2 (count at > eTOX.scala:53) finished in 0.195 s > 15/11/25 16:27:44 INFO DAGScheduler: Job 1 finished: count at > eTOX.scala:53, took 1.249107 s > Num Findings: 2775141 > Exception in thread "main" java.lang.UnsatisfiedLinkError: > org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Ljava/lang/String;)J > at org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Native Method) > at org.RDKit.RWMol.MolFromSmiles(RWMol.java:426) > at models.spark.sources.eTOX_DB$.main(eTOX.scala:54) > at models.spark.sources.eTOX_DB.main(eTOX.scala) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:727) > at > org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > 15/11/25 16:27:44 INFO SparkContext: Invoking stop() from shutdown hook > 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_3_piece0 on > 10.0.2.15:35769 in memory (size: 6.5 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_3_piece0 on > 10.0.2.15:39373 in memory (size: 6.5 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO ContextCleaner: Cleaned accumulator 12 > 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_2_piece0 on > 10.0.2.15:35769 in memory (size: 6.0 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_2_piece0 on > 10.0.2.15:39373 in memory (size: 6.0 KB, free: 736.1 MB) > 15/11/25 16:27:44 INFO SparkUI: Stopped Spark web UI at > http://10.0.2.15:4040 > 15/11/25 16:27:44 INFO ContextCleaner: Cleaned accumulator 11 > 15/11/25 16:27:44 INFO ContextCleaner: Cleaned shuffle 0 > 15/11/25 16:27:44 INFO DAGScheduler: Stopping DAGScheduler > 15/11/25 16:27:44 INFO SparkDeploySchedulerBackend: Shutting down all > executors > 15/11/25 16:27:44 INFO SparkDeploySchedulerBackend: Asking each executor > to shut down > 15/11/25 16:27:44 INFO MapOutputTrackerMasterEndpoint: > MapOutputTrackerMasterEndpoint stopped! > 15/11/25 16:27:44 INFO MemoryStore: MemoryStore cleared > 15/11/25 16:27:44 INFO BlockManager: BlockManager stopped > 15/11/25 16:27:44 INFO BlockManagerMaster: BlockManagerMaster stopped > 15/11/25 16:27:44 INFO > OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: > OutputCommitCoordinator stopped! > 15/11/25 16:27:44 INFO RemoteActorRefProvider$RemotingTerminator: Shutting > down remote daemon. > 15/11/25 16:27:44 INFO SparkContext: Successfully stopped SparkContext > 15/11/25 16:27:44 INFO ShutdownHookManager: Shutdown hook called > 15/11/25 16:27:44 INFO RemoteActorRefProvider$RemotingTerminator: Remote > daemon shut down; proceeding with flushing remote transports. > 15/11/25 16:27:44 INFO ShutdownHookManager: Deleting directory > /tmp/spark-3f269496-c3aa-4bbc-8955-1ab88a179420 > > > > Oriol. >