The EC2 versiĆ³n is 1.1.0 and this is my build.sbt:

libraryDependencies ++= Seq(
  jdbc,
  anorm,
  cache,
  "org.apache.spark"  %% "spark-core"              % "1.1.0",
  "com.typesafe.akka" %% "akka-actor"              % "2.2.3",
  "com.typesafe.akka" %% "akka-slf4j"              % "2.2.3",
  "org.apache.spark"  %% "spark-streaming-twitter" % "1.1.0",
  "org.apache.spark"  %% "spark-sql"               % "1.1.0",
  "org.apache.spark"  %% "spark-mllib"             % "1.1.0"
  )



On Sun, Jan 11, 2015 at 3:01 AM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> What is your spark version that is running on the EC2 cluster? From the build
> file <https://github.com/knoldus/Play-Spark-Scala/blob/master/build.sbt>
> of your play application it seems that it uses Spark 1.0.1.
>
> Thanks
> Best Regards
>
> On Fri, Jan 9, 2015 at 7:17 PM, Eduardo Cusa <
> eduardo.c...@usmediaconsulting.com> wrote:
>
>> Hi guys, I running the following example :
>> https://github.com/knoldus/Play-Spark-Scala in the same machine as the
>> spark master, and the spark cluster was lauched with ec2 script.
>>
>>
>> I'm stuck with this errors, any idea how to fix it?
>>
>> Regards
>> Eduardo
>>
>>
>> call the play app prints the following exception :
>>
>>
>> [*error*] a.r.EndpointWriter - AssociationError 
>> [akka.tcp://sparkDriver@ip-10-28-236-122.ec2.internal:47481] <- 
>> [akka.tcp://driverPropsFetcher@ip-10-158-18-250.ec2.internal:52575]: Error 
>> [Shut down address: 
>> akka.tcp://driverPropsFetcher@ip-10-158-18-250.ec2.internal:52575] [
>> akka.remote.ShutDownAssociation: Shut down address: 
>> akka.tcp://driverPropsFetcher@ip-10-158-18-250.ec2.internal:52575
>> Caused by: akka.remote.transport.Transport$InvalidAssociationException: The 
>> remote system terminated the association because it is shutting down.
>>
>>
>>
>>
>> The master recive the spark application and generate the following stderr
>> log page:
>>
>>
>> 15/01/09 13:31:23 INFO Remoting: Remoting started; listening on addresses 
>> :[akka.tcp://sparkExecutor@ip-10-158-18-250.ec2.internal:37856]
>> 15/01/09 13:31:23 INFO Remoting: Remoting now listens on addresses: 
>> [akka.tcp://sparkExecutor@ip-10-158-18-250.ec2.internal:37856]
>> 15/01/09 13:31:23 INFO util.Utils: Successfully started service 
>> 'sparkExecutor' on port 37856.
>> 15/01/09 13:31:23 INFO util.AkkaUtils: Connecting to MapOutputTracker: 
>> akka.tcp://sparkDriver@ip-10-28-236-122.ec2.internal:47481/user/MapOutputTracker
>> 15/01/09 13:31:23 INFO util.AkkaUtils: Connecting to BlockManagerMaster: 
>> akka.tcp://sparkDriver@ip-10-28-236-122.ec2.internal:47481/user/BlockManagerMaster
>> 15/01/09 13:31:23 INFO storage.DiskBlockManager: Created local directory at 
>> /mnt/spark/spark-local-20150109133123-3805
>> 15/01/09 13:31:23 INFO storage.DiskBlockManager: Created local directory at 
>> /mnt2/spark/spark-local-20150109133123-b05e
>> 15/01/09 13:31:23 INFO util.Utils: Successfully started service 'Connection 
>> manager for block manager' on port 36936.
>> 15/01/09 13:31:23 INFO network.ConnectionManager: Bound socket to port 36936 
>> with id = ConnectionManagerId(ip-10-158-18-250.ec2.internal,36936)
>> 15/01/09 13:31:23 INFO storage.MemoryStore: MemoryStore started with 
>> capacity 265.4 MB
>> 15/01/09 13:31:23 INFO storage.BlockManagerMaster: Trying to register 
>> BlockManager
>> 15/01/09 13:31:23 INFO storage.BlockManagerMaster: Registered BlockManager
>> 15/01/09 13:31:23 INFO util.AkkaUtils: Connecting to HeartbeatReceiver: 
>> akka.tcp://sparkDriver@ip-10-28-236-122.ec2.internal:47481/user/HeartbeatReceiver
>> 15/01/09 13:31:54 ERROR executor.CoarseGrainedExecutorBackend: Driver 
>> Disassociated [akka.tcp://sparkExecutor@ip-10-158-18-250.ec2.internal:57671] 
>> -> [akka.tcp://sparkDriver@ip-10-28-236-122.ec2.internal:47481] 
>> disassociated! Shutting down.
>>
>>
>

Reply via email to