This looks like your app is not finding your Typesafe config. The config should usually be placed in particular folder under your app to be seen correctly. If it's in a non-standard location you can pass -Dconfig.file=alt_reference1.conf to java to tell it where to look. If this is a config that belogs to Spark and not your app, I'd recommend running your jar via spark submit (that should run) and dump out the classpath/variables that spark submit sets up...
On Tue, Jun 2, 2015 at 6:58 AM, Pa Rö <paul.roewer1...@googlemail.com> wrote: > hello community, > > i have build a jar file from my spark app with maven (mvn clean compile > assembly:single) and the following pom file: > > <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi=" > http://www.w3.org/2001/XMLSchema-instance" > xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 > http://maven.apache.org/xsd/maven-4.0.0.xsd"> > <modelVersion>4.0.0</modelVersion> > > <groupId>mgm.tp.bigdata</groupId> > <artifactId>ma-spark</artifactId> > <version>0.0.1-SNAPSHOT</version> > <packaging>jar</packaging> > > <name>ma-spark</name> > <url>http://maven.apache.org</url> > > <properties> > <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> > </properties> > > <repositories> > <repository> > <id>cloudera</id> > <url>https://repository.cloudera.com/artifactory/cloudera-repos/ > </url> > </repository> > </repositories> > > <dependencies> > <dependency> > <groupId>junit</groupId> > <artifactId>junit</artifactId> > <version>3.8.1</version> > <scope>test</scope> > </dependency> > <dependency> > <groupId>org.apache.spark</groupId> > <artifactId>spark-core_2.10</artifactId> > <version>1.1.0-cdh5.2.5</version> > </dependency> > <dependency> > <groupId>mgm.tp.bigdata</groupId> > <artifactId>ma-commons</artifactId> > <version>0.0.1-SNAPSHOT</version> > </dependency> > </dependencies> > > <build> > <plugins> > <plugin> > <artifactId>maven-assembly-plugin</artifactId> > <configuration> > <archive> > <manifest> > <mainClass>mgm.tp.bigdata.ma_spark.SparkMain</mainClass> > </manifest> > </archive> > <descriptorRefs> > <descriptorRef>jar-with-dependencies</descriptorRef> > </descriptorRefs> > </configuration> > </plugin> > </plugins> > </build> > </project> > > if i run my app with java -jar > ma-spark-0.0.1-SNAPSHOT-jar-with-dependencies.jar on terminal, i get the > following error message: > > proewer@proewer-VirtualBox:~/Schreibtisch$ java -jar > ma-spark-0.0.1-SNAPSHOT-jar-with-dependencies.jar > 2015-Jun-02 12:53:36,348 [main] org.apache.spark.util.Utils > WARN - Your hostname, proewer-VirtualBox resolves to a loopback address: > 127.0.1.1; using 10.0.2.15 instead (on interface eth0) > 2015-Jun-02 12:53:36,350 [main] org.apache.spark.util.Utils > WARN - Set SPARK_LOCAL_IP if you need to bind to another address > 2015-Jun-02 12:53:36,401 [main] org.apache.spark.SecurityManager > INFO - Changing view acls to: proewer > 2015-Jun-02 12:53:36,402 [main] org.apache.spark.SecurityManager > INFO - Changing modify acls to: proewer > 2015-Jun-02 12:53:36,403 [main] org.apache.spark.SecurityManager > INFO - SecurityManager: authentication disabled; ui acls disabled; users > with view permissions: Set(proewer); users with modify permissions: > Set(proewer) > Exception in thread "main" com.typesafe.config.ConfigException$Missing: No > configuration setting found for key 'akka.version' > at com.typesafe.config.impl.SimpleConfig.findKey(SimpleConfig.java:115) > at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:136) > at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:142) > at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:150) > at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:155) > at > com.typesafe.config.impl.SimpleConfig.getString(SimpleConfig.java:197) > at akka.actor.ActorSystem$Settings.<init>(ActorSystem.scala:136) > at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:470) > at akka.actor.ActorSystem$.apply(ActorSystem.scala:111) > at akka.actor.ActorSystem$.apply(ActorSystem.scala:104) > at > org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121) > at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:54) > at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53) > at > org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1454) > at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141) > at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1450) > at > org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:56) > at org.apache.spark.SparkEnv$.create(SparkEnv.scala:156) > at org.apache.spark.SparkContext.<init>(SparkContext.scala:203) > at > org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:53) > at mgm.tp.bigdata.ma_spark.SparkMain.main(SparkMain.java:38) > > what i do wrong? > > best regards, > paul > >