This looks like your app is not finding your Typesafe config. The config
should usually be placed in particular folder under your app to be seen
correctly. If it's in a non-standard location you can
pass  -Dconfig.file=alt_reference1.conf to java to tell it where to look.
If this is a config that belogs to Spark and not your app, I'd recommend
running your jar via spark submit (that should run) and dump out the
classpath/variables that spark submit sets up...

On Tue, Jun 2, 2015 at 6:58 AM, Pa Rö <paul.roewer1...@googlemail.com>
wrote:

> hello community,
>
> i have build a jar file from my spark app with maven (mvn clean compile
> assembly:single) and the following pom file:
>
> <project xmlns="http://maven.apache.org/POM/4.0.0"; xmlns:xsi="
> http://www.w3.org/2001/XMLSchema-instance";
>   xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
> http://maven.apache.org/xsd/maven-4.0.0.xsd";>
>   <modelVersion>4.0.0</modelVersion>
>
>   <groupId>mgm.tp.bigdata</groupId>
>   <artifactId>ma-spark</artifactId>
>   <version>0.0.1-SNAPSHOT</version>
>   <packaging>jar</packaging>
>
>   <name>ma-spark</name>
>   <url>http://maven.apache.org</url>
>
>   <properties>
>     <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
>   </properties>
>
>   <repositories>
>     <repository>
>       <id>cloudera</id>
>       <url>https://repository.cloudera.com/artifactory/cloudera-repos/
> </url>
>     </repository>
>   </repositories>
>
>   <dependencies>
>     <dependency>
>       <groupId>junit</groupId>
>       <artifactId>junit</artifactId>
>       <version>3.8.1</version>
>       <scope>test</scope>
>     </dependency>
>     <dependency>
>         <groupId>org.apache.spark</groupId>
>         <artifactId>spark-core_2.10</artifactId>
>         <version>1.1.0-cdh5.2.5</version>
>     </dependency>
>     <dependency>
>         <groupId>mgm.tp.bigdata</groupId>
>         <artifactId>ma-commons</artifactId>
>         <version>0.0.1-SNAPSHOT</version>
>     </dependency>
>   </dependencies>
>
>   <build>
>   <plugins>
>     <plugin>
>       <artifactId>maven-assembly-plugin</artifactId>
>       <configuration>
>         <archive>
>           <manifest>
>             <mainClass>mgm.tp.bigdata.ma_spark.SparkMain</mainClass>
>           </manifest>
>         </archive>
>         <descriptorRefs>
>           <descriptorRef>jar-with-dependencies</descriptorRef>
>         </descriptorRefs>
>       </configuration>
>     </plugin>
>   </plugins>
> </build>
> </project>
>
> if i run my app with  java -jar
> ma-spark-0.0.1-SNAPSHOT-jar-with-dependencies.jar on terminal, i get the
> following error message:
>
> proewer@proewer-VirtualBox:~/Schreibtisch$ java -jar
> ma-spark-0.0.1-SNAPSHOT-jar-with-dependencies.jar
> 2015-Jun-02 12:53:36,348 [main] org.apache.spark.util.Utils
>  WARN  - Your hostname, proewer-VirtualBox resolves to a loopback address:
> 127.0.1.1; using 10.0.2.15 instead (on interface eth0)
> 2015-Jun-02 12:53:36,350 [main] org.apache.spark.util.Utils
>  WARN  - Set SPARK_LOCAL_IP if you need to bind to another address
> 2015-Jun-02 12:53:36,401 [main] org.apache.spark.SecurityManager
>  INFO  - Changing view acls to: proewer
> 2015-Jun-02 12:53:36,402 [main] org.apache.spark.SecurityManager
>  INFO  - Changing modify acls to: proewer
> 2015-Jun-02 12:53:36,403 [main] org.apache.spark.SecurityManager
>  INFO  - SecurityManager: authentication disabled; ui acls disabled; users
> with view permissions: Set(proewer); users with modify permissions:
> Set(proewer)
> Exception in thread "main" com.typesafe.config.ConfigException$Missing: No
> configuration setting found for key 'akka.version'
>     at com.typesafe.config.impl.SimpleConfig.findKey(SimpleConfig.java:115)
>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:136)
>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:142)
>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:150)
>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:155)
>     at
> com.typesafe.config.impl.SimpleConfig.getString(SimpleConfig.java:197)
>     at akka.actor.ActorSystem$Settings.<init>(ActorSystem.scala:136)
>     at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:470)
>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:111)
>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:104)
>     at
> org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121)
>     at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:54)
>     at org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53)
>     at
> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1454)
>     at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141)
>     at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1450)
>     at
> org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:56)
>     at org.apache.spark.SparkEnv$.create(SparkEnv.scala:156)
>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:203)
>     at
> org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:53)
>     at mgm.tp.bigdata.ma_spark.SparkMain.main(SparkMain.java:38)
>
> what i do wrong?
>
> best regards,
> paul
>
>

Reply via email to