Hi, I ran into in error trying to run spark-shell with an external package that I built and published locally using the spark-package sbt plugin ( https://github.com/databricks/sbt-spark-package).
To my understanding, spark packages can be published simply as maven artifacts, yet after running "publishLocal" in my package project ( https://github.com/jodersky/spark-paperui), the following command park-shell --packages ch.jodersky:spark-paperui-server_2.10:0.1-SNAPSHOT gives an error: :::::::::::::::::::::::::::::::::::::::::::::: :: UNRESOLVED DEPENDENCIES :: :::::::::::::::::::::::::::::::::::::::::::::: :: ch.jodersky#spark-paperui-server_2.10;0.1: configuration not found in ch.jodersky#spark-paperui-server_2.10;0.1: 'default'. It was required from org.apache.spark#spark-submit-parent;1.0 default :::::::::::::::::::::::::::::::::::::::::::::: :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: ch.jodersky#spark-paperui-server_2.10;0.1: configuration not found in ch.jodersky#spark-paperui-server_2.10;0.1: 'default'. It was required from org.apache.spark#spark-submit-parent;1.0 default] at org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:1011) at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:286) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:153) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:12 Do I need to include some default configuration? If so where and how should I do it? All other packages I looked at had no such thing. Btw, I am using spark-shell from a