I would remove your entire local Maven repo (~/.m2/repo in linux) and try
again. I'm able to compile sample code with your build.sbt and sbt
v.0.13.12.

-Don

On Mon, Nov 14, 2016 at 3:11 PM, Marco Mistroni <mmistr...@gmail.com> wrote:

> uhm.sorry...... still same issues. this is hte new version
>
> name := "SparkExamples"
> version := "1.0"
> scalaVersion := "2.11.8"
> val sparkVersion = "2.0.1"
>
> // Add a single dependency
> libraryDependencies += "junit" % "junit" % "4.8" % "test"
> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5",
>                             "org.slf4j" % "slf4j-simple" % "1.7.5",
>                             "org.clapper" %% "grizzled-slf4j" % "1.0.2")
> libraryDependencies += "org.apache.spark"%%"spark-core"   % sparkVersion
> libraryDependencies += "org.apache.spark"%%"spark-streaming"   %
> sparkVersion
> libraryDependencies += "org.apache.spark"%%"spark-mllib"   % sparkVersion
> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume-sink" %
> "2.0.1"
> libraryDependencies += "org.apache.spark"%%"spark-sql"   % sparkVersion
>
>
> resolvers += "softprops-maven" at "http://dl.bintray.com/
> content/softprops/maven"
>
> Still seeing these kinds of errors  which seems to lead to the fact that
> somehow sbt is getting confused..
>
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:2:
> object mllib is not a member of package org.apache.spark
> [error] import org.apache.spark.mllib.linalg.{ Vector, Vectors }
> [error]                         ^
> [error] 
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:3:
> object mllib is not a member of package org.apache.spark
> [error] import org.apache.spark.mllib.regression.LabeledPoint
> [error]                         ^
> [error] 
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:4:
> object classification is not a member of package org.apache.spark.ml
> [error] import org.apache.spark.ml.classification.{
> RandomForestClassifier, RandomForestClassificationModel }
> [error]                            ^
> [error] 
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:6:
> object feature is not a member of package org.apache.spark.ml
> [error] import org.apache.spark.ml.feature.{ StringIndexer, IndexToString,
> VectorIndexer, VectorAssembler }
> [error]                            ^
> [error] 
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:7:
> object evaluation is not a member of package org.apache.spark.ml
> [error] import org.apache.spark.ml.evaluation.{ RegressionEvaluator,
> MulticlassClassificationEvaluator }
> [error]                            ^
> [error] 
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:8:
> object classification is not a member of package org.apache.spark.ml
> [error] import org.apache.spark.ml.classification._
> [error]                            ^
> [error] 
> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:9:
> object tuning is not a member of package org.apache.spark.ml
> [error] import org.apache.spark.ml.tuning.{ CrossValidator,
> ParamGridBuilder }
> [error]                            ^
> [error] C:\Users\marco\SparkExamples\src\main\scala\
> DecisionTreeExampleML.scala:10: object tuning is not a member of package
> org.apache.spark.ml
> [error] import org.apache.spark.ml.tuning.{ ParamGridBuilder,
> TrainValidationSplit }
> [error]                            ^
> [error] C:\Users\marco\SparkExamples\src\main\scala\
> DecisionTreeExampleML.scala:16: object Pipeline is not a member of
> package org.apache.spark.ml
> [error] import org.apache.spark.ml.{ Pipeline, PipelineModel }
>
> any other hints?
>
> thanks and regarsd
>  marco
>
>
>
>
> On Sun, Nov 13, 2016 at 10:52 PM, Don Drake <dondr...@gmail.com> wrote:
>
>> I would upgrade your Scala version to 2.11.8 as Spark 2.0 uses Scala 2.11
>> by default.
>>
>> On Sun, Nov 13, 2016 at 3:01 PM, Marco Mistroni <mmistr...@gmail.com>
>> wrote:
>>
>>> HI all
>>>  i have a small Spark-based project which at the moment depends on jar
>>> from Spark 1.6.0
>>> The project has few Spark examples plus one which depends on Flume
>>> libraries
>>>
>>>
>>> I am attempting to move to Spark 2.0, but i am having issues with
>>> my dependencies
>>> The stetup below works fine when compiled against 1.6.0 dependencies
>>>
>>> name := "SparkExamples"
>>> version := "1.0"
>>> scalaVersion := "2.10.5"
>>> val sparkVersion = "1.6.0"
>>>
>>>
>>> // Add a single dependency
>>> libraryDependencies += "junit" % "junit" % "4.8" % "test"
>>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5",
>>>                             "org.slf4j" % "slf4j-simple" % "1.7.5",
>>>                             "org.clapper" %% "grizzled-slf4j" % "1.0.2")
>>> libraryDependencies += "org.apache.spark"%%"spark-core"   %
>>> sparkVersion
>>> libraryDependencies += "org.apache.spark"%%"spark-streaming"   %
>>> sparkVersion
>>> libraryDependencies += "org.apache.spark"%%"spark-mllib"   %
>>> sparkVersion
>>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume" %
>>> "1.3.0"
>>> libraryDependencies += "org.apache.spark"%%"spark-sql"   % sparkVersion
>>>
>>>
>>> resolvers += "softprops-maven" at "http://dl.bintray.com/content
>>> /softprops/maven"
>>>
>>>
>>>
>>> This is the build.sbt version for using Spark 2 dependencies
>>>
>>> name := "SparkExamples"
>>> version := "1.0"
>>> scalaVersion := "2.10.6"
>>> val sparkVersion = "2.0.1"
>>>
>>>
>>> // Add a single dependency
>>> libraryDependencies += "junit" % "junit" % "4.8" % "test"
>>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5",
>>>                             "org.slf4j" % "slf4j-simple" % "1.7.5",
>>>                             "org.clapper" %% "grizzled-slf4j" % "1.0.2")
>>> libraryDependencies += "org.apache.spark"%%"spark-core"   %
>>> sparkVersion
>>> libraryDependencies += "org.apache.spark"%%"spark-streaming"   %
>>> sparkVersion
>>> libraryDependencies += "org.apache.spark"%%"spark-mllib"   %
>>> sparkVersion
>>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume-sink"
>>> % "2.0.1"
>>> libraryDependencies += "org.apache.spark"%%"spark-sql"   % sparkVersion
>>> resolvers += "softprops-maven" at "http://dl.bintray.com/content
>>> /softprops/maven"
>>>
>>> but the sbt compile fails miserably...below few of the errors (it
>>> actually compiles like i forgot all the depencencies as it is complaining
>>> on all org.apache.spark.ml and mllib packages
>>>
>>> [warn] Multiple dependencies with the same organization/name but
>>> different versions. To avoid conflict, pick one version:
>>> [warn]  * org.apache.spark:spark-core_2.10:(1.6.1, 2.0.1)
>>> [warn]  * org.apache.spark:spark-streaming_2.10:(1.6.1, 2.0.1)
>>> [warn]  * org.apache.spark:spark-sql_2.10:(1.6.1, 2.0.1)
>>> [warn]  * org.apache.spark:spark-mllib_2.10:(1.6.1, 2.0.1)
>>> [info] Resolving org.scala-lang#scala-library;2.10.6 ...
>>> .........
>>> [warn]     * org.apache.spark:spark-mllib_2.10:1.6.1 -> 2.0.1
>>> [warn]     * org.apache.spark:spark-sql_2.10:1.6.1 -> 2.0.1
>>> [warn]     * org.apache.spark:spark-streaming_2.10:1.6.1 -> 2.0.1
>>> [warn]     * org.apache.spark:spark-core_2.10:1.6.1 -> 2.0.1
>>> [warn] Run 'evicted' to see detailed eviction warnings
>>> [info] Compiling 18 Scala sources to C:\Users\marco\SparkExamples\t
>>> arget\scala-2.10\classes...
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:2:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.linalg.{ Vector, Vectors }
>>> [error]                         ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:3:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.regression.LabeledPoint
>>> [error]                         ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:4:
>>> object classification is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.classification._
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:5:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.tree.DecisionTree
>>> [error]                         ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:6:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.tree.model.DecisionTreeModel
>>> [error]                         ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:7:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.util.MLUtils
>>> [error]                         ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:9:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.stat.{
>>> MultivariateStatisticalSummary, Statistics }
>>> [error]                         ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:10:
>>> object mllib is not a member of package org.apache.spark
>>> [error] import org.apache.spark.mllib.linalg.Vector
>>> ........
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:4:
>>> object classification is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.classification.{
>>> RandomForestClassifier, RandomForestClassificationModel }
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:6:
>>> object feature is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.feature.{ StringIndexer,
>>> IndexToString, VectorIndexer, VectorAssembler }
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:7:
>>> object evaluation is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.evaluation.{ RegressionEvaluator,
>>> MulticlassClassificationEvaluator }
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:8:
>>> object classification is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.classification._
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:9:
>>> object tuning is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.tuning.{ CrossValidator,
>>> ParamGridBuilder }
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:10:
>>> object tuning is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.tuning.{ ParamGridBuilder,
>>> TrainValidationSplit }
>>> [error]                            ^
>>> [error] 
>>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:16:
>>> object Pipeline is not a member of package org.apache.spark.ml
>>> [error] import org.apache.spark.ml.{ Pipeline, PipelineModel }
>>> [error]
>>> ....
>>> and many more
>>>
>>> Could anyone assist or pass me the right dependencies? Or am i using
>>> wrong resolvers?
>>>
>>> kr
>>>  marco
>>>
>>
>>
>>
>> --
>> Donald Drake
>> Drake Consulting
>> http://www.drakeconsulting.com/
>> https://twitter.com/dondrake <http://www.MailLaunder.com/>
>> 800-733-2143
>>
>
>


-- 
Donald Drake
Drake Consulting
http://www.drakeconsulting.com/
https://twitter.com/dondrake <http://www.MailLaunder.com/>
800-733-2143

Reply via email to