I would remove your entire local Maven repo (~/.m2/repo in linux) and try again. I'm able to compile sample code with your build.sbt and sbt v.0.13.12.
-Don On Mon, Nov 14, 2016 at 3:11 PM, Marco Mistroni <mmistr...@gmail.com> wrote: > uhm.sorry...... still same issues. this is hte new version > > name := "SparkExamples" > version := "1.0" > scalaVersion := "2.11.8" > val sparkVersion = "2.0.1" > > // Add a single dependency > libraryDependencies += "junit" % "junit" % "4.8" % "test" > libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5", > "org.slf4j" % "slf4j-simple" % "1.7.5", > "org.clapper" %% "grizzled-slf4j" % "1.0.2") > libraryDependencies += "org.apache.spark"%%"spark-core" % sparkVersion > libraryDependencies += "org.apache.spark"%%"spark-streaming" % > sparkVersion > libraryDependencies += "org.apache.spark"%%"spark-mllib" % sparkVersion > libraryDependencies += "org.apache.spark"%%"spark-streaming-flume-sink" % > "2.0.1" > libraryDependencies += "org.apache.spark"%%"spark-sql" % sparkVersion > > > resolvers += "softprops-maven" at "http://dl.bintray.com/ > content/softprops/maven" > > Still seeing these kinds of errors which seems to lead to the fact that > somehow sbt is getting confused.. > > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:2: > object mllib is not a member of package org.apache.spark > [error] import org.apache.spark.mllib.linalg.{ Vector, Vectors } > [error] ^ > [error] > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:3: > object mllib is not a member of package org.apache.spark > [error] import org.apache.spark.mllib.regression.LabeledPoint > [error] ^ > [error] > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:4: > object classification is not a member of package org.apache.spark.ml > [error] import org.apache.spark.ml.classification.{ > RandomForestClassifier, RandomForestClassificationModel } > [error] ^ > [error] > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:6: > object feature is not a member of package org.apache.spark.ml > [error] import org.apache.spark.ml.feature.{ StringIndexer, IndexToString, > VectorIndexer, VectorAssembler } > [error] ^ > [error] > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:7: > object evaluation is not a member of package org.apache.spark.ml > [error] import org.apache.spark.ml.evaluation.{ RegressionEvaluator, > MulticlassClassificationEvaluator } > [error] ^ > [error] > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:8: > object classification is not a member of package org.apache.spark.ml > [error] import org.apache.spark.ml.classification._ > [error] ^ > [error] > C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:9: > object tuning is not a member of package org.apache.spark.ml > [error] import org.apache.spark.ml.tuning.{ CrossValidator, > ParamGridBuilder } > [error] ^ > [error] C:\Users\marco\SparkExamples\src\main\scala\ > DecisionTreeExampleML.scala:10: object tuning is not a member of package > org.apache.spark.ml > [error] import org.apache.spark.ml.tuning.{ ParamGridBuilder, > TrainValidationSplit } > [error] ^ > [error] C:\Users\marco\SparkExamples\src\main\scala\ > DecisionTreeExampleML.scala:16: object Pipeline is not a member of > package org.apache.spark.ml > [error] import org.apache.spark.ml.{ Pipeline, PipelineModel } > > any other hints? > > thanks and regarsd > marco > > > > > On Sun, Nov 13, 2016 at 10:52 PM, Don Drake <dondr...@gmail.com> wrote: > >> I would upgrade your Scala version to 2.11.8 as Spark 2.0 uses Scala 2.11 >> by default. >> >> On Sun, Nov 13, 2016 at 3:01 PM, Marco Mistroni <mmistr...@gmail.com> >> wrote: >> >>> HI all >>> i have a small Spark-based project which at the moment depends on jar >>> from Spark 1.6.0 >>> The project has few Spark examples plus one which depends on Flume >>> libraries >>> >>> >>> I am attempting to move to Spark 2.0, but i am having issues with >>> my dependencies >>> The stetup below works fine when compiled against 1.6.0 dependencies >>> >>> name := "SparkExamples" >>> version := "1.0" >>> scalaVersion := "2.10.5" >>> val sparkVersion = "1.6.0" >>> >>> >>> // Add a single dependency >>> libraryDependencies += "junit" % "junit" % "4.8" % "test" >>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5", >>> "org.slf4j" % "slf4j-simple" % "1.7.5", >>> "org.clapper" %% "grizzled-slf4j" % "1.0.2") >>> libraryDependencies += "org.apache.spark"%%"spark-core" % >>> sparkVersion >>> libraryDependencies += "org.apache.spark"%%"spark-streaming" % >>> sparkVersion >>> libraryDependencies += "org.apache.spark"%%"spark-mllib" % >>> sparkVersion >>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume" % >>> "1.3.0" >>> libraryDependencies += "org.apache.spark"%%"spark-sql" % sparkVersion >>> >>> >>> resolvers += "softprops-maven" at "http://dl.bintray.com/content >>> /softprops/maven" >>> >>> >>> >>> This is the build.sbt version for using Spark 2 dependencies >>> >>> name := "SparkExamples" >>> version := "1.0" >>> scalaVersion := "2.10.6" >>> val sparkVersion = "2.0.1" >>> >>> >>> // Add a single dependency >>> libraryDependencies += "junit" % "junit" % "4.8" % "test" >>> libraryDependencies ++= Seq("org.slf4j" % "slf4j-api" % "1.7.5", >>> "org.slf4j" % "slf4j-simple" % "1.7.5", >>> "org.clapper" %% "grizzled-slf4j" % "1.0.2") >>> libraryDependencies += "org.apache.spark"%%"spark-core" % >>> sparkVersion >>> libraryDependencies += "org.apache.spark"%%"spark-streaming" % >>> sparkVersion >>> libraryDependencies += "org.apache.spark"%%"spark-mllib" % >>> sparkVersion >>> libraryDependencies += "org.apache.spark"%%"spark-streaming-flume-sink" >>> % "2.0.1" >>> libraryDependencies += "org.apache.spark"%%"spark-sql" % sparkVersion >>> resolvers += "softprops-maven" at "http://dl.bintray.com/content >>> /softprops/maven" >>> >>> but the sbt compile fails miserably...below few of the errors (it >>> actually compiles like i forgot all the depencencies as it is complaining >>> on all org.apache.spark.ml and mllib packages >>> >>> [warn] Multiple dependencies with the same organization/name but >>> different versions. To avoid conflict, pick one version: >>> [warn] * org.apache.spark:spark-core_2.10:(1.6.1, 2.0.1) >>> [warn] * org.apache.spark:spark-streaming_2.10:(1.6.1, 2.0.1) >>> [warn] * org.apache.spark:spark-sql_2.10:(1.6.1, 2.0.1) >>> [warn] * org.apache.spark:spark-mllib_2.10:(1.6.1, 2.0.1) >>> [info] Resolving org.scala-lang#scala-library;2.10.6 ... >>> ......... >>> [warn] * org.apache.spark:spark-mllib_2.10:1.6.1 -> 2.0.1 >>> [warn] * org.apache.spark:spark-sql_2.10:1.6.1 -> 2.0.1 >>> [warn] * org.apache.spark:spark-streaming_2.10:1.6.1 -> 2.0.1 >>> [warn] * org.apache.spark:spark-core_2.10:1.6.1 -> 2.0.1 >>> [warn] Run 'evicted' to see detailed eviction warnings >>> [info] Compiling 18 Scala sources to C:\Users\marco\SparkExamples\t >>> arget\scala-2.10\classes... >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:2: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.linalg.{ Vector, Vectors } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:3: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.regression.LabeledPoint >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:4: >>> object classification is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.classification._ >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:5: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.tree.DecisionTree >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:6: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.tree.model.DecisionTreeModel >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:7: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.util.MLUtils >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:9: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.stat.{ >>> MultivariateStatisticalSummary, Statistics } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\AnotherDecisionTreeExample.scala:10: >>> object mllib is not a member of package org.apache.spark >>> [error] import org.apache.spark.mllib.linalg.Vector >>> ........ >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:4: >>> object classification is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.classification.{ >>> RandomForestClassifier, RandomForestClassificationModel } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:6: >>> object feature is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.feature.{ StringIndexer, >>> IndexToString, VectorIndexer, VectorAssembler } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:7: >>> object evaluation is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.evaluation.{ RegressionEvaluator, >>> MulticlassClassificationEvaluator } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:8: >>> object classification is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.classification._ >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:9: >>> object tuning is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.tuning.{ CrossValidator, >>> ParamGridBuilder } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:10: >>> object tuning is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.tuning.{ ParamGridBuilder, >>> TrainValidationSplit } >>> [error] ^ >>> [error] >>> C:\Users\marco\SparkExamples\src\main\scala\DecisionTreeExampleML.scala:16: >>> object Pipeline is not a member of package org.apache.spark.ml >>> [error] import org.apache.spark.ml.{ Pipeline, PipelineModel } >>> [error] >>> .... >>> and many more >>> >>> Could anyone assist or pass me the right dependencies? Or am i using >>> wrong resolvers? >>> >>> kr >>> marco >>> >> >> >> >> -- >> Donald Drake >> Drake Consulting >> http://www.drakeconsulting.com/ >> https://twitter.com/dondrake <http://www.MailLaunder.com/> >> 800-733-2143 >> > > -- Donald Drake Drake Consulting http://www.drakeconsulting.com/ https://twitter.com/dondrake <http://www.MailLaunder.com/> 800-733-2143