I am using Spark 1.3.0 and Scala 2.10. Thanks
On Tue, May 5, 2015 at 6:48 PM, Ted Yu <yuzhih...@gmail.com> wrote: > Can you give us a bit more information ? > Such as release of Spark you're using, version of Scala, etc. > > Thanks > > On Tue, May 5, 2015 at 6:37 PM, xweb <ashish8...@gmail.com> wrote: > >> I am getting on following code >> Error:(164, 25) *overloaded method constructor Strategy with >> alternatives:* >> (algo: org.apache.spark.mllib.tree.configuration.Algo.Algo,impurity: >> org.apache.spark.mllib.tree.impurity.Impurity,maxDepth: Int,numClasses: >> Int,maxBins: Int,categoricalFeaturesInfo: >> >> java.util.Map[Integer,Integer])org.apache.spark.mllib.tree.configuration.Strategy >> <and> >> (algo: org.apache.spark.mllib.tree.configuration.Algo.Algo,impurity: >> org.apache.spark.mllib.tree.impurity.Impurity,maxDepth: Int,numClasses: >> Int,maxBins: Int,quantileCalculationStrategy: >> >> org.apache.spark.mllib.tree.configuration.QuantileStrategy.QuantileStrategy,categoricalFeaturesInfo: >> scala.collection.immutable.Map[Int,Int],minInstancesPerNode: >> Int,minInfoGain: Double,maxMemoryInMB: Int,subsamplingRate: >> Double,useNodeIdCache: Boolean,checkpointInterval: >> Int)org.apache.spark.mllib.tree.configuration.Strategy >> cannot be applied to >> (org.apache.spark.mllib.tree.configuration.Algo.Value, >> org.apache.spark.mllib.tree.impurity.Gini.type, Int, Int, Int, >> scala.collection.immutable.Map[Int,Int]) >> val dTreeStrategy = new Strategy(algo, impurity, maxDepth, numClasses, >> maxBins, categoricalFeaturesInfo) >> ^ >> <code> >> val categoricalFeaturesInfo = Map[Int, Int]() >> val impurity = Gini >> val maxDepth = 4 >> val maxBins = 32 >> val algo = Algo.Classification >> >> val numClasses = 7 >> >> val dTreeStrategy = new Strategy(algo, impurity, maxDepth, numClasses, >> maxBins, categoricalFeaturesInfo) >> </code> >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/overloaded-method-constructor-Strategy-with-alternatives-tp22777.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >