As a remedy, i'd suggest to branch out spark 1.2 work and rollback 1.2.1
commit on master until 1.2 branch is fixed.

On Tue, Feb 24, 2015 at 10:19 AM, Dmitriy Lyubimov <dlie...@gmail.com>
wrote:

> oops.
>
> tests dont test shell startup.
>
> apparently stuff got out of sync with 1.2
>
> On Tue, Feb 24, 2015 at 10:02 AM, Pat Ferrel <p...@occamsmachete.com>
> wrote:
>
>> Me too and I built with 1.2.1
>>
>> On Feb 24, 2015, at 9:50 AM, Andrew Musselman <andrew.mussel...@gmail.com>
>> wrote:
>>
>> I've just rebuild mahout master and spark v1.2.1-rc2 and am getting this
>> error when I try out the spark-shell; am I missing something?
>>
>> $ bin/mahout spark-shell
>> SLF4J: Class path contains multiple SLF4J bindings.
>> SLF4J: Found binding in
>>
>> [jar:file:/home/akm/mahout/mrlegacy/target/mahout-mrlegacy-1.0-SNAPSHOT-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: Found binding in
>>
>> [jar:file:/home/akm/mahout/spark/target/mahout-spark_2.10-1.0-SNAPSHOT-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: Found binding in
>>
>> [jar:file:/home/akm/spark/assembly/target/scala-2.10/spark-assembly-1.2.1-hadoop1.1.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>> explanation.
>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>> error:
>>     while compiling: <init>
>>        during phase: typer
>>     library version: version 2.10.4
>>    compiler version: version 2.10.0
>>  reconstructed args:
>>
>>  last tree to typer: Literal(Constant(()))
>>              symbol: null
>>   symbol definition: null
>>                 tpe: Unit
>>       symbol owners:
>>      context owners: package <empty>
>>
>> == Enclosing template or block ==
>>
>> Block( // tree.tpe=Unit
>>  {}
>>  ()
>> )
>>
>> == Expanded type of tree ==
>>
>> TypeRef(TypeSymbol(final abstract class Unit extends AnyVal))
>>
>> uncaught exception during compilation: java.lang.NoSuchMethodError
>>
>> Failed to initialize compiler: NoSuchMethodError.
>> This is most often remedied by a full clean and recompile.
>> Otherwise, your classpath may continue bytecode compiled by
>> different and incompatible versions of scala.
>>
>> java.lang.NoSuchMethodError:
>>
>> scala.reflect.internal.TreeInfo.firstArgument(Lscala/reflect/internal/Trees$Tree;)Lscala/reflect/internal/Trees$Tree;
>>    at
>> scala.tools.nsc.typechecker.Typers$Typer.parentTypes(Typers.scala:1550)
>>    at
>> scala.tools.nsc.typechecker.Namers$Namer.templateSig(Namers.scala:861)
>>    at scala.tools.nsc.typechecker.Namers$Namer.classSig(Namers.scala:907)
>>    at scala.tools.nsc.typechecker.Namers$Namer.getSig$1(Namers.scala:1289)
>>    at scala.tools.nsc.typechecker.Namers$Namer.typeSig(Namers.scala:1347)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1$$anonfun$apply$1.apply$mcV$sp(Namers.scala:709)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1$$anonfun$apply$1.apply(Namers.scala:708)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1$$anonfun$apply$1.apply(Namers.scala:708)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer.scala$tools$nsc$typechecker$Namers$Namer$$logAndValidate(Namers.scala:1385)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1.apply(Namers.scala:708)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1.apply(Namers.scala:707)
>>    at
>> scala.tools.nsc.typechecker.Namers$$anon$1.completeImpl(Namers.scala:1496)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$LockingTypeCompleter$class.complete(Namers.scala:1504)
>>    at
>> scala.tools.nsc.typechecker.Namers$$anon$1.complete(Namers.scala:1494)
>>    at scala.reflect.internal.Symbols$Symbol.info(Symbols.scala:1231)
>>    at scala.reflect.internal.Symbols$Symbol.initialize(Symbols.scala:1374)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed1(Typers.scala:5119)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed(Typers.scala:5458)
>>    at
>>
>> scala.tools.nsc.interpreter.ReplGlobal$$anon$1$$anon$2.typed(ReplGlobal.scala:29)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer.scala$tools$nsc$typechecker$Typers$Typer$$typedStat$1(Typers.scala:2770)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer$$anonfun$55.apply(Typers.scala:2870)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer$$anonfun$55.apply(Typers.scala:2870)
>>    at scala.collection.immutable.List.loop$1(List.scala:170)
>>    at scala.collection.immutable.List.mapConserve(List.scala:186)
>>    at
>> scala.tools.nsc.typechecker.Typers$Typer.typedStats(Typers.scala:2870)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer.typedPackageDef$1(Typers.scala:5127)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed1(Typers.scala:5404)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed(Typers.scala:5458)
>>    at
>>
>> scala.tools.nsc.interpreter.ReplGlobal$$anon$1$$anon$2.typed(ReplGlobal.scala:29)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed(Typers.scala:5509)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3.apply(Analyzer.scala:98)
>>    at scala.tools.nsc.Global$GlobalPhase.applyPhase(Global.scala:461)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3$$anonfun$run$1.apply(Analyzer.scala:90)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3$$anonfun$run$1.apply(Analyzer.scala:90)
>>    at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>>    at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3.run(Analyzer.scala:90)
>>    at scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1574)
>>    at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1548)
>>    at scala.tools.nsc.Global$Run.compileSources(Global.scala:1544)
>>    at org.apache.spark.repl.SparkIMain.org
>> $apache$spark$repl$SparkIMain$$_initialize(SparkIMain.scala:187)
>>    at
>>
>> org.apache.spark.repl.SparkIMain.initializeSynchronous(SparkIMain.scala:208)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:961)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:916)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:916)
>>    at
>>
>> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
>>    at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:916)
>>    at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1011)
>>    at org.apache.mahout.sparkbindings.shell.Main$.main(Main.scala:39)
>>    at org.apache.mahout.sparkbindings.shell.Main.main(Main.scala)
>> 15/02/24 09:42:52 WARN SparkILoop$SparkILoopInterpreter: Warning: compiler
>> accessed before init set up.  Assuming no postInit code.
>> error:
>>     while compiling: <console>
>>        during phase: typer
>>     library version: version 2.10.4
>>    compiler version: version 2.10.0
>>  reconstructed args:
>>
>>  last tree to typer: Literal(Constant(()))
>>              symbol: null
>>   symbol definition: null
>>                 tpe: Unit
>>       symbol owners:
>>      context owners: package $line1
>>
>> == Enclosing template or block ==
>>
>> Block( // tree.tpe=Unit
>>  {}
>>  ()
>> )
>>
>> == Expanded type of tree ==
>>
>> TypeRef(TypeSymbol(final abstract class Unit extends AnyVal))
>>
>> uncaught exception during compilation: java.lang.NoSuchMethodError
>> Exception in thread "main" java.lang.NoSuchMethodError:
>>
>> scala.reflect.internal.TreeInfo.firstArgument(Lscala/reflect/internal/Trees$Tree;)Lscala/reflect/internal/Trees$Tree;
>>    at
>> scala.tools.nsc.typechecker.Typers$Typer.parentTypes(Typers.scala:1550)
>>    at
>> scala.tools.nsc.typechecker.Namers$Namer.templateSig(Namers.scala:861)
>>    at scala.tools.nsc.typechecker.Namers$Namer.getSig$1(Namers.scala:1300)
>>    at scala.tools.nsc.typechecker.Namers$Namer.typeSig(Namers.scala:1347)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1$$anonfun$apply$1.apply$mcV$sp(Namers.scala:709)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1$$anonfun$apply$1.apply(Namers.scala:708)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1$$anonfun$apply$1.apply(Namers.scala:708)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer.scala$tools$nsc$typechecker$Namers$Namer$$logAndValidate(Namers.scala:1385)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1.apply(Namers.scala:708)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$Namer$$anonfun$monoTypeCompleter$1.apply(Namers.scala:707)
>>    at
>> scala.tools.nsc.typechecker.Namers$$anon$1.completeImpl(Namers.scala:1496)
>>    at
>>
>> scala.tools.nsc.typechecker.Namers$LockingTypeCompleter$class.complete(Namers.scala:1504)
>>    at
>> scala.tools.nsc.typechecker.Namers$$anon$1.complete(Namers.scala:1494)
>>    at scala.reflect.internal.Symbols$Symbol.info(Symbols.scala:1231)
>>    at scala.reflect.internal.Symbols$Symbol.initialize(Symbols.scala:1374)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed1(Typers.scala:5119)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed(Typers.scala:5458)
>>    at
>>
>> scala.tools.nsc.interpreter.ReplGlobal$$anon$1$$anon$2.typed(ReplGlobal.scala:29)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer.scala$tools$nsc$typechecker$Typers$Typer$$typedStat$1(Typers.scala:2770)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer$$anonfun$55.apply(Typers.scala:2870)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer$$anonfun$55.apply(Typers.scala:2870)
>>    at scala.collection.immutable.List.loop$1(List.scala:170)
>>    at scala.collection.immutable.List.mapConserve(List.scala:186)
>>    at
>> scala.tools.nsc.typechecker.Typers$Typer.typedStats(Typers.scala:2870)
>>    at
>>
>> scala.tools.nsc.typechecker.Typers$Typer.typedPackageDef$1(Typers.scala:5127)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed1(Typers.scala:5404)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed(Typers.scala:5458)
>>    at
>>
>> scala.tools.nsc.interpreter.ReplGlobal$$anon$1$$anon$2.typed(ReplGlobal.scala:29)
>>    at scala.tools.nsc.typechecker.Typers$Typer.typed(Typers.scala:5509)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3.apply(Analyzer.scala:98)
>>    at scala.tools.nsc.Global$GlobalPhase.applyPhase(Global.scala:461)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3$$anonfun$run$1.apply(Analyzer.scala:90)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3$$anonfun$run$1.apply(Analyzer.scala:90)
>>    at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>>    at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>>    at
>>
>> scala.tools.nsc.typechecker.Analyzer$typerFactory$$anon$3.run(Analyzer.scala:90)
>>    at scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1574)
>>    at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1548)
>>    at scala.tools.nsc.Global$Run.compileSources(Global.scala:1544)
>>    at
>>
>> org.apache.spark.repl.SparkIMain.compileSourcesKeepingRun(SparkIMain.scala:528)
>>    at
>>
>> org.apache.spark.repl.SparkIMain$ReadEvalPrint.compileAndSaveRun(SparkIMain.scala:923)
>>    at
>>
>> org.apache.spark.repl.SparkIMain$ReadEvalPrint.compile(SparkIMain.scala:879)
>>    at org.apache.spark.repl.SparkIMain.bind(SparkIMain.scala:719)
>>    at org.apache.spark.repl.SparkIMain.bind(SparkIMain.scala:762)
>>    at
>>
>> org.apache.spark.repl.SparkIMain$$anonfun$quietBind$1.apply(SparkIMain.scala:761)
>>    at
>>
>> org.apache.spark.repl.SparkIMain$$anonfun$quietBind$1.apply(SparkIMain.scala:761)
>>    at org.apache.spark.repl.SparkIMain.beQuietDuring(SparkIMain.scala:270)
>>    at org.apache.spark.repl.SparkIMain.quietBind(SparkIMain.scala:761)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1$$anonfun$apply$mcZ$sp$2.apply$mcV$sp(SparkILoop.scala:935)
>>    at
>>
>> org.apache.spark.repl.SparkILoopInit$class.runThunks(SparkILoopInit.scala:147)
>>    at org.apache.spark.repl.SparkILoop.runThunks(SparkILoop.scala:60)
>>    at
>>
>> org.apache.spark.repl.SparkILoopInit$class.postInitialization(SparkILoopInit.scala:106)
>>    at
>>
>> org.apache.mahout.sparkbindings.shell.MahoutSparkILoop.postInitialization(MahoutSparkILoop.scala:24)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:962)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:916)
>>    at
>>
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:916)
>>    at
>>
>> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
>>    at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:916)
>>    at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1011)
>>    at org.apache.mahout.sparkbindings.shell.Main$.main(Main.scala:39)
>>    at org.apache.mahout.sparkbindings.shell.Main.main(Main.scala)
>>
>>
>

Reply via email to