Thanks for the stack trace, Cory. Looks like you were on the right path with the Spark issue. We will file an issue and correct it soon.
Thanks, Max On Mon, Dec 7, 2015 at 8:20 PM, Stephan Ewen <se...@apache.org> wrote: > Sorry, correcting myself: > > The ClosureCleaner uses Kryo's bundled ASM 4 without any reason - simply > adjusting the imports to use the common ASM (which is 5.0) should do it ;-) > > On Mon, Dec 7, 2015 at 8:18 PM, Stephan Ewen <se...@apache.org> wrote: >> >> Flink's own asm is 5.0, but the Kryo version used in Flink bundles >> reflectasm with a dedicated asm version 4 (no lambdas supported). >> >> Might be as simple as bumping the kryo version... >> >> >> >> On Mon, Dec 7, 2015 at 7:59 PM, Cory Monty <cory.mo...@getbraintree.com> >> wrote: >>> >>> Thanks, Max. >>> >>> Here is the stack trace I receive: >>> >>> java.lang.IllegalArgumentException: >>> at >>> com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.<init>(Unknown >>> Source) >>> at >>> com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.<init>(Unknown >>> Source) >>> at >>> com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.<init>(Unknown >>> Source) >>> at >>> org.apache.flink.api.scala.ClosureCleaner$.org$apache$flink$api$scala$ClosureCleaner$$getClassReader(ClosureCleaner.scala:47) >>> at >>> org.apache.flink.api.scala.ClosureCleaner$.getInnerClasses(ClosureCleaner.scala:90) >>> at >>> org.apache.flink.api.scala.ClosureCleaner$.clean(ClosureCleaner.scala:113) >>> at >>> org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.scalaClean(StreamExecutionEnvironment.scala:555) >>> at >>> org.apache.flink.streaming.api.scala.DataStream.clean(DataStream.scala:764) >>> at >>> org.apache.flink.streaming.api.scala.DataStream.flatMap(DataStream.scala:473) >>> >>> On Mon, Dec 7, 2015 at 11:58 AM, Maximilian Michels <m...@apache.org> >>> wrote: >>>> >>>> For completeness, could you provide a stack trace of the error message? >>>> >>>> On Mon, Dec 7, 2015 at 6:56 PM, Maximilian Michels <m...@apache.org> >>>> wrote: >>>> > Hi Cory, >>>> > >>>> > Thanks for reporting the issue. Scala should run independently of the >>>> > Java version. We are already using ASM version 5.0.4. However, some >>>> > code uses the ASM4 op codes which don't seem to be work with Java 8. >>>> > This needs to be fixed. I'm filing a JIRA. >>>> > >>>> > Cheers, >>>> > Max >>>> > >>>> > On Mon, Dec 7, 2015 at 4:15 PM, Cory Monty >>>> > <cory.mo...@getbraintree.com> wrote: >>>> >> Is it possible to use Scala 2.11 and Java 8? >>>> >> >>>> >> I'm able to get our project to compile correctly, however there are >>>> >> runtime >>>> >> errors with the Reflectasm library (I'm guessing due to Kyro). I >>>> >> looked into >>>> >> the error and it seems Spark had the same issue >>>> >> (https://issues.apache.org/jira/browse/SPARK-6152, >>>> >> https://github.com/EsotericSoftware/reflectasm/issues/35) because of >>>> >> an >>>> >> outdated version of Kyro. >>>> >> >>>> >> I'm also unsure if maybe we have to build Flink with Scala 2.11 >>>> >> >>>> >> (https://ci.apache.org/projects/flink/flink-docs-release-0.10/setup/building.html) >>>> >> in order to run Flink correctly with Java 8. >>>> >> >>>> >> Cheers, >>>> >> >>>> >> Cory >>> >>> >> >