Thanks! On Thu, Dec 10, 2015 at 12:32 PM, Maximilian Michels <m...@apache.org> wrote:
> Hi Cory, > > The issue has been fixed in the master and the latest Maven snapshot. > https://issues.apache.org/jira/browse/FLINK-3143 > > Cheers, > Max > > On Tue, Dec 8, 2015 at 12:35 PM, Maximilian Michels <m...@apache.org> > wrote: > > Thanks for the stack trace, Cory. Looks like you were on the right > > path with the Spark issue. We will file an issue and correct it soon. > > > > Thanks, > > Max > > > > On Mon, Dec 7, 2015 at 8:20 PM, Stephan Ewen <se...@apache.org> wrote: > >> Sorry, correcting myself: > >> > >> The ClosureCleaner uses Kryo's bundled ASM 4 without any reason - simply > >> adjusting the imports to use the common ASM (which is 5.0) should do it > ;-) > >> > >> On Mon, Dec 7, 2015 at 8:18 PM, Stephan Ewen <se...@apache.org> wrote: > >>> > >>> Flink's own asm is 5.0, but the Kryo version used in Flink bundles > >>> reflectasm with a dedicated asm version 4 (no lambdas supported). > >>> > >>> Might be as simple as bumping the kryo version... > >>> > >>> > >>> > >>> On Mon, Dec 7, 2015 at 7:59 PM, Cory Monty < > cory.mo...@getbraintree.com> > >>> wrote: > >>>> > >>>> Thanks, Max. > >>>> > >>>> Here is the stack trace I receive: > >>>> > >>>> java.lang.IllegalArgumentException: > >>>> at > >>>> > com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.<init>(Unknown > >>>> Source) > >>>> at > >>>> > com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.<init>(Unknown > >>>> Source) > >>>> at > >>>> > com.esotericsoftware.reflectasm.shaded.org.objectweb.asm.ClassReader.<init>(Unknown > >>>> Source) > >>>> at > >>>> > org.apache.flink.api.scala.ClosureCleaner$.org$apache$flink$api$scala$ClosureCleaner$$getClassReader(ClosureCleaner.scala:47) > >>>> at > >>>> > org.apache.flink.api.scala.ClosureCleaner$.getInnerClasses(ClosureCleaner.scala:90) > >>>> at > >>>> > org.apache.flink.api.scala.ClosureCleaner$.clean(ClosureCleaner.scala:113) > >>>> at > >>>> > org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.scalaClean(StreamExecutionEnvironment.scala:555) > >>>> at > >>>> > org.apache.flink.streaming.api.scala.DataStream.clean(DataStream.scala:764) > >>>> at > >>>> > org.apache.flink.streaming.api.scala.DataStream.flatMap(DataStream.scala:473) > >>>> > >>>> On Mon, Dec 7, 2015 at 11:58 AM, Maximilian Michels <m...@apache.org> > >>>> wrote: > >>>>> > >>>>> For completeness, could you provide a stack trace of the error > message? > >>>>> > >>>>> On Mon, Dec 7, 2015 at 6:56 PM, Maximilian Michels <m...@apache.org> > >>>>> wrote: > >>>>> > Hi Cory, > >>>>> > > >>>>> > Thanks for reporting the issue. Scala should run independently of > the > >>>>> > Java version. We are already using ASM version 5.0.4. However, some > >>>>> > code uses the ASM4 op codes which don't seem to be work with Java > 8. > >>>>> > This needs to be fixed. I'm filing a JIRA. > >>>>> > > >>>>> > Cheers, > >>>>> > Max > >>>>> > > >>>>> > On Mon, Dec 7, 2015 at 4:15 PM, Cory Monty > >>>>> > <cory.mo...@getbraintree.com> wrote: > >>>>> >> Is it possible to use Scala 2.11 and Java 8? > >>>>> >> > >>>>> >> I'm able to get our project to compile correctly, however there > are > >>>>> >> runtime > >>>>> >> errors with the Reflectasm library (I'm guessing due to Kyro). I > >>>>> >> looked into > >>>>> >> the error and it seems Spark had the same issue > >>>>> >> (https://issues.apache.org/jira/browse/SPARK-6152, > >>>>> >> https://github.com/EsotericSoftware/reflectasm/issues/35) > because of > >>>>> >> an > >>>>> >> outdated version of Kyro. > >>>>> >> > >>>>> >> I'm also unsure if maybe we have to build Flink with Scala 2.11 > >>>>> >> > >>>>> >> ( > https://ci.apache.org/projects/flink/flink-docs-release-0.10/setup/building.html > ) > >>>>> >> in order to run Flink correctly with Java 8. > >>>>> >> > >>>>> >> Cheers, > >>>>> >> > >>>>> >> Cory > >>>> > >>>> > >>> > >> >