Looks like there is an issue with Maven (likely just the test itself
though). We should look into it.


On Wed, Nov 2, 2016 at 11:32 AM, Dongjoon Hyun <dongj...@apache.org> wrote:

> Hi, Sean.
>
> The same failure blocks me, too.
>
> - SPARK-18189: Fix serialization issue in KeyValueGroupedDataset ***
> FAILED ***
>
> I used `-Pyarn -Phadoop-2.7 -Pkinesis-asl -Phive -Phive-thriftserver
> -Dsparkr` on CentOS 7 / OpenJDK1.8.0_111.
>
> Dongjoon.
>
> On 2016-11-02 10:44 (-0700), Sean Owen <so...@cloudera.com> wrote:
> > Sigs, license, etc are OK. There are no Blockers for 2.0.2, though here
> are
> > the 4 issues still open:
> >
> > SPARK-14387 Enable Hive-1.x ORC compatibility with
> > spark.sql.hive.convertMetastoreOrc
> > SPARK-17957 Calling outer join and na.fill(0) and then inner join will
> miss
> > rows
> > SPARK-17981 Incorrectly Set Nullability to False in FilterExec
> > SPARK-18160 spark.files & spark.jars should not be passed to driver in
> yarn
> > mode
> >
> > Running with Java 8, -Pyarn -Phive -Phive-thriftserver -Phadoop-2.7 on
> > Ubuntu 16, I am seeing consistent failures in this test below. I think we
> > very recently changed this so it could be legitimate. But does anyone
> else
> > see something like this? I have seen other failures in this test due to
> OOM
> > but my MAVEN_OPTS allows 6g of heap, which ought to be plenty.
> >
> >
> > - SPARK-18189: Fix serialization issue in KeyValueGroupedDataset ***
> FAILED
> > ***
> >   isContain was true Interpreter output contained 'Exception':
> >   Welcome to
> >         ____              __
> >        / __/__  ___ _____/ /__
> >       _\ \/ _ \/ _ `/ __/  '_/
> >      /___/ .__/\_,_/_/ /_/\_\   version 2.0.2
> >         /_/
> >
> >   Using Scala version 2.11.8 (OpenJDK 64-Bit Server VM, Java 1.8.0_102)
> >   Type in expressions to have them evaluated.
> >   Type :help for more information.
> >
> >   scala>
> >   scala> keyValueGrouped:
> > org.apache.spark.sql.KeyValueGroupedDataset[Int,(Int, Int)] =
> > org.apache.spark.sql.KeyValueGroupedDataset@70c30f72
> >
> >   scala> mapGroups: org.apache.spark.sql.Dataset[(Int, Int)] = [_1: int,
> > _2: int]
> >
> >   scala> broadcasted: org.apache.spark.broadcast.Broadcast[Int] =
> > Broadcast(0)
> >
> >   scala>
> >   scala>
> >   scala> dataset: org.apache.spark.sql.Dataset[Int] = [value: int]
> >
> >   scala> org.apache.spark.SparkException: Job aborted due to stage
> failure:
> > Task 0 in stage 0.0 failed 1 times, most recent failure: Lost task 0.0 in
> > stage 0.0 (TID 0, localhost):
> > com.google.common.util.concurrent.ExecutionError:
> > java.lang.ClassCircularityError:
> > io/netty/util/internal/__matchers__/org/apache/spark/network/protocol/
> MessageMatcher
> >   at com.google.common.cache.LocalCache$Segment.get(
> LocalCache.java:2261)
> >   at com.google.common.cache.LocalCache.get(LocalCache.java:4000)
> >   at com.google.common.cache.LocalCache.getOrLoad(LocalCache.java:4004)
> >   at
> > com.google.common.cache.LocalCache$LocalLoadingCache.
> get(LocalCache.java:4874)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> CodeGenerator$.compile(CodeGenerator.scala:841)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> GenerateSafeProjection$.create(GenerateSafeProjection.scala:188)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> GenerateSafeProjection$.create(GenerateSafeProjection.scala:36)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> CodeGenerator.generate(CodeGenerator.scala:825)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> CodeGenerator.generate(CodeGenerator.scala:822)
> >   at
> > org.apache.spark.sql.execution.ObjectOperator$.deserializeRowToObject(
> objects.scala:137)
> >   at
> > org.apache.spark.sql.execution.AppendColumnsExec$$
> anonfun$9.apply(objects.scala:251)
> >   at
> > org.apache.spark.sql.execution.AppendColumnsExec$$
> anonfun$9.apply(objects.scala:250)
> >   at
> > org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$
> 1$$anonfun$apply$24.apply(RDD.scala:803)
> >   at
> > org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$
> 1$$anonfun$apply$24.apply(RDD.scala:803)
> >   at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
> >   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
> >   at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
> >   at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
> >   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:319)
> >   at org.apache.spark.rdd.RDD.iterator(RDD.scala:283)
> >   at
> > org.apache.spark.scheduler.ShuffleMapTask.runTask(
> ShuffleMapTask.scala:79)
> >   at
> > org.apache.spark.scheduler.ShuffleMapTask.runTask(
> ShuffleMapTask.scala:47)
> >   at org.apache.spark.scheduler.Task.run(Task.scala:86)
> >   at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:274)
> >   at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1142)
> >   at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:617)
> >   at java.lang.Thread.run(Thread.java:745)
> >   Caused by: java.lang.ClassCircularityError:
> > io/netty/util/internal/__matchers__/org/apache/spark/network/protocol/
> MessageMatcher
> >   at java.lang.Class.forName0(Native Method)
> >   at java.lang.Class.forName(Class.java:348)
> >   at
> > io.netty.util.internal.JavassistTypeParameterMatcherGenerator.generate(
> JavassistTypeParameterMatcherGenerator.java:62)
> >   at
> > io.netty.util.internal.JavassistTypeParameterMatcherGenerator.generate(
> JavassistTypeParameterMatcherGenerator.java:54)
> >   at
> > io.netty.util.internal.TypeParameterMatcher.get(
> TypeParameterMatcher.java:42)
> >   at
> > io.netty.util.internal.TypeParameterMatcher.find(
> TypeParameterMatcher.java:78)
> >   at
> > io.netty.handler.codec.MessageToMessageEncoder.<init>
> (MessageToMessageEncoder.java:60)
> >   at
> > org.apache.spark.network.protocol.MessageEncoder.<init>
> (MessageEncoder.java:34)
> >   at
> > org.apache.spark.network.TransportContext.<init>(
> TransportContext.java:78)
> >   at
> > org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient(
> NettyRpcEnv.scala:354)
> >   at
> > org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(
> NettyRpcEnv.scala:324)
> >   at org.apache.spark.repl.ExecutorClassLoader.org
> > $apache$spark$repl$ExecutorClassLoader$$getClassFileInputStreamFromSpa
> rkRPC(ExecutorClassLoader.scala:90)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader$$anonfun$
> 1.apply(ExecutorClassLoader.scala:57)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader$$anonfun$
> 1.apply(ExecutorClassLoader.scala:57)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader.findClassLocally(
> ExecutorClassLoader.scala:161)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader.findClass(
> ExecutorClassLoader.scala:80)
> >   at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
> >   at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
> >   at java.lang.Class.forName0(Native Method)
> >   at java.lang.Class.forName(Class.java:348)
> >   at
> > io.netty.util.internal.JavassistTypeParameterMatcherGenerator.generate(
> JavassistTypeParameterMatcherGenerator.java:62)
> >   at
> > io.netty.util.internal.JavassistTypeParameterMatcherGenerator.generate(
> JavassistTypeParameterMatcherGenerator.java:54)
> >   at
> > io.netty.util.internal.TypeParameterMatcher.get(
> TypeParameterMatcher.java:42)
> >   at
> > io.netty.util.internal.TypeParameterMatcher.find(
> TypeParameterMatcher.java:78)
> >   at
> > io.netty.handler.codec.MessageToMessageEncoder.<init>
> (MessageToMessageEncoder.java:60)
> >   at
> > org.apache.spark.network.protocol.MessageEncoder.<init>
> (MessageEncoder.java:34)
> >   at
> > org.apache.spark.network.TransportContext.<init>(
> TransportContext.java:78)
> >   at
> > org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient(
> NettyRpcEnv.scala:354)
> >   at
> > org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(
> NettyRpcEnv.scala:324)
> >   at org.apache.spark.repl.ExecutorClassLoader.org
> > $apache$spark$repl$ExecutorClassLoader$$getClassFileInputStreamFromSpa
> rkRPC(ExecutorClassLoader.scala:90)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader$$anonfun$
> 1.apply(ExecutorClassLoader.scala:57)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader$$anonfun$
> 1.apply(ExecutorClassLoader.scala:57)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader.findClassLocally(
> ExecutorClassLoader.scala:161)
> >   at
> > org.apache.spark.repl.ExecutorClassLoader.findClass(
> ExecutorClassLoader.scala:80)
> >   at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
> >   at java.lang.ClassLoader.loadClass(ClassLoader.java:411)
> >   at
> > org.apache.spark.util.ParentClassLoader.loadClass(
> ParentClassLoader.scala:34)
> >   at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
> >   at
> > org.apache.spark.util.ParentClassLoader.loadClass(
> ParentClassLoader.scala:30)
> >   at java.lang.Class.forName0(Native Method)
> >   at java.lang.Class.forName(Class.java:348)
> >   at
> > org.codehaus.janino.ClassLoaderIClassLoader.findIClass(
> ClassLoaderIClassLoader.java:78)
> >   at org.codehaus.janino.IClassLoader.loadIClass(IClassLoader.java:254)
> >   at org.codehaus.janino.UnitCompiler.findTypeByName(
> UnitCompiler.java:6893)
> >   at
> > org.codehaus.janino.UnitCompiler.getReferenceType(
> UnitCompiler.java:5331)
> >   at
> > org.codehaus.janino.UnitCompiler.getReferenceType(
> UnitCompiler.java:5207)
> >   at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:5188)
> >   at org.codehaus.janino.UnitCompiler.access$12600(
> UnitCompiler.java:185)
> >   at
> > org.codehaus.janino.UnitCompiler$16.visitReferenceType(
> UnitCompiler.java:5119)
> >   at org.codehaus.janino.Java$ReferenceType.accept(Java.java:2880)
> >   at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:5159)
> >   at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:5414)
> >   at org.codehaus.janino.UnitCompiler.access$12400(
> UnitCompiler.java:185)
> >   at
> > org.codehaus.janino.UnitCompiler$16.visitArrayType(UnitCompiler.
> java:5117)
> >   at org.codehaus.janino.Java$ArrayType.accept(Java.java:2954)
> >   at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:5159)
> >   at org.codehaus.janino.UnitCompiler.access$16700(
> UnitCompiler.java:185)
> >   at
> > org.codehaus.janino.UnitCompiler$31.getParameterTypes2(
> UnitCompiler.java:8533)
> >   at
> > org.codehaus.janino.IClass$IInvocable.getParameterTypes(IClass.java:835)
> >   at org.codehaus.janino.IClass$IMethod.getDescriptor2(IClass.java:1063)
> >   at org.codehaus.janino.IClass$IInvocable.getDescriptor(
> IClass.java:849)
> >   at org.codehaus.janino.IClass.getIMethods(IClass.java:211)
> >   at org.codehaus.janino.IClass.getIMethods(IClass.java:199)
> >   at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:409)
> >   at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:393)
> >   at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:185)
> >   at
> > org.codehaus.janino.UnitCompiler$2.visitPackageMemberClassDeclara
> tion(UnitCompiler.java:347)
> >   at
> > org.codehaus.janino.Java$PackageMemberClassDeclaration.
> accept(Java.java:1139)
> >   at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:354)
> >   at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:322)
> >   at
> > org.codehaus.janino.SimpleCompiler.compileToClassLoader(
> SimpleCompiler.java:383)
> >   at
> > org.codehaus.janino.ClassBodyEvaluator.compileToClass(
> ClassBodyEvaluator.java:315)
> >   at
> > org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:233)
> >   at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:192)
> >   at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:84)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> CodeGenerator$.org$apache$spark$sql$catalyst$expressions$codegen$
> CodeGenerator$$doCompile(CodeGenerator.scala:887)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> CodeGenerator$$anon$1.load(CodeGenerator.scala:950)
> >   at
> > org.apache.spark.sql.catalyst.expressions.codegen.
> CodeGenerator$$anon$1.load(CodeGenerator.scala:947)
> >   at
> > com.google.common.cache.LocalCache$LoadingValueReference.
> loadFuture(LocalCache.java:3599)
> >   at
> > com.google.common.cache.LocalCache$Segment.loadSync(
> LocalCache.java:2379)
> >   at
> > com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.
> java:2342)
> >   at com.google.common.cache.LocalCache$Segment.get(
> LocalCache.java:2257)
> >   ... 26 more
> >
> >
> >
> > On Wed, Nov 2, 2016 at 4:52 AM Reynold Xin <r...@databricks.com> wrote:
> >
> > > Please vote on releasing the following candidate as Apache Spark
> version
> > > 2.0.2. The vote is open until Fri, Nov 4, 2016 at 22:00 PDT and passes
> if a
> > > majority of at least 3+1 PMC votes are cast.
> > >
> > > [ ] +1 Release this package as Apache Spark 2.0.2
> > > [ ] -1 Do not release this package because ...
> > >
> > >
> > > The tag to be voted on is v2.0.2-rc2
> > > (a6abe1ee22141931614bf27a4f371c46d8379e33)
> > >
> > > This release candidate resolves 84 issues:
> > > https://s.apache.org/spark-2.0.2-jira
> > >
> > > The release files, including signatures, digests, etc. can be found at:
> > > http://people.apache.org/~pwendell/spark-releases/spark-2.0.2-rc2-bin/
> > >
> > > Release artifacts are signed with the following key:
> > > https://people.apache.org/keys/committer/pwendell.asc
> > >
> > > The staging repository for this release can be found at:
> > > https://repository.apache.org/content/repositories/
> orgapachespark-1210/
> > >
> > > The documentation corresponding to this release can be found at:
> > > http://people.apache.org/~pwendell/spark-releases/spark-
> 2.0.2-rc2-docs/
> > >
> > >
> > > Q: How can I help test this release?
> > > A: If you are a Spark user, you can help us test this release by
> taking an
> > > existing Spark workload and running on this release candidate, then
> > > reporting any regressions from 2.0.1.
> > >
> > > Q: What justifies a -1 vote for this release?
> > > A: This is a maintenance release in the 2.0.x series. Bugs already
> present
> > > in 2.0.1, missing features, or bugs related to new features will not
> > > necessarily block this release.
> > >
> > > Q: What fix version should I use for patches merging into branch-2.0
> from
> > > now on?
> > > A: Please mark the fix version as 2.0.3, rather than 2.0.2. If a new RC
> > > (i.e. RC3) is cut, I will change the fix version of those patches to
> 2.0.2.
> > >
> >
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to