Yes. I think Felix checked in a fix to ignore tests run on java
versions that are not Java 8 (I think the fix was in
https://github.com/apache/spark/pull/21666 which is in 2.3.2)

Shivaram
On Mon, Jul 9, 2018 at 5:39 PM Sean Owen <sro...@gmail.com> wrote:
>
> Yes, this flavor of error should only come up in Java 9. Spark doesn't 
> support that. Is there any way to tell CRAN this should not be tested?
>
> On Mon, Jul 9, 2018, 4:17 PM Shivaram Venkataraman 
> <shiva...@eecs.berkeley.edu> wrote:
>>
>> The upcoming 2.2.2 release was submitted to CRAN. I think there are
>> some knows issues on Windows, but does anybody know what the following
>> error with Netty is ?
>>
>> >     WARNING: Illegal reflective access by 
>> > io.netty.util.internal.PlatformDependent0$1 
>> > (file:/home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/jars/netty-all-4.0.43.Final.jar)
>> >  to field java.nio.Buffer.address
>>
>> Thanks
>> Shivaram
>>
>>
>> ---------- Forwarded message ---------
>> From: <lig...@statistik.tu-dortmund.de>
>> Date: Mon, Jul 9, 2018 at 12:12 PM
>> Subject: [CRAN-pretest-archived] CRAN submission SparkR 2.2.2
>> To: <shiva...@cs.berkeley.edu>
>> Cc: <cran-submissi...@r-project.org>
>>
>>
>> Dear maintainer,
>>
>> package SparkR_2.2.2.tar.gz does not pass the incoming checks
>> automatically, please see the following pre-tests:
>> Windows: 
>> <https://win-builder.r-project.org/incoming_pretest/SparkR_2.2.2_20180709_175630/Windows/00check.log>
>> Status: 1 ERROR, 1 WARNING
>> Debian: 
>> <https://win-builder.r-project.org/incoming_pretest/SparkR_2.2.2_20180709_175630/Debian/00check.log>
>> Status: 1 ERROR, 2 WARNINGs
>>
>> Last released version's CRAN status: ERROR: 1, OK: 1
>> See: <https://CRAN.R-project.org/web/checks/check_results_SparkR.html>
>>
>> CRAN Web: <https://cran.r-project.org/package=SparkR>
>>
>> Please fix all problems and resubmit a fixed version via the webform.
>> If you are not sure how to fix the problems shown, please ask for help
>> on the R-package-devel mailing list:
>> <https://stat.ethz.ch/mailman/listinfo/r-package-devel>
>> If you are fairly certain the rejection is a false positive, please
>> reply-all to this message and explain.
>>
>> More details are given in the directory:
>> <https://win-builder.r-project.org/incoming_pretest/SparkR_2.2.2_20180709_175630/>
>> The files will be removed after roughly 7 days.
>>
>> No strong reverse dependencies to be checked.
>>
>> Best regards,
>> CRAN teams' auto-check service
>> Flavor: r-devel-linux-x86_64-debian-gcc, r-devel-windows-ix86+x86_64
>> Check: CRAN incoming feasibility, Result: WARNING
>>   Maintainer: 'Shivaram Venkataraman <shiva...@cs.berkeley.edu>'
>>
>>   New submission
>>
>>   Package was archived on CRAN
>>
>>   Insufficient package version (submitted: 2.2.2, existing: 2.3.0)
>>
>>   Possibly mis-spelled words in DESCRIPTION:
>>     Frontend (4:10, 5:28)
>>
>>   CRAN repository db overrides:
>>     X-CRAN-Comment: Archived on 2018-05-01 as check problems were not
>>       corrected despite reminders.
>>
>>   Found the following (possibly) invalid URLs:
>>     URL: http://spark.apache.org/docs/latest/api/R/mean.html
>>       From: inst/doc/sparkr-vignettes.html
>>       Status: 404
>>       Message: Not Found
>>
>> Flavor: r-devel-windows-ix86+x86_64
>> Check: running tests for arch 'x64', Result: ERROR
>>     Running 'run-all.R' [175s]
>>   Running the tests in 'tests/run-all.R' failed.
>>   Complete output:
>>     > #
>>     > # Licensed to the Apache Software Foundation (ASF) under one or more
>>     > # contributor license agreements.  See the NOTICE file distributed with
>>     > # this work for additional information regarding copyright ownership.
>>     > # The ASF licenses this file to You under the Apache License, Version 
>> 2.0
>>     > # (the "License"); you may not use this file except in compliance with
>>     > # the License.  You may obtain a copy of the License at
>>     > #
>>     > #    http://www.apache.org/licenses/LICENSE-2.0
>>     > #
>>     > # Unless required by applicable law or agreed to in writing, software
>>     > # distributed under the License is distributed on an "AS IS" BASIS,
>>     > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or 
>> implied.
>>     > # See the License for the specific language governing permissions and
>>     > # limitations under the License.
>>     > #
>>     >
>>     > library(testthat)
>>     > library(SparkR)
>>
>>     Attaching package: 'SparkR'
>>
>>     The following object is masked from 'package:testthat':
>>
>>         describe
>>
>>     The following objects are masked from 'package:stats':
>>
>>         cov, filter, lag, na.omit, predict, sd, var, window
>>
>>     The following objects are masked from 'package:base':
>>
>>         as.data.frame, colnames, colnames<-, drop, endsWith, intersect,
>>         rank, rbind, sample, startsWith, subset, summary, transform, union
>>
>>     >
>>     > # Turn all warnings into errors
>>     > options("warn" = 2)
>>     >
>>     > if (.Platform$OS.type == "windows") {
>>     +   Sys.setenv(TZ = "GMT")
>>     + }
>>     >
>>     > # Setup global test environment
>>     > # Install Spark first to set SPARK_HOME
>>     >
>>     > # NOTE(shivaram): We set overwrite to handle any old tar.gz
>> files or directories left behind on
>>     > # CRAN machines. For Jenkins we should already have SPARK_HOME set.
>>     > install.spark(overwrite = TRUE)
>>     Overwrite = TRUE: download and overwrite the tar fileand Spark
>> package directory if they exist.
>>     Spark not found in the cache directory. Installation will start.
>>     MirrorUrl not provided.
>>     Looking for preferred site from apache website...
>>     Preferred mirror site found: http://mirror.dkd.de/apache/spark
>>     Downloading spark-2.2.2 for Hadoop 2.7 from:
>>     - 
>> http://mirror.dkd.de/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz
>>     trying URL 
>> 'http://mirror.dkd.de/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz'
>>     Content type 'application/x-gzip' length 200743115 bytes (191.4 MB)
>>     ==================================================
>>     downloaded 191.4 MB
>>
>>     Installing to C:\Users\ligges\AppData\Local\Apache\Spark\Cache
>>     DONE.
>>     SPARK_HOME set to
>> C:\Users\ligges\AppData\Local\Apache\Spark\Cache/spark-2.2.2-bin-hadoop2.7
>>     >
>>     > sparkRDir <- file.path(Sys.getenv("SPARK_HOME"), "R")
>>     > sparkRWhitelistSQLDirs <- c("spark-warehouse", "metastore_db")
>>     > invisible(lapply(sparkRWhitelistSQLDirs,
>>     +                  function(x) { unlink(file.path(sparkRDir, x),
>> recursive = TRUE, force = TRUE)}))
>>     > sparkRFilesBefore <- list.files(path = sparkRDir, all.files = TRUE)
>>     >
>>     > sparkRTestMaster <- "local[1]"
>>     > sparkRTestConfig <- list()
>>     > if (identical(Sys.getenv("NOT_CRAN"), "true")) {
>>     +   sparkRTestMaster <- ""
>>     + } else {
>>     +   # Disable hsperfdata on CRAN
>>     +   old_java_opt <- Sys.getenv("_JAVA_OPTIONS")
>>     +   Sys.setenv("_JAVA_OPTIONS" = paste("-XX:-UsePerfData", old_java_opt))
>>     +   tmpDir <- tempdir()
>>     +   tmpArg <- paste0("-Djava.io.tmpdir=", tmpDir)
>>     +   sparkRTestConfig <- list(spark.driver.extraJavaOptions = tmpArg,
>>     +                            spark.executor.extraJavaOptions = tmpArg)
>>     + }
>>     >
>>     > test_package("SparkR")
>>     Launching java with spark-submit command
>> C:\Users\ligges\AppData\Local\Apache\Spark\Cache/spark-2.2.2-bin-hadoop2.7/bin/spark-submit2.cmd
>>   --driver-java-options "-Djava.io.tmpdir=D:\temp\RtmpABZLQj"
>> sparkr-shell D:\temp\RtmpABZLQj\backend_port16d0838283f7e
>>     Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>     Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>>     Setting default log level to "WARN".
>>     To adjust logging level use sc.setLogLevel(newLevel). For SparkR,
>> use setLogLevel(newLevel).
>>     -- 1. Error: create DataFrame from list or data.frame
>> (@test_basic.R#21)  ------
>>     cannot open the connection
>>     1: sparkR.session(master = sparkRTestMaster, enableHiveSupport =
>> FALSE, sparkConfig = sparkRTestConfig) at
>> D:/temp/Rtmp8IKu99/RLIBS_77d8215b7bce/SparkR/tests/testthat/test_basic.R:21
>>     2: sparkR.sparkContext(master, appName, sparkHome, sparkConfigMap,
>> sparkExecutorEnvMap,
>>            sparkJars, sparkPackages)
>>     3: file(path, open = "rb")
>>
>>     Launching java with spark-submit command
>> C:\Users\ligges\AppData\Local\Apache\Spark\Cache/spark-2.2.2-bin-hadoop2.7/bin/spark-submit2.cmd
>>   --driver-java-options "-Djava.io.tmpdir=D:\temp\RtmpABZLQj"
>> sparkr-shell D:\temp\RtmpABZLQj\backend_port16d085df97d88
>>     Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>     Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>>     Setting default log level to "WARN".
>>     To adjust logging level use sc.setLogLevel(newLevel). For SparkR,
>> use setLogLevel(newLevel).
>>     18/07/09 18:10:43 ERROR Shell: Failed to locate the winutils
>> binary in the hadoop binary path
>>     java.io.IOException: Could not locate executable
>> null\bin\winutils.exe in the Hadoop binaries.
>>         at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:379)
>>         at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:394)
>>         at org.apache.hadoop.util.Shell.<clinit>(Shell.java:387)
>>         at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:80)
>>         at 
>> org.apache.hadoop.security.SecurityUtil.getAuthenticationMethod(SecurityUtil.java:611)
>>         at 
>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:273)
>>         at 
>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:261)
>>         at 
>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:791)
>>         at 
>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:761)
>>         at 
>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:634)
>>         at 
>> org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2427)
>>         at 
>> org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2427)
>>         at scala.Option.getOrElse(Option.scala:121)
>>         at org.apache.spark.util.Utils$.getCurrentUserName(Utils.scala:2427)
>>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:295)
>>         at 
>> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2516)
>>         at org.apache.spark.api.r.RRDD$.createSparkContext(RRDD.scala:139)
>>         at org.apache.spark.api.r.RRDD.createSparkContext(RRDD.scala)
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>         at 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>         at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.lang.reflect.Method.invoke(Method.java:498)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
>>         at 
>> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
>>         at 
>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>>         at 
>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
>>         at 
>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>>         at java.lang.Thread.run(Thread.java:748)
>>     18/07/09 18:10:43 WARN NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes
>> where applicable
>>     Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>     18/07/09 18:10:54 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:10:55 WARN BLAS: Failed to load implementation from:
>> com.github.fommil.netlib.NativeSystemBLAS
>>     18/07/09 18:10:55 WARN BLAS: Failed to load implementation from:
>> com.github.fommil.netlib.NativeRefBLAS
>>     18/07/09 18:10:55 WARN LAPACK: Failed to load implementation from:
>> com.github.fommil.netlib.NativeSystemLAPACK
>>     18/07/09 18:10:55 WARN LAPACK: Failed to load implementation from:
>> com.github.fommil.netlib.NativeRefLAPACK
>>     18/07/09 18:11:12 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:14 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:15 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:17 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:18 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:19 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:21 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:22 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:23 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:25 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:26 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:28 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:29 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:46 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:47 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:49 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     18/07/09 18:11:50 WARN WeightedLeastSquares: regParam is zero,
>> which might cause numerical instability and overfitting.
>>     == testthat results
>> ===========================================================
>>     OK: 6 SKIPPED: 0 FAILED: 1
>>     1. Error: create DataFrame from list or data.frame (@test_basic.R#21)
>>
>>     Error: testthat unit tests failed
>>     Execution halted
>>     Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>
>> Flavor: r-devel-linux-x86_64-debian-gcc
>> Check: tests, Result: ERROR
>>     Running 'run-all.R' [6s/15s]
>>   Running the tests in 'tests/run-all.R' failed.
>>   Complete output:
>>     > #
>>     > # Licensed to the Apache Software Foundation (ASF) under one or more
>>     > # contributor license agreements.  See the NOTICE file distributed with
>>     > # this work for additional information regarding copyright ownership.
>>     > # The ASF licenses this file to You under the Apache License, Version 
>> 2.0
>>     > # (the "License"); you may not use this file except in compliance with
>>     > # the License.  You may obtain a copy of the License at
>>     > #
>>     > #    http://www.apache.org/licenses/LICENSE-2.0
>>     > #
>>     > # Unless required by applicable law or agreed to in writing, software
>>     > # distributed under the License is distributed on an "AS IS" BASIS,
>>     > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or 
>> implied.
>>     > # See the License for the specific language governing permissions and
>>     > # limitations under the License.
>>     > #
>>     >
>>     > library(testthat)
>>     > library(SparkR)
>>
>>     Attaching package: 'SparkR'
>>
>>     The following object is masked from 'package:testthat':
>>
>>         describe
>>
>>     The following objects are masked from 'package:stats':
>>
>>         cov, filter, lag, na.omit, predict, sd, var, window
>>
>>     The following objects are masked from 'package:base':
>>
>>         as.data.frame, colnames, colnames<-, drop, endsWith, intersect,
>>         rank, rbind, sample, startsWith, subset, summary, transform, union
>>
>>     >
>>     > # Turn all warnings into errors
>>     > options("warn" = 2)
>>     >
>>     > if (.Platform$OS.type == "windows") {
>>     +   Sys.setenv(TZ = "GMT")
>>     + }
>>     >
>>     > # Setup global test environment
>>     > # Install Spark first to set SPARK_HOME
>>     >
>>     > # NOTE(shivaram): We set overwrite to handle any old tar.gz
>> files or directories left behind on
>>     > # CRAN machines. For Jenkins we should already have SPARK_HOME set.
>>     > install.spark(overwrite = TRUE)
>>     Overwrite = TRUE: download and overwrite the tar fileand Spark
>> package directory if they exist.
>>     Spark not found in the cache directory. Installation will start.
>>     MirrorUrl not provided.
>>     Looking for preferred site from apache website...
>>     Preferred mirror site found: http://mirror.klaus-uwe.me/apache/spark
>>     Downloading spark-2.2.2 for Hadoop 2.7 from:
>>     - 
>> http://mirror.klaus-uwe.me/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz
>>     trying URL 
>> 'http://mirror.klaus-uwe.me/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz'
>>     Content type 'application/octet-stream' length 200743115 bytes (191.4 MB)
>>     ==================================================
>>     downloaded 191.4 MB
>>
>>     Installing to /home/hornik/.cache/spark
>>     DONE.
>>     SPARK_HOME set to /home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7
>>     >
>>     > sparkRDir <- file.path(Sys.getenv("SPARK_HOME"), "R")
>>     > sparkRWhitelistSQLDirs <- c("spark-warehouse", "metastore_db")
>>     > invisible(lapply(sparkRWhitelistSQLDirs,
>>     +                  function(x) { unlink(file.path(sparkRDir, x),
>> recursive = TRUE, force = TRUE)}))
>>     > sparkRFilesBefore <- list.files(path = sparkRDir, all.files = TRUE)
>>     >
>>     > sparkRTestMaster <- "local[1]"
>>     > sparkRTestConfig <- list()
>>     > if (identical(Sys.getenv("NOT_CRAN"), "true")) {
>>     +   sparkRTestMaster <- ""
>>     + } else {
>>     +   # Disable hsperfdata on CRAN
>>     +   old_java_opt <- Sys.getenv("_JAVA_OPTIONS")
>>     +   Sys.setenv("_JAVA_OPTIONS" = paste("-XX:-UsePerfData", old_java_opt))
>>     +   tmpDir <- tempdir()
>>     +   tmpArg <- paste0("-Djava.io.tmpdir=", tmpDir)
>>     +   sparkRTestConfig <- list(spark.driver.extraJavaOptions = tmpArg,
>>     +                            spark.executor.extraJavaOptions = tmpArg)
>>     + }
>>     >
>>     > test_package("SparkR")
>>     Launching java with spark-submit command
>> /home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/bin/spark-submit
>> --driver-java-options "-Djava.io.tmpdir=/tmp/Rtmpkd8Lf6" sparkr-shell
>> /tmp/Rtmpkd8Lf6/backend_port289f65a5f5e0
>>     Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>     Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>     Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>>     Setting default log level to "WARN".
>>     To adjust logging level use sc.setLogLevel(newLevel). For SparkR,
>> use setLogLevel(newLevel).
>>     WARNING: An illegal reflective access operation has occurred
>>     WARNING: Illegal reflective access by
>> io.netty.util.internal.PlatformDependent0$1
>> (file:/home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/jars/netty-all-4.0.43.Final.jar)
>> to field java.nio.Buffer.address
>>     WARNING: Please consider reporting this to the maintainers of
>> io.netty.util.internal.PlatformDependent0$1
>>     WARNING: Use --illegal-access=warn to enable warnings of further
>> illegal reflective access operations
>>     WARNING: All illegal access operations will be denied in a future release
>>     18/07/09 17:58:50 WARN NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes
>> where applicable
>>     18/07/09 17:58:54 ERROR RBackendHandler: count on 13 failed
>>     java.lang.reflect.InvocationTargetException
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>         at 
>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.base/java.lang.reflect.Method.invoke(Method.java:564)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
>>         at 
>> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
>>         at 
>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>>         at 
>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
>>         at 
>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>>         at java.base/java.lang.Thread.run(Thread.java:844)
>>     Caused by: java.lang.IllegalArgumentException
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426)
>>         at 
>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103)
>>         at 
>> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426)
>>         at org.apache.xbean.asm5.ClassReader.a(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.b(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256)
>>         at scala.collection.immutable.List.foreach(List.scala:381)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156)
>>         at org.apache.spark.SparkContext.clean(SparkContext.scala:2294)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094)
>>         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.collect(RDD.scala:935)
>>         at 
>> org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:278)
>>         at 
>> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2439)
>>         at 
>> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2438)
>>         at org.apache.spark.sql.Dataset$$anonfun$55.apply(Dataset.scala:2846)
>>         at 
>> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65)
>>         at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2845)
>>         at org.apache.spark.sql.Dataset.count(Dataset.scala:2438)
>>         ... 36 more
>>     ── 1. Error: create DataFrame from list or data.frame
>> (@test_basic.R#26)  ──────
>>     java.lang.IllegalArgumentException
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426)
>>         at 
>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103)
>>         at 
>> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426)
>>         at org.apache.xbean.asm5.ClassReader.a(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.b(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256)
>>         at scala.collection.immutable.List.foreach(List.scala:381)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156)
>>         at org.apache.spark.SparkContext.clean(SparkContext.scala:2294)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094)
>>         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.collect(RDD.scala:935)
>>         at 
>> org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:278)
>>         at 
>> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2439)
>>         at 
>> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2438)
>>         at org.apache.spark.sql.Dataset$$anonfun$55.apply(Dataset.scala:2846)
>>         at 
>> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65)
>>         at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2845)
>>         at org.apache.spark.sql.Dataset.count(Dataset.scala:2438)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>         at 
>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.base/java.lang.reflect.Method.invoke(Method.java:564)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
>>         at 
>> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
>>         at 
>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>>         at 
>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
>>         at 
>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>>         at java.base/java.lang.Thread.run(Thread.java:844)
>>     1: expect_equal(count(df), i) at
>> /srv/hornik/tmp/CRAN/SparkR.Rcheck/SparkR/tests/testthat/test_basic.R:26
>>     2: quasi_label(enquo(object), label)
>>     3: eval_bare(get_expr(quo), get_env(quo))
>>     4: count(df)
>>     5: count(df)
>>     6: callJMethod(x@sdf, "count")
>>     7: invokeJava(isStatic = FALSE, objId$id, methodName, ...)
>>     8: handleErrors(returnStatus, conn)
>>     9: stop(readString(conn))
>>
>>     18/07/09 17:58:54 ERROR RBackendHandler: fit on
>> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper failed
>>     java.lang.reflect.InvocationTargetException
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>         at 
>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.base/java.lang.reflect.Method.invoke(Method.java:564)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
>>         at 
>> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
>>         at 
>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>>         at 
>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
>>         at 
>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>>         at java.base/java.lang.Thread.run(Thread.java:844)
>>     Caused by: java.lang.IllegalArgumentException
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426)
>>         at 
>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103)
>>         at 
>> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426)
>>         at org.apache.xbean.asm5.ClassReader.a(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.b(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256)
>>         at scala.collection.immutable.List.foreach(List.scala:381)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156)
>>         at org.apache.spark.SparkContext.clean(SparkContext.scala:2294)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094)
>>         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.collect(RDD.scala:935)
>>         at 
>> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373)
>>         at 
>> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at 
>> org.apache.spark.rdd.PairRDDFunctions.countByKey(PairRDDFunctions.scala:372)
>>         at 
>> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204)
>>         at 
>> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.countByValue(RDD.scala:1203)
>>         at 
>> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:113)
>>         at 
>> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:88)
>>         at 
>> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:153)
>>         at 
>> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:149)
>>         at scala.collection.Iterator$class.foreach(Iterator.scala:893)
>>         at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
>>         at 
>> scala.collection.IterableViewLike$Transformed$class.foreach(IterableViewLike.scala:44)
>>         at 
>> scala.collection.SeqViewLike$AbstractTransformed.foreach(SeqViewLike.scala:37)
>>         at org.apache.spark.ml.Pipeline.fit(Pipeline.scala:149)
>>         at org.apache.spark.ml.feature.RFormula.fit(RFormula.scala:198)
>>         at 
>> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper$.fit(GeneralizedLinearRegressionWrapper.scala:81)
>>         at 
>> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper.fit(GeneralizedLinearRegressionWrapper.scala)
>>         ... 36 more
>>     ── 2. Error: spark.glm and predict (@test_basic.R#58)
>> ─────────────────────────
>>     java.lang.IllegalArgumentException
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426)
>>         at 
>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103)
>>         at 
>> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426)
>>         at org.apache.xbean.asm5.ClassReader.a(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.b(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256)
>>         at scala.collection.immutable.List.foreach(List.scala:381)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156)
>>         at org.apache.spark.SparkContext.clean(SparkContext.scala:2294)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094)
>>         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.collect(RDD.scala:935)
>>         at 
>> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373)
>>         at 
>> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at 
>> org.apache.spark.rdd.PairRDDFunctions.countByKey(PairRDDFunctions.scala:372)
>>         at 
>> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204)
>>         at 
>> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.countByValue(RDD.scala:1203)
>>         at 
>> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:113)
>>         at 
>> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:88)
>>         at 
>> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:153)
>>         at 
>> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:149)
>>         at scala.collection.Iterator$class.foreach(Iterator.scala:893)
>>         at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
>>         at 
>> scala.collection.IterableViewLike$Transformed$class.foreach(IterableViewLike.scala:44)
>>         at 
>> scala.collection.SeqViewLike$AbstractTransformed.foreach(SeqViewLike.scala:37)
>>         at org.apache.spark.ml.Pipeline.fit(Pipeline.scala:149)
>>         at org.apache.spark.ml.feature.RFormula.fit(RFormula.scala:198)
>>         at 
>> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper$.fit(GeneralizedLinearRegressionWrapper.scala:81)
>>         at 
>> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper.fit(GeneralizedLinearRegressionWrapper.scala)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>         at 
>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.base/java.lang.reflect.Method.invoke(Method.java:564)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
>>         at 
>> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
>>         at 
>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>>         at 
>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
>>         at 
>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>>         at java.base/java.lang.Thread.run(Thread.java:844)
>>     1: spark.glm(training, Sepal_Width ~ Sepal_Length + Species) at
>> /srv/hornik/tmp/CRAN/SparkR.Rcheck/SparkR/tests/testthat/test_basic.R:58
>>     2: spark.glm(training, Sepal_Width ~ Sepal_Length + Species)
>>     3: .local(data, formula, ...)
>>     4: 
>> callJStatic("org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper",
>> "fit", formula,
>>            data@sdf, tolower(family$family), family$link, tol,
>> as.integer(maxIter), weightCol,
>>            regParam, as.double(var.power), as.double(link.power))
>>     5: invokeJava(isStatic = TRUE, className, methodName, ...)
>>     6: handleErrors(returnStatus, conn)
>>     7: stop(readString(conn))
>>
>>     ══ testthat results
>> ═══════════════════════════════════════════════════════════
>>     OK: 0 SKIPPED: 0 FAILED: 2
>>     1. Error: create DataFrame from list or data.frame (@test_basic.R#26)
>>     2. Error: spark.glm and predict (@test_basic.R#58)
>>
>>     Error: testthat unit tests failed
>>     Execution halted
>>
>> Flavor: r-devel-linux-x86_64-debian-gcc
>> Check: re-building of vignette outputs, Result: WARNING
>>   Error in re-building vignettes:
>>     ...
>>
>>   Attaching package: 'SparkR'
>>
>>   The following objects are masked from 'package:stats':
>>
>>       cov, filter, lag, na.omit, predict, sd, var, window
>>
>>   The following objects are masked from 'package:base':
>>
>>       as.data.frame, colnames, colnames<-, drop, endsWith,
>>       intersect, rank, rbind, sample, startsWith, subset, summary,
>>       transform, union
>>
>>   Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>   Picked up _JAVA_OPTIONS: -XX:-UsePerfData
>>   Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>>   Setting default log level to "WARN".
>>   To adjust logging level use sc.setLogLevel(newLevel). For SparkR,
>> use setLogLevel(newLevel).
>>   WARNING: An illegal reflective access operation has occurred
>>   WARNING: Illegal reflective access by
>> io.netty.util.internal.PlatformDependent0$1
>> (file:/home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/jars/netty-all-4.0.43.Final.jar)
>> to field java.nio.Buffer.address
>>   WARNING: Please consider reporting this to the maintainers of
>> io.netty.util.internal.PlatformDependent0$1
>>   WARNING: Use --illegal-access=warn to enable warnings of further
>> illegal reflective access operations
>>   WARNING: All illegal access operations will be denied in a future release
>>   18/07/09 17:58:59 WARN NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes
>> where applicable
>>   18/07/09 17:59:07 ERROR RBackendHandler: dfToCols on
>> org.apache.spark.sql.api.r.SQLUtils failed
>>   java.lang.reflect.InvocationTargetException
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>         at 
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>         at 
>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>         at java.base/java.lang.reflect.Method.invoke(Method.java:564)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108)
>>         at 
>> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40)
>>         at 
>> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
>>         at 
>> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336)
>>         at 
>> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357)
>>         at 
>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343)
>>         at 
>> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911)
>>         at 
>> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
>>         at 
>> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
>>         at 
>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
>>         at 
>> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
>>         at java.base/java.lang.Thread.run(Thread.java:844)
>>   Caused by: java.lang.IllegalArgumentException
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426)
>>         at 
>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103)
>>         at 
>> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426)
>>         at org.apache.xbean.asm5.ClassReader.a(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.b(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.accept(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257)
>>         at 
>> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256)
>>         at scala.collection.immutable.List.foreach(List.scala:381)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156)
>>         at org.apache.spark.SparkContext.clean(SparkContext.scala:2294)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068)
>>         at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094)
>>         at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>>         at 
>> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>>         at org.apache.spark.rdd.RDD.withScope(RDD.scala:362)
>>         at org.apache.spark.rdd.RDD.collect(RDD.scala:935)
>>         at 
>> org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:278)
>>         at 
>> org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$collectFromPlan(Dataset.scala:2865)
>>         at 
>> org.apache.spark.sql.Dataset$$anonfun$collect$1.apply(Dataset.scala:2391)
>>         at 
>> org.apache.spark.sql.Dataset$$anonfun$collect$1.apply(Dataset.scala:2391)
>>         at org.apache.spark.sql.Dataset$$anonfun$55.apply(Dataset.scala:2846)
>>         at 
>> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65)
>>         at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2845)
>>         at org.apache.spark.sql.Dataset.collect(Dataset.scala:2391)
>>         at org.apache.spark.sql.api.r.SQLUtils$.dfToCols(SQLUtils.scala:212)
>>         at org.apache.spark.sql.api.r.SQLUtils.dfToCols(SQLUtils.scala)
>>         ... 36 more
>>   Quitting from lines 102-104 (sparkr-vignettes.Rmd)
>>   Error: processing vignette 'sparkr-vignettes.Rmd' failed with diagnostics:
>>   java.lang.IllegalArgumentException
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source)
>>         at 
>> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443)
>>         at 
>> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426)
>>         at 
>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103)
>>         at 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230)
>>         at scala.collection.mutable.HashMap.fore
>>   Execution halted
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to