I dont think we need to respin 2.2.2 -- Given that 2.3.2 is on the way we can just submit that.
Shivaram On Mon, Jul 9, 2018 at 6:19 PM Tom Graves <tgraves...@yahoo.com> wrote: > > is there anyway to push it to CRAN without this fix, I don't really want to > respin 2.2.2 just with the test fix. > > Tom > > On Monday, July 9, 2018, 4:50:18 PM CDT, Shivaram Venkataraman > <shiva...@eecs.berkeley.edu> wrote: > > > Yes. I think Felix checked in a fix to ignore tests run on java > versions that are not Java 8 (I think the fix was in > https://github.com/apache/spark/pull/21666 which is in 2.3.2) > > Shivaram > On Mon, Jul 9, 2018 at 5:39 PM Sean Owen <sro...@gmail.com> wrote: > > > > Yes, this flavor of error should only come up in Java 9. Spark doesn't > > support that. Is there any way to tell CRAN this should not be tested? > > > > On Mon, Jul 9, 2018, 4:17 PM Shivaram Venkataraman > > <shiva...@eecs.berkeley.edu> wrote: > >> > >> The upcoming 2.2.2 release was submitted to CRAN. I think there are > >> some knows issues on Windows, but does anybody know what the following > >> error with Netty is ? > >> > >> > WARNING: Illegal reflective access by > >> > io.netty.util.internal.PlatformDependent0$1 > >> > (file:/home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/jars/netty-all-4.0.43.Final.jar) > >> > to field java.nio.Buffer.address > >> > >> Thanks > >> Shivaram > >> > >> > >> ---------- Forwarded message --------- > >> From: <lig...@statistik.tu-dortmund.de> > >> Date: Mon, Jul 9, 2018 at 12:12 PM > >> Subject: [CRAN-pretest-archived] CRAN submission SparkR 2.2.2 > >> To: <shiva...@cs.berkeley.edu> > >> Cc: <cran-submissi...@r-project.org> > >> > >> > >> Dear maintainer, > >> > >> package SparkR_2.2.2.tar.gz does not pass the incoming checks > >> automatically, please see the following pre-tests: > >> Windows: > >> <https://win-builder.r-project.org/incoming_pretest/SparkR_2.2.2_20180709_175630/Windows/00check.log> > >> Status: 1 ERROR, 1 WARNING > >> Debian: > >> <https://win-builder.r-project.org/incoming_pretest/SparkR_2.2.2_20180709_175630/Debian/00check.log> > >> Status: 1 ERROR, 2 WARNINGs > >> > >> Last released version's CRAN status: ERROR: 1, OK: 1 > >> See: <https://CRAN.R-project.org/web/checks/check_results_SparkR.html> > >> > >> CRAN Web: <https://cran.r-project.org/package=SparkR> > >> > >> Please fix all problems and resubmit a fixed version via the webform. > >> If you are not sure how to fix the problems shown, please ask for help > >> on the R-package-devel mailing list: > >> <https://stat.ethz.ch/mailman/listinfo/r-package-devel> > >> If you are fairly certain the rejection is a false positive, please > >> reply-all to this message and explain. > >> > >> More details are given in the directory: > >> <https://win-builder.r-project.org/incoming_pretest/SparkR_2.2.2_20180709_175630/> > >> The files will be removed after roughly 7 days. > >> > >> No strong reverse dependencies to be checked. > >> > >> Best regards, > >> CRAN teams' auto-check service > >> Flavor: r-devel-linux-x86_64-debian-gcc, r-devel-windows-ix86+x86_64 > >> Check: CRAN incoming feasibility, Result: WARNING > >> Maintainer: 'Shivaram Venkataraman <shiva...@cs.berkeley.edu>' > >> > >> New submission > >> > >> Package was archived on CRAN > >> > >> Insufficient package version (submitted: 2.2.2, existing: 2.3.0) > >> > >> Possibly mis-spelled words in DESCRIPTION: > >> Frontend (4:10, 5:28) > >> > >> CRAN repository db overrides: > >> X-CRAN-Comment: Archived on 2018-05-01 as check problems were not > >> corrected despite reminders. > >> > >> Found the following (possibly) invalid URLs: > >> URL: http://spark.apache.org/docs/latest/api/R/mean.html > >> From: inst/doc/sparkr-vignettes.html > >> Status: 404 > >> Message: Not Found > >> > >> Flavor: r-devel-windows-ix86+x86_64 > >> Check: running tests for arch 'x64', Result: ERROR > >> Running 'run-all.R' [175s] > >> Running the tests in 'tests/run-all.R' failed. > >> Complete output: > >> > # > >> > # Licensed to the Apache Software Foundation (ASF) under one or more > >> > # contributor license agreements. See the NOTICE file distributed > >> with > >> > # this work for additional information regarding copyright ownership. > >> > # The ASF licenses this file to You under the Apache License, Version > >> 2.0 > >> > # (the "License"); you may not use this file except in compliance with > >> > # the License. You may obtain a copy of the License at > >> > # > >> > # http://www.apache.org/licenses/LICENSE-2.0 > >> > # > >> > # Unless required by applicable law or agreed to in writing, software > >> > # distributed under the License is distributed on an "AS IS" BASIS, > >> > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or > >> implied. > >> > # See the License for the specific language governing permissions and > >> > # limitations under the License. > >> > # > >> > > >> > library(testthat) > >> > library(SparkR) > >> > >> Attaching package: 'SparkR' > >> > >> The following object is masked from 'package:testthat': > >> > >> describe > >> > >> The following objects are masked from 'package:stats': > >> > >> cov, filter, lag, na.omit, predict, sd, var, window > >> > >> The following objects are masked from 'package:base': > >> > >> as.data.frame, colnames, colnames<-, drop, endsWith, intersect, > >> rank, rbind, sample, startsWith, subset, summary, transform, union > >> > >> > > >> > # Turn all warnings into errors > >> > options("warn" = 2) > >> > > >> > if (.Platform$OS.type == "windows") { > >> + Sys.setenv(TZ = "GMT") > >> + } > >> > > >> > # Setup global test environment > >> > # Install Spark first to set SPARK_HOME > >> > > >> > # NOTE(shivaram): We set overwrite to handle any old tar.gz > >> files or directories left behind on > >> > # CRAN machines. For Jenkins we should already have SPARK_HOME set. > >> > install.spark(overwrite = TRUE) > >> Overwrite = TRUE: download and overwrite the tar fileand Spark > >> package directory if they exist. > >> Spark not found in the cache directory. Installation will start. > >> MirrorUrl not provided. > >> Looking for preferred site from apache website... > >> Preferred mirror site found: http://mirror.dkd.de/apache/spark > >> Downloading spark-2.2.2 for Hadoop 2.7 from: > >> - > >> http://mirror.dkd.de/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz > >> trying URL > >> 'http://mirror.dkd.de/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz' > >> Content type 'application/x-gzip' length 200743115 bytes (191.4 MB) > >> ================================================== > >> downloaded 191.4 MB > >> > >> Installing to C:\Users\ligges\AppData\Local\Apache\Spark\Cache > >> DONE. > >> SPARK_HOME set to > >> C:\Users\ligges\AppData\Local\Apache\Spark\Cache/spark-2.2.2-bin-hadoop2.7 > >> > > >> > sparkRDir <- file.path(Sys.getenv("SPARK_HOME"), "R") > >> > sparkRWhitelistSQLDirs <- c("spark-warehouse", "metastore_db") > >> > invisible(lapply(sparkRWhitelistSQLDirs, > >> + function(x) { unlink(file.path(sparkRDir, x), > >> recursive = TRUE, force = TRUE)})) > >> > sparkRFilesBefore <- list.files(path = sparkRDir, all.files = TRUE) > >> > > >> > sparkRTestMaster <- "local[1]" > >> > sparkRTestConfig <- list() > >> > if (identical(Sys.getenv("NOT_CRAN"), "true")) { > >> + sparkRTestMaster <- "" > >> + } else { > >> + # Disable hsperfdata on CRAN > >> + old_java_opt <- Sys.getenv("_JAVA_OPTIONS") > >> + Sys.setenv("_JAVA_OPTIONS" = paste("-XX:-UsePerfData", old_java_opt)) > >> + tmpDir <- tempdir() > >> + tmpArg <- paste0("-Djava.io.tmpdir=", tmpDir) > >> + sparkRTestConfig <- list(spark.driver.extraJavaOptions = tmpArg, > >> + spark.executor.extraJavaOptions = tmpArg) > >> + } > >> > > >> > test_package("SparkR") > >> Launching java with spark-submit command > >> C:\Users\ligges\AppData\Local\Apache\Spark\Cache/spark-2.2.2-bin-hadoop2.7/bin/spark-submit2.cmd > >> --driver-java-options "-Djava.io.tmpdir=D:\temp\RtmpABZLQj" > >> sparkr-shell D:\temp\RtmpABZLQj\backend_port16d0838283f7e > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> Using Spark's default log4j profile: > >> org/apache/spark/log4j-defaults.properties > >> Setting default log level to "WARN". > >> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, > >> use setLogLevel(newLevel). > >> -- 1. Error: create DataFrame from list or data.frame > >> (@test_basic.R#21) ------ > >> cannot open the connection > >> 1: sparkR.session(master = sparkRTestMaster, enableHiveSupport = > >> FALSE, sparkConfig = sparkRTestConfig) at > >> D:/temp/Rtmp8IKu99/RLIBS_77d8215b7bce/SparkR/tests/testthat/test_basic.R:21 > >> 2: sparkR.sparkContext(master, appName, sparkHome, sparkConfigMap, > >> sparkExecutorEnvMap, > >> sparkJars, sparkPackages) > >> 3: file(path, open = "rb") > >> > >> Launching java with spark-submit command > >> C:\Users\ligges\AppData\Local\Apache\Spark\Cache/spark-2.2.2-bin-hadoop2.7/bin/spark-submit2.cmd > >> --driver-java-options "-Djava.io.tmpdir=D:\temp\RtmpABZLQj" > >> sparkr-shell D:\temp\RtmpABZLQj\backend_port16d085df97d88 > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> Using Spark's default log4j profile: > >> org/apache/spark/log4j-defaults.properties > >> Setting default log level to "WARN". > >> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, > >> use setLogLevel(newLevel). > >> 18/07/09 18:10:43 ERROR Shell: Failed to locate the winutils > >> binary in the hadoop binary path > >> java.io.IOException: Could not locate executable > >> null\bin\winutils.exe in the Hadoop binaries. > >> at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:379) > >> at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:394) > >> at org.apache.hadoop.util.Shell.<clinit>(Shell.java:387) > >> at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:80) > >> at > >> org.apache.hadoop.security.SecurityUtil.getAuthenticationMethod(SecurityUtil.java:611) > >> at > >> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:273) > >> at > >> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:261) > >> at > >> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:791) > >> at > >> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:761) > >> at > >> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:634) > >> at > >> org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2427) > >> at > >> org.apache.spark.util.Utils$$anonfun$getCurrentUserName$1.apply(Utils.scala:2427) > >> at scala.Option.getOrElse(Option.scala:121) > >> at org.apache.spark.util.Utils$.getCurrentUserName(Utils.scala:2427) > >> at org.apache.spark.SparkContext.<init>(SparkContext.scala:295) > >> at > >> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2516) > >> at org.apache.spark.api.r.RRDD$.createSparkContext(RRDD.scala:139) > >> at org.apache.spark.api.r.RRDD.createSparkContext(RRDD.scala) > >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > >> at > >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >> at > >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >> at java.lang.reflect.Method.invoke(Method.java:498) > >> at > >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40) > >> at > >> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) > >> at > >> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) > >> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) > >> at > >> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) > >> at > >> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144) > >> at java.lang.Thread.run(Thread.java:748) > >> 18/07/09 18:10:43 WARN NativeCodeLoader: Unable to load > >> native-hadoop library for your platform... using builtin-java classes > >> where applicable > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> 18/07/09 18:10:54 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:10:55 WARN BLAS: Failed to load implementation from: > >> com.github.fommil.netlib.NativeSystemBLAS > >> 18/07/09 18:10:55 WARN BLAS: Failed to load implementation from: > >> com.github.fommil.netlib.NativeRefBLAS > >> 18/07/09 18:10:55 WARN LAPACK: Failed to load implementation from: > >> com.github.fommil.netlib.NativeSystemLAPACK > >> 18/07/09 18:10:55 WARN LAPACK: Failed to load implementation from: > >> com.github.fommil.netlib.NativeRefLAPACK > >> 18/07/09 18:11:12 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:14 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:15 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:17 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:18 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:19 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:21 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:22 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:23 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:25 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:26 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:28 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:29 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:46 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:47 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:49 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> 18/07/09 18:11:50 WARN WeightedLeastSquares: regParam is zero, > >> which might cause numerical instability and overfitting. > >> == testthat results > >> =========================================================== > >> OK: 6 SKIPPED: 0 FAILED: 1 > >> 1. Error: create DataFrame from list or data.frame (@test_basic.R#21) > >> > >> Error: testthat unit tests failed > >> Execution halted > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> > >> Flavor: r-devel-linux-x86_64-debian-gcc > >> Check: tests, Result: ERROR > >> Running 'run-all.R' [6s/15s] > >> Running the tests in 'tests/run-all.R' failed. > >> Complete output: > >> > # > >> > # Licensed to the Apache Software Foundation (ASF) under one or more > >> > # contributor license agreements. See the NOTICE file distributed > >> with > >> > # this work for additional information regarding copyright ownership. > >> > # The ASF licenses this file to You under the Apache License, Version > >> 2.0 > >> > # (the "License"); you may not use this file except in compliance with > >> > # the License. You may obtain a copy of the License at > >> > # > >> > # http://www.apache.org/licenses/LICENSE-2.0 > >> > # > >> > # Unless required by applicable law or agreed to in writing, software > >> > # distributed under the License is distributed on an "AS IS" BASIS, > >> > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or > >> implied. > >> > # See the License for the specific language governing permissions and > >> > # limitations under the License. > >> > # > >> > > >> > library(testthat) > >> > library(SparkR) > >> > >> Attaching package: 'SparkR' > >> > >> The following object is masked from 'package:testthat': > >> > >> describe > >> > >> The following objects are masked from 'package:stats': > >> > >> cov, filter, lag, na.omit, predict, sd, var, window > >> > >> The following objects are masked from 'package:base': > >> > >> as.data.frame, colnames, colnames<-, drop, endsWith, intersect, > >> rank, rbind, sample, startsWith, subset, summary, transform, union > >> > >> > > >> > # Turn all warnings into errors > >> > options("warn" = 2) > >> > > >> > if (.Platform$OS.type == "windows") { > >> + Sys.setenv(TZ = "GMT") > >> + } > >> > > >> > # Setup global test environment > >> > # Install Spark first to set SPARK_HOME > >> > > >> > # NOTE(shivaram): We set overwrite to handle any old tar.gz > >> files or directories left behind on > >> > # CRAN machines. For Jenkins we should already have SPARK_HOME set. > >> > install.spark(overwrite = TRUE) > >> Overwrite = TRUE: download and overwrite the tar fileand Spark > >> package directory if they exist. > >> Spark not found in the cache directory. Installation will start. > >> MirrorUrl not provided. > >> Looking for preferred site from apache website... > >> Preferred mirror site found: http://mirror.klaus-uwe.me/apache/spark > >> Downloading spark-2.2.2 for Hadoop 2.7 from: > >> - > >> http://mirror.klaus-uwe.me/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz > >> trying URL > >> 'http://mirror.klaus-uwe.me/apache/spark/spark-2.2.2/spark-2.2.2-bin-hadoop2.7.tgz' > >> Content type 'application/octet-stream' length 200743115 bytes (191.4 > >> MB) > >> ================================================== > >> downloaded 191.4 MB > >> > >> Installing to /home/hornik/.cache/spark > >> DONE. > >> SPARK_HOME set to /home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7 > >> > > >> > sparkRDir <- file.path(Sys.getenv("SPARK_HOME"), "R") > >> > sparkRWhitelistSQLDirs <- c("spark-warehouse", "metastore_db") > >> > invisible(lapply(sparkRWhitelistSQLDirs, > >> + function(x) { unlink(file.path(sparkRDir, x), > >> recursive = TRUE, force = TRUE)})) > >> > sparkRFilesBefore <- list.files(path = sparkRDir, all.files = TRUE) > >> > > >> > sparkRTestMaster <- "local[1]" > >> > sparkRTestConfig <- list() > >> > if (identical(Sys.getenv("NOT_CRAN"), "true")) { > >> + sparkRTestMaster <- "" > >> + } else { > >> + # Disable hsperfdata on CRAN > >> + old_java_opt <- Sys.getenv("_JAVA_OPTIONS") > >> + Sys.setenv("_JAVA_OPTIONS" = paste("-XX:-UsePerfData", old_java_opt)) > >> + tmpDir <- tempdir() > >> + tmpArg <- paste0("-Djava.io.tmpdir=", tmpDir) > >> + sparkRTestConfig <- list(spark.driver.extraJavaOptions = tmpArg, > >> + spark.executor.extraJavaOptions = tmpArg) > >> + } > >> > > >> > test_package("SparkR") > >> Launching java with spark-submit command > >> /home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/bin/spark-submit > >> --driver-java-options "-Djava.io.tmpdir=/tmp/Rtmpkd8Lf6" sparkr-shell > >> /tmp/Rtmpkd8Lf6/backend_port289f65a5f5e0 > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> Using Spark's default log4j profile: > >> org/apache/spark/log4j-defaults.properties > >> Setting default log level to "WARN". > >> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, > >> use setLogLevel(newLevel). > >> WARNING: An illegal reflective access operation has occurred > >> WARNING: Illegal reflective access by > >> io.netty.util.internal.PlatformDependent0$1 > >> (file:/home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/jars/netty-all-4.0.43.Final.jar) > >> to field java.nio.Buffer.address > >> WARNING: Please consider reporting this to the maintainers of > >> io.netty.util.internal.PlatformDependent0$1 > >> WARNING: Use --illegal-access=warn to enable warnings of further > >> illegal reflective access operations > >> WARNING: All illegal access operations will be denied in a future > >> release > >> 18/07/09 17:58:50 WARN NativeCodeLoader: Unable to load > >> native-hadoop library for your platform... using builtin-java classes > >> where applicable > >> 18/07/09 17:58:54 ERROR RBackendHandler: count on 13 failed > >> java.lang.reflect.InvocationTargetException > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > >> Method) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >> at > >> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >> at java.base/java.lang.reflect.Method.invoke(Method.java:564) > >> at > >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40) > >> at > >> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) > >> at > >> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) > >> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) > >> at > >> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) > >> at > >> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144) > >> at java.base/java.lang.Thread.run(Thread.java:844) > >> Caused by: java.lang.IllegalArgumentException > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426) > >> at > >> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230) > >> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > >> at > >> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103) > >> at > >> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426) > >> at org.apache.xbean.asm5.ClassReader.a(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.b(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256) > >> at scala.collection.immutable.List.foreach(List.scala:381) > >> at > >> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256) > >> at > >> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156) > >> at org.apache.spark.SparkContext.clean(SparkContext.scala:2294) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094) > >> at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at org.apache.spark.rdd.RDD.collect(RDD.scala:935) > >> at > >> org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:278) > >> at > >> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2439) > >> at > >> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2438) > >> at > >> org.apache.spark.sql.Dataset$$anonfun$55.apply(Dataset.scala:2846) > >> at > >> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65) > >> at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2845) > >> at org.apache.spark.sql.Dataset.count(Dataset.scala:2438) > >> ... 36 more > >> ── 1. Error: create DataFrame from list or data.frame > >> (@test_basic.R#26) ────── > >> java.lang.IllegalArgumentException > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426) > >> at > >> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230) > >> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > >> at > >> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103) > >> at > >> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426) > >> at org.apache.xbean.asm5.ClassReader.a(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.b(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256) > >> at scala.collection.immutable.List.foreach(List.scala:381) > >> at > >> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256) > >> at > >> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156) > >> at org.apache.spark.SparkContext.clean(SparkContext.scala:2294) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094) > >> at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at org.apache.spark.rdd.RDD.collect(RDD.scala:935) > >> at > >> org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:278) > >> at > >> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2439) > >> at > >> org.apache.spark.sql.Dataset$$anonfun$count$1.apply(Dataset.scala:2438) > >> at > >> org.apache.spark.sql.Dataset$$anonfun$55.apply(Dataset.scala:2846) > >> at > >> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65) > >> at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2845) > >> at org.apache.spark.sql.Dataset.count(Dataset.scala:2438) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > >> Method) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >> at > >> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >> at java.base/java.lang.reflect.Method.invoke(Method.java:564) > >> at > >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40) > >> at > >> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) > >> at > >> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) > >> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) > >> at > >> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) > >> at > >> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144) > >> at java.base/java.lang.Thread.run(Thread.java:844) > >> 1: expect_equal(count(df), i) at > >> /srv/hornik/tmp/CRAN/SparkR.Rcheck/SparkR/tests/testthat/test_basic.R:26 > >> 2: quasi_label(enquo(object), label) > >> 3: eval_bare(get_expr(quo), get_env(quo)) > >> 4: count(df) > >> 5: count(df) > >> 6: callJMethod(x@sdf, "count") > >> 7: invokeJava(isStatic = FALSE, objId$id, methodName, ...) > >> 8: handleErrors(returnStatus, conn) > >> 9: stop(readString(conn)) > >> > >> 18/07/09 17:58:54 ERROR RBackendHandler: fit on > >> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper failed > >> java.lang.reflect.InvocationTargetException > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > >> Method) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >> at > >> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >> at java.base/java.lang.reflect.Method.invoke(Method.java:564) > >> at > >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40) > >> at > >> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) > >> at > >> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) > >> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) > >> at > >> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) > >> at > >> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144) > >> at java.base/java.lang.Thread.run(Thread.java:844) > >> Caused by: java.lang.IllegalArgumentException > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426) > >> at > >> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230) > >> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > >> at > >> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103) > >> at > >> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426) > >> at org.apache.xbean.asm5.ClassReader.a(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.b(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256) > >> at scala.collection.immutable.List.foreach(List.scala:381) > >> at > >> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256) > >> at > >> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156) > >> at org.apache.spark.SparkContext.clean(SparkContext.scala:2294) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094) > >> at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at org.apache.spark.rdd.RDD.collect(RDD.scala:935) > >> at > >> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373) > >> at > >> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at > >> org.apache.spark.rdd.PairRDDFunctions.countByKey(PairRDDFunctions.scala:372) > >> at > >> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204) > >> at > >> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at org.apache.spark.rdd.RDD.countByValue(RDD.scala:1203) > >> at > >> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:113) > >> at > >> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:88) > >> at > >> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:153) > >> at > >> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:149) > >> at scala.collection.Iterator$class.foreach(Iterator.scala:893) > >> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) > >> at > >> scala.collection.IterableViewLike$Transformed$class.foreach(IterableViewLike.scala:44) > >> at > >> scala.collection.SeqViewLike$AbstractTransformed.foreach(SeqViewLike.scala:37) > >> at org.apache.spark.ml.Pipeline.fit(Pipeline.scala:149) > >> at org.apache.spark.ml.feature.RFormula.fit(RFormula.scala:198) > >> at > >> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper$.fit(GeneralizedLinearRegressionWrapper.scala:81) > >> at > >> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper.fit(GeneralizedLinearRegressionWrapper.scala) > >> ... 36 more > >> ── 2. Error: spark.glm and predict (@test_basic.R#58) > >> ───────────────────────── > >> java.lang.IllegalArgumentException > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.<init>(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:46) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:443) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:426) > >> at > >> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:103) > >> at > >> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:230) > >> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) > >> at > >> scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:103) > >> at > >> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732) > >> at > >> org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:426) > >> at org.apache.xbean.asm5.ClassReader.a(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.b(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at org.apache.xbean.asm5.ClassReader.accept(Unknown Source) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:257) > >> at > >> org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:256) > >> at scala.collection.immutable.List.foreach(List.scala:381) > >> at > >> org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:256) > >> at > >> org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:156) > >> at org.apache.spark.SparkContext.clean(SparkContext.scala:2294) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2068) > >> at org.apache.spark.SparkContext.runJob(SparkContext.scala:2094) > >> at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:936) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at org.apache.spark.rdd.RDD.collect(RDD.scala:935) > >> at > >> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373) > >> at > >> org.apache.spark.rdd.PairRDDFunctions$$anonfun$countByKey$1.apply(PairRDDFunctions.scala:373) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at > >> org.apache.spark.rdd.PairRDDFunctions.countByKey(PairRDDFunctions.scala:372) > >> at > >> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204) > >> at > >> org.apache.spark.rdd.RDD$$anonfun$countByValue$1.apply(RDD.scala:1204) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > >> at > >> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > >> at org.apache.spark.rdd.RDD.withScope(RDD.scala:362) > >> at org.apache.spark.rdd.RDD.countByValue(RDD.scala:1203) > >> at > >> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:113) > >> at > >> org.apache.spark.ml.feature.StringIndexer.fit(StringIndexer.scala:88) > >> at > >> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:153) > >> at > >> org.apache.spark.ml.Pipeline$$anonfun$fit$2.apply(Pipeline.scala:149) > >> at scala.collection.Iterator$class.foreach(Iterator.scala:893) > >> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) > >> at > >> scala.collection.IterableViewLike$Transformed$class.foreach(IterableViewLike.scala:44) > >> at > >> scala.collection.SeqViewLike$AbstractTransformed.foreach(SeqViewLike.scala:37) > >> at org.apache.spark.ml.Pipeline.fit(Pipeline.scala:149) > >> at org.apache.spark.ml.feature.RFormula.fit(RFormula.scala:198) > >> at > >> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper$.fit(GeneralizedLinearRegressionWrapper.scala:81) > >> at > >> org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper.fit(GeneralizedLinearRegressionWrapper.scala) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > >> Method) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >> at > >> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >> at java.base/java.lang.reflect.Method.invoke(Method.java:564) > >> at > >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40) > >> at > >> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1294) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:911) > >> at > >> io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:131) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:643) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566) > >> at > >> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480) > >> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442) > >> at > >> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131) > >> at > >> io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144) > >> at java.base/java.lang.Thread.run(Thread.java:844) > >> 1: spark.glm(training, Sepal_Width ~ Sepal_Length + Species) at > >> /srv/hornik/tmp/CRAN/SparkR.Rcheck/SparkR/tests/testthat/test_basic.R:58 > >> 2: spark.glm(training, Sepal_Width ~ Sepal_Length + Species) > >> 3: .local(data, formula, ...) > >> 4: > >> callJStatic("org.apache.spark.ml.r.GeneralizedLinearRegressionWrapper", > >> "fit", formula, > >> data@sdf, tolower(family$family), family$link, tol, > >> as.integer(maxIter), weightCol, > >> regParam, as.double(var.power), as.double(link.power)) > >> 5: invokeJava(isStatic = TRUE, className, methodName, ...) > >> 6: handleErrors(returnStatus, conn) > >> 7: stop(readString(conn)) > >> > >> ══ testthat results > >> ═══════════════════════════════════════════════════════════ > >> OK: 0 SKIPPED: 0 FAILED: 2 > >> 1. Error: create DataFrame from list or data.frame (@test_basic.R#26) > >> 2. Error: spark.glm and predict (@test_basic.R#58) > >> > >> Error: testthat unit tests failed > >> Execution halted > >> > >> Flavor: r-devel-linux-x86_64-debian-gcc > >> Check: re-building of vignette outputs, Result: WARNING > >> Error in re-building vignettes: > >> ... > >> > >> Attaching package: 'SparkR' > >> > >> The following objects are masked from 'package:stats': > >> > >> cov, filter, lag, na.omit, predict, sd, var, window > >> > >> The following objects are masked from 'package:base': > >> > >> as.data.frame, colnames, colnames<-, drop, endsWith, > >> intersect, rank, rbind, sample, startsWith, subset, summary, > >> transform, union > >> > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> Picked up _JAVA_OPTIONS: -XX:-UsePerfData > >> Using Spark's default log4j profile: > >> org/apache/spark/log4j-defaults.properties > >> Setting default log level to "WARN". > >> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, > >> use setLogLevel(newLevel). > >> WARNING: An illegal reflective access operation has occurred > >> WARNING: Illegal reflective access by > >> io.netty.util.internal.PlatformDependent0$1 > >> (file:/home/hornik/.cache/spark/spark-2.2.2-bin-hadoop2.7/jars/netty-all-4.0.43.Final.jar) > >> to field java.nio.Buffer.address > >> WARNING: Please consider reporting this to the maintainers of > >> io.netty.util.internal.PlatformDependent0$1 > >> WARNING: Use --illegal-access=warn to enable warnings of further > >> illegal reflective access operations > >> WARNING: All illegal access operations will be denied in a future release > >> 18/07/09 17:58:59 WARN NativeCodeLoader: Unable to load > >> native-hadoop library for your platform... using builtin-java classes > >> where applicable > >> 18/07/09 17:59:07 ERROR RBackendHandler: dfToCols on > >> org.apache.spark.sql.api.r.SQLUtils failed > >> java.lang.reflect.InvocationTargetException > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > >> Method) > >> at > >> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >> at > >> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >> at java.base/java.lang.reflect.Method.invoke(Method.java:564) > >> at > >> org.apache.spark.api.r.RBackendHandler.handleMethodCall(RBackendHandler.scala:167) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:108) > >> at > >> org.apache.spark.api.r.RBackendHandler.channelRead0(RBackendHandler.scala:40) > >> at > >> io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.timeout.IdleStateHandler.channelRead(IdleStateHandler.java:287) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:102) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:357) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:343) > >> at > >> io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:336) > >> at > >> io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293 --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org