Yes, I guess so. But R 3.6.2 is just released this month, I think we should
use an older version to build SparkR.

Felix Cheung <felixcheun...@hotmail.com> 于2019年12月27日周五 上午10:43写道:

> Maybe it’s the reverse - the package is built to run in latest but not
> compatible with slightly older (3.5.2 was Dec 2018)
>
> ------------------------------
> *From:* Jeff Zhang <zjf...@gmail.com>
> *Sent:* Thursday, December 26, 2019 5:36:50 PM
> *To:* Felix Cheung <felixcheun...@hotmail.com>
> *Cc:* user.spark <user@spark.apache.org>
> *Subject:* Re: Fail to use SparkR of 3.0 preview 2
>
> I use R 3.5.2
>
> Felix Cheung <felixcheun...@hotmail.com> 于2019年12月27日周五 上午4:32写道:
>
> It looks like a change in the method signature in R base packages.
>
> Which version of R are you running on?
>
> ------------------------------
> *From:* Jeff Zhang <zjf...@gmail.com>
> *Sent:* Thursday, December 26, 2019 12:46:12 AM
> *To:* user.spark <user@spark.apache.org>
> *Subject:* Fail to use SparkR of 3.0 preview 2
>
> I tried SparkR of spark 3.0 preview 2, but hit the following issue.
>
> Error in rbind(info, getNamespaceInfo(env, "S3methods")) :
>   number of columns of matrices must match (see arg 2)
> Error: package or namespace load failed for ‘SparkR’ in rbind(info,
> getNamespaceInfo(env, "S3methods")):
>  number of columns of matrices must match (see arg 2)
> During startup - Warning messages:
> 1: package ‘SparkR’ was built under R version 3.6.2
> 2: package ‘SparkR’ in options("defaultPackages") was not found
>
> Does anyone know what might be wrong ? Thanks
>
>
>
> --
> Best Regards
>
> Jeff Zhang
>
>
>
> --
> Best Regards
>
> Jeff Zhang
>


-- 
Best Regards

Jeff Zhang

Reply via email to