As R is single-threaded, SparkR launches one R process per-executor on
the worker side.

Thanks
Shivaram

On Thu, Sep 18, 2014 at 7:49 AM, oppokui <oppo...@gmail.com> wrote:
> Shivaram,
>
> As I know, SparkR used rJava package. In work node, spark code will execute R 
> code by launching R process and send/receive byte array.
> I have a question on when to launch R process. R process is per Work process, 
> or per executor thread, or per each RDD processing?
>
> Thanks and Regards.
>
> Kui
>
>> On Sep 6, 2014, at 5:53 PM, oppokui <oppo...@gmail.com> wrote:
>>
>> Cool! It is a very good news. Can’t wait for it.
>>
>> Kui
>>
>>> On Sep 5, 2014, at 1:58 AM, Shivaram Venkataraman 
>>> <shiva...@eecs.berkeley.edu> wrote:
>>>
>>> Thanks Kui. SparkR is a pretty young project, but there are a bunch of
>>> things we are working on. One of the main features is to expose a data
>>> frame API (https://sparkr.atlassian.net/browse/SPARKR-1) and we will
>>> be integrating this with Spark's MLLib.  At a high-level this will
>>> allow R users to use a familiar API but make use of MLLib's efficient
>>> distributed implementation. This is the same strategy used in Python
>>> as well.
>>>
>>> Also we do hope to merge SparkR with mainline Spark -- we have a few
>>> features to complete before that and plan to shoot for integration by
>>> Spark 1.3.
>>>
>>> Thanks
>>> Shivaram
>>>
>>> On Wed, Sep 3, 2014 at 9:24 PM, oppokui <oppo...@gmail.com> wrote:
>>>> Thanks, Shivaram.
>>>>
>>>> No specific use case yet. We try to use R in our project as data scientest
>>>> are all knowing R. We had a concern that how R handles the mass data. Spark
>>>> does a better work on big data area, and Spark ML is focusing on predictive
>>>> analysis area. Then we are thinking whether we can merge R and Spark
>>>> together. We tried SparkR and it is pretty easy to use. But we didn’t see
>>>> any feedback on this package in industry. It will be better if Spark team
>>>> has R support just like scala/Java/Python.
>>>>
>>>> Another question is that MLlib will re-implement all famous data mining
>>>> algorithms in Spark, then what is the purpose of using R?
>>>>
>>>> There is another technique for us H2O which support R natively. H2O is more
>>>> friendly to data scientist. I saw H2O can also work on Spark (Sparkling
>>>> Water).  It is better than using SparkR?
>>>>
>>>> Thanks and Regards.
>>>>
>>>> Kui
>>>>
>>>>
>>>> On Sep 4, 2014, at 1:47 AM, Shivaram Venkataraman
>>>> <shiva...@eecs.berkeley.edu> wrote:
>>>>
>>>> Hi
>>>>
>>>> Do you have a specific use-case where SparkR doesn't work well ? We'd love
>>>> to hear more about use-cases and features that can be improved with SparkR.
>>>>
>>>> Thanks
>>>> Shivaram
>>>>
>>>>
>>>> On Wed, Sep 3, 2014 at 3:19 AM, oppokui <oppo...@gmail.com> wrote:
>>>>>
>>>>> Does spark ML team have plan to support R script natively? There is a
>>>>> SparkR project, but not from spark team. Spark ML used netlib-java to talk
>>>>> with native fortran routines or use NumPy, why not try to use R in some
>>>>> sense.
>>>>>
>>>>> R had lot of useful packages. If spark ML team can include R support, it
>>>>> will be a very powerful.
>>>>>
>>>>> Any comment?
>>>>>
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>>>
>>>>
>>>>
>>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to