[ 
https://issues.apache.org/jira/browse/SPARK-5223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Davies Liu updated SPARK-5223:
------------------------------
    Description: 
It will introduce problems if the object in dict/list/tuple can not support by 
py4j, such as Vector.

Also, pickle may have better performance for larger object (less RPC).

In some cases that the object in dict/list can not be pickled (such as 
JavaObject), we should still use MapConvert/ListConvert.

discussion: 
http://apache-spark-developers-list.1001551.n3.nabble.com/Python-to-Java-object-conversion-of-numpy-array-td10065.html

  was:
It will introduce problems if the object in dict/list/tuple can not support by 
py4j, such as Vector.

Also, pickle may have better performance for larger object (less RPC).

In some cases that the object in dict/list can not be pickled (such as 
JavaObject), we should still use MapConvert/ListConvert.


> Use pickle instead of MapConvert and ListConvert in MLlib Python API
> --------------------------------------------------------------------
>
>                 Key: SPARK-5223
>                 URL: https://issues.apache.org/jira/browse/SPARK-5223
>             Project: Spark
>          Issue Type: Bug
>          Components: MLlib, PySpark
>            Reporter: Davies Liu
>            Priority: Critical
>
> It will introduce problems if the object in dict/list/tuple can not support 
> by py4j, such as Vector.
> Also, pickle may have better performance for larger object (less RPC).
> In some cases that the object in dict/list can not be pickled (such as 
> JavaObject), we should still use MapConvert/ListConvert.
> discussion: 
> http://apache-spark-developers-list.1001551.n3.nabble.com/Python-to-Java-object-conversion-of-numpy-array-td10065.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to