Github user HyukjinKwon commented on the issue:

    https://github.com/apache/spark/pull/19498
  
    I just took a look. This looks actually going to bigger change because we 
will need manually deal with the different de/serializers cases in Python side 
as well, for example:
    
    ```
    >>> rdd = sc.parallelize(range(10)).cartesian(sc.parallelize(range(10)))
    >>> rdd._jrdd_deserializer = sc.serializer
    >>> rdd.collect()
    [0, 0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0, 8, 0, 9, 1, 0, 1, 1, 1, 
2, 1, 3, 1, 4, 1, 5, 1, 6, 1, 7, 1, 8, 1, 9, 2, 0, 2, 1, 2, 2, 2, 3, 2, 4, 2, 
5, 2, 6, 2, 7, 2, 8, 2, 9, 3, 0, 4, 0, 3, 1, 4, 1, 3, 2, 4, 2, 3, 3, 3, 4, 4, 
3, 4, 4, 3, 5, 4, 5, 3, 6, 4, 6, 3, 7, 4, 7, 3, 8, 3, 9, 4, 8, 4, 9, 5, 0, 5, 
1, 5, 2, 5, 3, 5, 4, 5, 5, 5, 6, 5, 7, 5, 8, 5, 9, 6, 0, 6, 1, 6, 2, 6, 3, 6, 
4, 6, 5, 6, 6, 6, 7, 6, 8, 6, 9, 7, 0, 7, 1, 7, 2, 7, 3, 7, 4, 7, 5, 7, 6, 7, 
7, 7, 8, 7, 9, 8, 0, 9, 0, 8, 1, 9, 1, 8, 2, 9, 2, 8, 3, 8, 4, 9, 3, 9, 4, 8, 
5, 9, 5, 8, 6, 9, 6, 8, 7, 9, 7, 8, 8, 8, 9, 9, 8, 9, 9]
    ```
    
    Seems we better just stick to the original workaround.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to