Github user HyukjinKwon commented on the issue: https://github.com/apache/spark/pull/19498 I just took a look. This looks actually going to bigger change because we will need manually deal with the different de/serializers cases in Python side as well, for example: ``` >>> rdd = sc.parallelize(range(10)).cartesian(sc.parallelize(range(10))) >>> rdd._jrdd_deserializer = sc.serializer >>> rdd.collect() [0, 0, 0, 1, 0, 2, 0, 3, 0, 4, 0, 5, 0, 6, 0, 7, 0, 8, 0, 9, 1, 0, 1, 1, 1, 2, 1, 3, 1, 4, 1, 5, 1, 6, 1, 7, 1, 8, 1, 9, 2, 0, 2, 1, 2, 2, 2, 3, 2, 4, 2, 5, 2, 6, 2, 7, 2, 8, 2, 9, 3, 0, 4, 0, 3, 1, 4, 1, 3, 2, 4, 2, 3, 3, 3, 4, 4, 3, 4, 4, 3, 5, 4, 5, 3, 6, 4, 6, 3, 7, 4, 7, 3, 8, 3, 9, 4, 8, 4, 9, 5, 0, 5, 1, 5, 2, 5, 3, 5, 4, 5, 5, 5, 6, 5, 7, 5, 8, 5, 9, 6, 0, 6, 1, 6, 2, 6, 3, 6, 4, 6, 5, 6, 6, 6, 7, 6, 8, 6, 9, 7, 0, 7, 1, 7, 2, 7, 3, 7, 4, 7, 5, 7, 6, 7, 7, 7, 8, 7, 9, 8, 0, 9, 0, 8, 1, 9, 1, 8, 2, 9, 2, 8, 3, 8, 4, 9, 3, 9, 4, 8, 5, 9, 5, 8, 6, 9, 6, 8, 7, 9, 7, 8, 8, 8, 9, 9, 8, 9, 9] ``` Seems we better just stick to the original workaround.
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org