Github user davies commented on the pull request:
https://github.com/apache/spark/pull/12620#issuecomment-214439666
Can we support dill directly and have a flag to choose from the two
serializer? cloud-pickler could be the default one.
---
If your project is set up for it, you can
Github user njwhite commented on the pull request:
https://github.com/apache/spark/pull/12620#issuecomment-214210492
@davies I'm using this to use the "dill" serializer, as it can pickle more
things (and allows more fine-grained control) than the cloud-pickle serializer.
What about
Github user davies commented on the pull request:
https://github.com/apache/spark/pull/12620#issuecomment-214147661
@njwhite We still use PickleSerializer to deserialize the functions, so it
means the serializer MUST be compatible with Pickle, I'm not sure make it
configurable will
Github user holdenk commented on the pull request:
https://github.com/apache/spark/pull/12620#issuecomment-214104132
If we do end up adding this we would probably want to add a test of using a
custom serializer (but maybe don't rush to do this since I think if we want to
expose this
Github user holdenk commented on the pull request:
https://github.com/apache/spark/pull/12620#issuecomment-214103364
Is this functionality we want to add? cc @davies ?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If
Github user AmplabJenkins commented on the pull request:
https://github.com/apache/spark/pull/12620#issuecomment-213586086
Can one of the admins verify this patch?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your
GitHub user njwhite opened a pull request:
https://github.com/apache/spark/pull/12620
[SPARK-14859][PYSPARK] Make Lambda Serializer Configurable
## What changes were proposed in this pull request?
Store the serializer that we should use to serialize RDD transformation