Abacn opened a new issue, #37893:
URL: https://github.com/apache/beam/issues/37893

   ### What needs to happen?
   
   Currently a Python Row (encoded with Row Coder) go through 
serialization/deserialization becomes a schema generated types named tuple. 
There are many caveats for this behavior
   
   - original type get lost
   
   - #22714 
   
   with cloudpickle becomes default and schema registry coder registry saved on 
pipeline submission, we should be able to use the schema id registered in the 
schema registry to obtain the user type, then use coder registry for the user 
type to get registered (row) coder, that makes user_type->GBK still produces 
user_type
   
   ### Issue Priority
   
   Priority: 2 (default / most normal work should be filed as P2)
   
   ### Issue Components
   
   - [ ] Component: Python SDK
   - [ ] Component: Java SDK
   - [ ] Component: Go SDK
   - [ ] Component: Typescript SDK
   - [ ] Component: IO connector
   - [ ] Component: Beam YAML
   - [ ] Component: Beam examples
   - [ ] Component: Beam playground
   - [ ] Component: Beam katas
   - [ ] Component: Website
   - [ ] Component: Infrastructure
   - [ ] Component: Spark Runner
   - [ ] Component: Flink Runner
   - [ ] Component: Samza Runner
   - [ ] Component: Twister2 Runner
   - [ ] Component: Hazelcast Jet Runner
   - [ ] Component: Google Cloud Dataflow Runner


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to