[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15224155#comment-15224155 ] ASF GitHub Bot commented on FLINK-3524: --- Github user asfgit closed the pull request at: https://github.com/apache/flink/pull/1834 > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > Fix For: 1.1.0 > > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15224149#comment-15224149 ] ASF GitHub Bot commented on FLINK-3524: --- Github user rmetzger commented on the pull request: https://github.com/apache/flink/pull/1834#issuecomment-205303136 Merging ... > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15217695#comment-15217695 ] ASF GitHub Bot commented on FLINK-3524: --- Github user zentol commented on the pull request: https://github.com/apache/flink/pull/1834#issuecomment-203335575 @StephanEwen yes, it's only for kafka. it relies on other classes (KeyedDeserializationSchema) that are only present in the kafka module. > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15216193#comment-15216193 ] ASF GitHub Bot commented on FLINK-3524: --- Github user StephanEwen commented on the pull request: https://github.com/apache/flink/pull/1834#issuecomment-202964440 Looks good. Is this a Kafka-only util? > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15210093#comment-15210093 ] ASF GitHub Bot commented on FLINK-3524: --- Github user rmetzger commented on the pull request: https://github.com/apache/flink/pull/1834#issuecomment-200795303 +1 to merge > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15208352#comment-15208352 ] ASF GitHub Bot commented on FLINK-3524: --- Github user rmetzger commented on the pull request: https://github.com/apache/flink/pull/1834#issuecomment-200330990 Otherwise, the change looks good to merge. > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15208351#comment-15208351 ] ASF GitHub Bot commented on FLINK-3524: --- Github user rmetzger commented on the pull request: https://github.com/apache/flink/pull/1834#issuecomment-200330964 Can you quickly mention the class in the Kafka documentation? > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (FLINK-3524) Provide a JSONDeserialisationSchema in the kafka connector package
[ https://issues.apache.org/jira/browse/FLINK-3524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15208337#comment-15208337 ] ASF GitHub Bot commented on FLINK-3524: --- GitHub user zentol opened a pull request: https://github.com/apache/flink/pull/1834 [FLINK-3524] [kafka] Add JSONDeserializationSchema This PR adds a JSON[KeyValue]DeserializationSchema to the kafka connector, which can be used to more easily read JSON data from Kafka. Instead of having to read them as strings and creating parsers themselves a user can now delegeta this to the DeserializationSchema, saving the string allocation and possible instantiation of multiple parser factories. You can merge this pull request into a Git repository by running: $ git pull https://github.com/zentol/flink 3524_json_schema Alternatively you can review and apply these changes as the patch at: https://github.com/apache/flink/pull/1834.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1834 commit 536b2888f1f8fa0a76753e0245378ceb0d242ef1 Author: zentolDate: 2016-03-23T10:24:50Z [FLINK-3524] [kafka] Add JSONDeserializationSchema > Provide a JSONDeserialisationSchema in the kafka connector package > -- > > Key: FLINK-3524 > URL: https://issues.apache.org/jira/browse/FLINK-3524 > Project: Flink > Issue Type: Improvement > Components: Kafka Connector >Reporter: Robert Metzger >Assignee: Chesnay Schepler > Labels: starter > > (I don't want to include this into 1.0.0) > Currently, there is no standardized way of parsing JSON data from a Kafka > stream. I see a lot of users using JSON in their topics. It would make things > easier for our users to provide a serializer for them. > I suggest to use the jackson library because we have that aready as a > dependency in Flink and it allows to parse from a byte[]. > I would suggest to provide the following classes: > - JSONDeserializationSchema() > - JSONDeKeyValueSerializationSchema(bool includeMetadata) > The second variant should produce a record like this: > {code} > {"key": "keydata", > "value": "valuedata", > "metadata": {"offset": 123, "topic": "", "partition": 2 } > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)