This is an automated email from the ASF dual-hosted git repository. jasonhuynh pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/geode-kafka-connector.git
The following commit(s) were added to refs/heads/master by this push: new bf89bf7 Update README.md bf89bf7 is described below commit bf89bf78fc4e58b13da2623e415693c59cfad5ef Author: Jason Huynh <huyn...@gmail.com> AuthorDate: Thu Feb 6 12:12:57 2020 -0800 Update README.md --- README.md | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index aeb3d87..39dd423 100644 --- a/README.md +++ b/README.md @@ -13,10 +13,13 @@ The GeodeKafkaSink allows Geode to consume data off of topics and store data fro Installation of the connector is similar in process to other Kafka Connectors. For now, we will follow the guide for [Manual Installation](https://docs.confluent.io/current/connect/managing/install.html#install-connector-manually). In summary, we will use the standalone worker for this example. -* Explode a zip file or build into a known (and Kafka accessible) location +* Build the jar into a known (and Kafka accessible) location (in the geode-kafka-connector project run './gradlew shadowJar' and the artifacts will be in build/lib/ * Modify the connect-standalone.properties and point to where the connector is installed. ``` -plugin.path=/Users/jhuynh/Pivotal/geode-kafka-connector/build/libs/ +plugin.path=(Path to your clone)/geode-kafka-connector/build/libs/ +#depending on object type, you might want to modify the converter (for manually testing we can use the JSON or string converter) +#key.converter= +#value.converter= ``` * Create and modify connect-geode-sink.properties file, for example ``` @@ -24,6 +27,7 @@ name=geode-kafka-sink connector.class=GeodeKafkaSink tasks.max=1 topicToRegions=[someTopicToSinkFrom:someRegionToConsume] +topics=someTopicToSinkFrom locators=localHost[10334] ``` * Create and modify connect-geode-source.properties files