Yes, though it's still awaiting some updates after some renaming and API modifications that happened in Kafka recently.
-Ewen On Thu, Nov 12, 2015 at 9:10 AM, Venkatesh Rudraraju < venkatengineer...@gmail.com> wrote: > Ewen, > > How do I use a HDFSSinkConnector. I see the sink as part of a confluent > project ( > > https://github.com/confluentinc/copycat-hdfs/blob/master/src/main/java/io/confluent/copycat/hdfs/HdfsSinkConnector.java > ). > Does it mean that I build this project and add the jar to kafka libs ? > > > > > On Tue, Nov 10, 2015 at 9:35 PM, Ewen Cheslack-Postava <e...@confluent.io> > wrote: > > > Venkatesh, > > > > 1. It only works with quotes because the message needs to be parsed as > JSON > > -- a bare string without quotes is not valid JSON. If you're just using a > > file sink, you can also try the StringConverter, which only supports > > strings and uses a fixed schema, but is also very easy to use since it > has > > minimal requirements. It's really meant for demonstration purposes more > > than anything else, but may be helpful just to get up and running. > > 2. Which JsonParser error? When processing a message fails, we need to be > > careful about how we handle it. Currently it will not proceed if it can't > > process a message since for a lot of applications it isn't acceptable to > > drop messages. By default, we want at least once semantics, with exactly > > once as long as we don't encounter any crashes or network errors. Manual > > intervention is currently required in that case. > > > > -Ewen > > > > On Tue, Nov 10, 2015 at 8:58 PM, Venkatesh Rudraraju < > > venkatengineer...@gmail.com> wrote: > > > > > Hi Ewen, > > > > > > Thanks for the explanation. with your suggested setting, I was able to > > > start just a sink connector like below : > > > > > > >* bin/connect-standalone.sh config/connect-standalone.properties > > > config/connect-file-sink.properties* > > > > > > But I have a couple of issues yet, > > > 1) Since I am only testing a simple file sink connector, I am manually > > > producing some messages to the 'connect-test' kafka topic, where the > > > sink-Task is reading from. And it works only if the message is within > > > double-quotes. > > > 2) Once I hit the above JsonParser error on the SinkTask, the connector > > is > > > hung, doesn't take any more messages even proper ones. > > > > > > > > > On Tue, Nov 10, 2015 at 1:59 PM, Ewen Cheslack-Postava < > > e...@confluent.io> > > > wrote: > > > > > > > Hi Venkatesh, > > > > > > > > If you're using the default settings included in the sample configs, > > > it'll > > > > expect JSON data in a special format to support passing schemas along > > > with > > > > the data. This is turned on by default because it makes it possible > to > > > work > > > > with a *lot* more connectors and data storage systems (many require > > > > schemas!), though it does mean consuming regular JSON data won't work > > out > > > > of the box. You can easily switch this off by changing these lines in > > the > > > > worker config: > > > > > > > > key.converter.schemas.enable=true > > > > value.converter.schemas.enable=true > > > > > > > > to be false instead. However, note that this will only work with > > > connectors > > > > that can work with "schemaless" data. This wouldn't work for, e.g., > > > writing > > > > Avro files in HDFS since they need schema information, but it might > > work > > > > for other formats. This would allow you to consume JSON data from any > > > topic > > > > it already existed in. > > > > > > > > Note that JSON is not the only format you can use. You can also > > > substitute > > > > other implementations of the Converter interface. Confluent has > > > implemented > > > > an Avro version that works well with our schema registry ( > > > > > > > > > > https://github.com/confluentinc/schema-registry/tree/master/avro-converter > > > > ). > > > > The JSON implementation made sense to add as the one included with > > Kafka > > > > simply because it didn't introduce any other dependencies that > weren't > > > > already in Kafka. It's also possible to write implementations for > other > > > > formats (e.g. Thrift, Protocol Buffers, Cap'n Proto, MessagePack, and > > > > more), but I'm not aware of anyone who has started to tackle those > > > > converters yet. > > > > > > > > -Ewen > > > > > > > > On Tue, Nov 10, 2015 at 1:23 PM, Venkatesh Rudraraju < > > > > venkatengineer...@gmail.com> wrote: > > > > > > > > > Hi, > > > > > > > > > > I am trying out the new kakfa connect service. > > > > > > > > > > version : kafka_2.11-0.9.0.0 > > > > > mode : standalone > > > > > > > > > > I have a conceptual question on the service. > > > > > > > > > > Can I just start a sink connector which reads from Kafka and writes > > to > > > > say > > > > > HDFS ? > > > > > From what I have tried, it's expecting a source-connector as well > > > because > > > > > the sink-connector is expecting a particular pattern of the message > > in > > > > > kafka-topic. > > > > > > > > > > Thanks, > > > > > Venkat > > > > > > > > > > > > > > > > > > > > > -- > > > > Thanks, > > > > Ewen > > > > > > > > > > > > > > > > -- > > > Victory awaits him who has everything in order--luck, people call it. > > > > > > > > > > > -- > > Thanks, > > Ewen > > > > > > -- > Victory awaits him who has everything in order--luck, people call it. > -- Thanks, Ewen