We rolled our own since we couldn't (1.5 years ago) find one. The code is quite simple and short.
On Tue, Dec 6, 2016 at 1:55 PM Aseem Bansal <asmbans...@gmail.com> wrote: > I just meant that is there an existing tool which does that. Basically I > tell it "Listen to all X streams and write them to S3/HDFS at Y path as > JSON". I know spark streaming can be used and there is flume. But I am not > sure about their scalability/reliability. That's why I thought to initiate > a discussion here to see whether someone knows about that already. > > On Tue, Dec 6, 2016 at 5:14 PM, Sharninder <sharnin...@gmail.com> wrote: > > > What do you mean by streaming way? The logic to push to S3 will be in > your > > consumer, so it totally depends on how you want to read and store. I > think > > that's an easier way to do what you want to, instead of trying to backup > > kafka and then read messages from there. Not even sure that's possible. > > > > On Tue, Dec 6, 2016 at 5:11 PM, Aseem Bansal <asmbans...@gmail.com> > wrote: > > > > > I get that we can read them and store them in batches but is there some > > > streaming way? > > > > > > On Tue, Dec 6, 2016 at 5:09 PM, Aseem Bansal <asmbans...@gmail.com> > > wrote: > > > > > > > Because we need to do exploratory data analysis and machine learning. > > We > > > > need to backup the messages somewhere so that the data scientists can > > > > query/load them. > > > > > > > > So we need something like a router that just opens up a new consumer > > > group > > > > which just keeps on storing them to S3. > > > > > > > > On Tue, Dec 6, 2016 at 5:05 PM, Sharninder Khera < > sharnin...@gmail.com > > > > > > > wrote: > > > > > > > >> Why not just have a parallel consumer read all messages from > whichever > > > >> topics you're interested in and store them wherever you want to? You > > > don't > > > >> need to "backup" Kafka messages. > > > >> > > > >> _____________________________ > > > >> From: Aseem Bansal <asmbans...@gmail.com> > > > >> Sent: Tuesday, December 6, 2016 4:55 PM > > > >> Subject: Storing Kafka Message JSON to deep storage like S3 > > > >> To: <users@kafka.apache.org> > > > >> > > > >> > > > >> Hi > > > >> > > > >> Has anyone done a storage of Kafka JSON messages to deep storage > like > > > S3. > > > >> We are looking to back up all of our raw Kafka JSON messages for > > > >> Exploration. S3, HDFS, MongoDB come to mind initially. > > > >> > > > >> I know that it can be stored in kafka itself but storing them in > Kafka > > > >> itself does not seem like a good option as we won't be able to query > > it > > > >> and > > > >> the configurations of machines containing kafka will have to be > > > increased > > > >> as we go. Something like S3 we won't have to manage. > > > >> > > > >> > > > >> > > > >> > > > >> > > > > > > > > > > > > > > > > > > > -- > > -- > > Sharninder > > >