Hi, all.  I'm just looking for quick guidance or confirmation that I'm going in 
the right direction here:

- There's a small Kotlin service that uses Camel to read from Kafka and write 
to Mongo.
- I need to replace Mongo with Apache Drill and write Parquet files to the file 
system.
  (I know nothing about Parquet but I know a little bit about Drill.)

- This service isn't used to do any queries, it's just for persisting data.
  So, given that, and the fact that Drill is just a query engine, I really 
can't use the "Drill" component for anything.

- But there is that "HDFS" component that I think I can use?
  Or maybe the "File" component is better here?

So my thinking is that I just need to:

1. write a Processor to transform the JSON data into Parquet
   (and keep in mind that I know nothing about Parquet...)

2. use the HDFS (or File) component to write it to a file
   (I think there's some Parquet set up to do (?) outside the scope of this 
service, but that's another matter...)

Seems pretty straight-forward.  Does that sound reasonable?

Are there any Camel examples I can look at?  The Google machine seems to not 
find anything related to Camel and Parquet...

Thank you so much!

Ron

Reply via email to