Hi Kailash On 6/11/19 9:24 AM, Kailash Kota wrote: > I understand Oracle Golden Gate is a data replication tool which uses > log-based technology to stream all changes to a database from source, to > target. Can you please help me in understand what is the role of Kafka after > the data is provided to it by OGG ?
The tool in question here is not "standard Oracle GoldenGate", but rather "Oracle GoldenGate for BigData" https://docs.oracle.com/goldengate/bd123010/gg-bd/. OGG-BD captures change records from the Oracle log (or other database) and publishes those changes to Kafka (or other typical "BigData technologies" like HDFS, HBase, Flume). The alternatives that I'm aware of that do similar things are Attunity https://www.attunity.com/content/change-data-capture-cdc-oracle/ (commercial), StreamSets https://streamsets.com/blog/change-data-capture-from-oracle-with-streamsets-data-collector/ (partially OSS), Debezium https://debezium.io/ and Kafka Connect Oracle https://github.com/erdemcer/kafka-connect-oracle There may be more. > Also if we just need to data replicate our oracle DB, is there any other ways > we can do it without Oracle Golden Gate ? Is Apache Storm/Flink a prospect > for us to look into (though they are streaming tools). As Robin Moffatt's article notes, besides log-based CDC there is also query based CDC. The tool choce there is much wider, and you could certainly use Storm of Flink to implement a solution. Since most people I have seen doing this want a persistent store/buffer for the change records, instead of processing them directly without any buffering, most solutions I have seen push the records into Kafka, by using some Kafka Connect setup. If your requirements are for direct processing, I would look into Flink for this. -- CU, Joe
