Hello We have an index Solr with 3 nodes, 1 shard et 2 replicas.
Our goal is to index 42 millions rows. Indexing time is important. The data source is an oracle database. Our indexing strategy is : * Reading from Oracle to a big CSV file. * Reading from 4 files (big file chunked) and injection via ConcurrentUpdateSolrClient Is it the optimal way of injecting such mass of data into Solr ? For information, estimated time for our solution is 6h.