First, it's totally impossible to answer in the abstract, see: https://lucidworks.com/2012/07/23/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/
Second, indexing DB tables directly into Solr is usually the wrong approach. Solr is not a replacement for a relational DB, it does not function as a DB, is not optimized for joins etc. It's a _search engine_ and does that superlatively. At the very least, the most common recommendation if you have the space is to de-normalize the data. My point is you need to think about this problem in terms of _search_, not "move some tables to Solr and use Solr like a DB". Which means that even if someone can answer your questions, it won't help much. Best, Erick On Thu, Jan 11, 2018 at 9:53 PM, Deepak Nair <deepak.n...@infosys.com> wrote: > Hello, > > We want to implement Solr 7.x for one of our client with below requirement. > > > 1. The data to be index will be from 2 Oracle databases with 2 tables > each and around 10 columns. > > 2. The data volume is expected to be reach around 10 million in each > table. > > 3. 4000+ users will query the indexed data from a UI. The peak load is > expected to be around 2000 queries/sec. > > 4. The implementation will be on a standalone or clustered Unix > environment. > > I want to know what should be the best server configuration for this kind of > requirement. Eg: how many VMs, what should be the RAM, Heap size etc. > > Thanks, > Deepak