Jeffery Yuan <yuanyun...@gmail.com> wrote:
>  In our application, every data there is about 800mb raw data, we are going
> to store this data for 5 years, then it's about 1 or 2 TB data.

>  I am wondering whether solr can support this much data?

Yes it can.

Or rather: You could probably construct a scenario where it is not feasible, 
but you would have to be very creative.

>  Usually how much data we store per node, how many nodes we can have in
> solr cloud, what hardware configuration each node should be?

As Shawn states, it is very hard to give advice on hardware (and I applaud him 
from refraining from giving the usual "free RAM == index size"-advice). 
However, we love to guesstimate, but to do that you really need to provide more 
details.


2TB of index that has hundreds of concurrent users, thousands of updates per 
seconds and heavy aggregations (grouping, faceting, streaming...) is a task 
that takes experimentation and beefy hardware.

2TB of index that is rarely updated and accessed by a few people at a time, 
which are okay with multi-second response times, can be handled by a 
desktop-class machine with SSDs.


Tell us about query types, update rates, latency requirements, document types 
and concurrent users. Then we can begin to guess.

- Toke Eskildsen

Reply via email to