[
https://issues.apache.org/jira/browse/SOLR-10676?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Jan Høydahl closed SOLR-10676.
------------------------------
Resolution: Invalid
Please ask your questions to the Sunspot community. This JIRA is a bug tracker
for bugs or new features in the core Solr product.
If you still need to ask a solr-core related question, please post that to the
[email protected] mailing list, and not to this bug tracker. See
http://lucene.apache.org/solr/community.html#mailing-lists-irc
PS: I would imagine that the Sunspot code could be optimized to not send
documents one by one but in batches and perhaps avoid explicit commits etc??
Closing this issue as invalid
> Optimize the reindexing of sunspot solr
> ---------------------------------------
>
> Key: SOLR-10676
> URL: https://issues.apache.org/jira/browse/SOLR-10676
> Project: Solr
> Issue Type: Improvement
> Security Level: Public(Default Security Level. Issues are Public)
> Components: clients - ruby - flare
> Affects Versions: 5.0
> Reporter: Krishna Sahoo
>
> We are using solr 5.0. <luceneMatchVersion>5.0.0</luceneMatchVersion>
> We have more than 5 million products. It is taking around 3.30 hours to
> reindex all the products.
> For optimizing the reindexing speed, we have used the following configurations
> <indexConfig>
> <ramBufferSizeMB>960</ramBufferSizeMB>
> <mergePolicyFactory>100</mergePolicyFactory>
> <mergeScheduler
> class="org.apache.lucene.index.ConcurrentMergeScheduler"/>+
> </indexConfig>
> <autoCommit>
> <maxTime>${solr.autoCommit.maxTime:15000}</maxTime>
> <openSearcher>false</openSearcher>
> </autoCommit>
> <autoSoftCommit>
> <maxTime>${solr.autoSoftCommit.maxTime:-1}</maxTime>
> </autoSoftCommit>
> We are indexing with the following option
> { :batch_commit => false,:batch_size=>20000 }
> We have set autocommit false in our model. So whenever a new record is
> inserted it is not automatically added to the solr index. But when a record
> is updated we manually call the Sunspot.index! method for that particular
> product data. Everyday we are inserting around .2 millions of records. We
> have target of 50 million products.
> Is there any way, that we can add to index only the new records or updated
> records?
> Can we increase the indexing speed by changing any of the current
> configurations?
> If we add new products to the solr through ruby code by using loop, it fails
> miserably as it takes too much time.
> Please help to find the best way to improve the indexing speed of solr.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]