Yes. Lucene / Solr supports multi threaded environment. You could do commit
from two different threads to same core or different core.
Regards
Aditya
www.findbestopensource.com
On Tue, May 22, 2012 at 12:35 AM, jame vaalet wrote:
> hi,
> my use case here is to search all the incoming documents
www.findbestopensource.com
On Tue, May 22, 2012 at 2:36 PM, Bruno Mannina wrote:
> My choice:
> http://www.ovh.com/fr/**serveurs_dedies/eg_best_of.xml<http://www.ovh.com/fr/serveurs_dedies/eg_best_of.xml>
>
> 24 Go DDR3
>
> Le 22/05/2012 10:26, findbestopensource a écrit :
>
> Dedica
Having cores per user is not good idea. The count is too high. Keep
everything in single core. You could filter the data based on user name or
user id.
Regards
Aditya
www.findbestopensource.com
On Tue, May 22, 2012 at 2:29 PM, Shanu Jha wrote:
> Hi all,
>
> greetings from my end. This is my f
Thats how de-normalization works. You need to update all child products.
If you just need the count and you are using facets then maintain a map
between category and main product, main product and child product. Lucene
db has no schema. You could retrieve the data based on its type.
Category reco
Dedicated Server may not be required. If you want to cut down cost, then
prefer shared server.
How much the RAM?
Regards
Aditya
www.findbestopensource.com
On Tue, May 22, 2012 at 12:36 PM, Bruno Mannina wrote:
> Dear Solr users,
>
> My company would like to use solr to index around 80 000 000
Hi Parvin,
Fault tolerant architecture is something you need to decide on your
requirement. At some point of time there may require some manual
intervention to recover from crash. You need to see how much percentage you
could support fault tolerant. It certainly may not be 100. We could handle
sit
You could very well use Solr. It has support to index the PDF and XML
files. If you want to index websites and search using page rank then choose
Nutch.
Regards
Aditya
www.findbestopensource.com
On Wed, May 16, 2012 at 1:13 PM, Tolga wrote:
> Hi,
>
> I have been trying for a week. I really wan
I have written an article on this. The various steps to restrict /
authenticate Solr admin interface.
http://www.findbestopensource.com/article-detail/restrict-solr-admin-access
Regards
Aditya
www.findbestopensource.com
On Thu, Mar 29, 2012 at 1:06 AM, geeky2 wrote:
> update -
>
> ok - i was
While indexing @ is removed. You need to use your own Tokenizer which will
consider "@rohit" as one word.
Another option is to break the tweet in to two fields, @ and the
tweet. Index both the fields but don't use any tokenizer for the field "
@". Just index as it is. While querying you need to se
Hello all,
Recently i saw couple of discussions in LinkedIn group about generating
large data set or data corpus. I have compiled the same in to an article.
Hope it would be helpful. If you have any other links where we could get
large data set for free, please reply to this mail thread, i will up
Hello Anurag
Google is always there to do internet search. You need to support search for
your university. My opinion would be don't crawl the sites. You require only
Solr and not Nutch.
1. Provide an interface to upload the documents by the university students.
The documents could be previous ye
1. Reports based on Location. Group by City / Country
2. Total search performed per hour / week / month
3. Frequently used search keywords
4. Analytics based on search keywords.
Regards
Aditya
www.findbestopensource.com
On Fri, May 6, 2011 at 3:55 AM, Otis Gospodnetic wrote:
> Hi,
>
> I'd like
Hello deniz,
You could create a new field say FullName which is a of
firstname and surname. Search on both the new field and location but boost
up the new field query.
Regards
Aditya
www.findbestopensource.com
On Thu, May 5, 2011 at 9:21 AM, deniz wrote:
> okay... let me make the situation
Hello Dominique Bejean,
Good job.
We identified almost 8 open source web crawlers
http://www.findbestopensource.com/tagged/webcrawler I don't know how far
yours would be different from the rest.
Your license states that it is not open source but it is free for personnel
use.
Regards
Aditya
ww
You may need to use Hebrew analyzer.
http://www.findbestopensource.com/search/?query=hebrew
Regards
Aditya
www.findbestopensource.com
On Tue, Jan 18, 2011 at 2:34 PM, prasad deshpande <
prasad.deshpand...@gmail.com> wrote:
> Hello,
>
> With reference to below links I haven't found Hebrew suppo
Some more pointers to spatial search,
http://www.jteam.nl/products/spatialsolrplugin.html
http://code.google.com/p/spatial-search-lucene/
http://sujitpal.blogspot.com/2008/02/spatial-search-with-lucene.html
Regards
Aditya
www.findbestopensource.com
On Thu, Jul 15, 2010 at 3:54 PM, Saïd Radhoua
s).
>
> 2010/7/14 findbestopensource :
> > I have just provided you two options. Since you already store as part of
> the
> > index, You could try external caching. Try using ehcache / Membase
> > http://www.findbestopensource.com/tagged/distributed-caching . The
>
ad more into
> memory, I want to compress it "in memory". I don't care much about
> disk space so whether or not it's compressed in lucene .
>
> 2010/7/14 findbestopensource :
> > You have two options
> > 1. Store the compressed text as part of store
You have two options
1. Store the compressed text as part of stored field in Solr.
2. Using external caching.
http://www.findbestopensource.com/tagged/distributed-caching
You could use ehcache / Memcache / Membase.
The problem with external caching is you need to synchronize the deletions
and
Refer http://wiki.apache.org/solr/Solrj#EmbeddedSolrServer
Regards
Aditya
www.findbestopensource.com
On Fri, Jun 11, 2010 at 2:25 PM, Robert Naczinski <
robert.naczin...@googlemail.com> wrote:
> Hello experts,
>
> we would like to use Solr in our search application. We want to index
> a large i
Could you tell us your schema used for indexing. In my opinion, using
standardanalyzer / Snowball analyzer will do the best. They will not break
the URLs. Add href, and other related html tags as part of stop words and it
will removed while indexing.
Regards
Aditya
www.findbestopensource.com
On
You need to do schedule your task. Check out schedulers available in all
programming languages.
http://www.findbestopensource.com/tagged/job-scheduler
Regards
Aditya
www.findbestopensource.com
On Wed, Jun 2, 2010 at 2:39 PM, Jonty Rhods wrote:
> Hi Peter,
>
> actually I want the index process
What analyzer you are using to index and search? Check out schema.xml. You
are currently using analyzer which breaks the words. If you don't want to
break then you need to use .
Regards
Aditya
www.findbestopensource.com
On Wed, Jun 2, 2010 at 2:41 PM, M.Rizwan wrote:
> Hi,
>
> I have solr 1.4.
Add commit after the loop. I would advise to use commit in a separate
thread. I do keep separate timer thread, where every minute I will do
commit and at the end of every day I will optimize the index.
Regards
Aditya
www.findbestopensource.com
On Tue, Jun 1, 2010 at 2:57 AM, Steve Kuo wrote:
>
go about that?
>
> Regards,
> Raakhi
>
>
> On Tue, May 25, 2010 at 5:07 PM, findbestopensource <
> findbestopensou...@gmail.com> wrote:
>
> > Resending it as there is a typo error.
> >
> > To reterive all documents, You need to use the query/filter
Resending it as there is a typo error.
To reterive all documents, You need to use the query/filter FieldName:*:* .
Regards
Aditya
www.findbestopensource.com
On Tue, May 25, 2010 at 4:29 PM, findbestopensource <
findbestopensou...@gmail.com> wrote:
> To reterive all documents, You ne
To reterive all documents, You need to use the query/filter *FieldName:*:**
Regards
Aditya
www.findbestopensource.com
On Tue, May 25, 2010 at 4:14 PM, Rakhi Khatwani wrote:
> Hi,
> Is there any way to get all the fields (irrespective of whether
> it contains a value or null) in solrDocu
To reterive all documents, You need to use the query/filter *FieldName:*:**
Regards
Aditya
www.findbestopensource.com
On Tue, May 25, 2010 at 4:14 PM, Rakhi Khatwani wrote:
> Hi,
> Is there any way to get all the fields (irrespective of whether
> it contains a value or null) in solrDo
Hi Rih,
You going to include either of the two field "bought" or "like" to per
member/visitor OR a unique field per member / visitor?
If it's one or two common fields are included then there will not be any
impact in performance. If you want to include unique field then you need to
consider multi
Hi Peter,
You need to use Lucene,
- To have more control
- You cannot depend on any Web server
- To use termvector, termdocs etc
- You could easily extend to have your own Analyzer
You need to use Solr,
- To index and search docs easily by writting few code
- Solr is a standal
ot;? No queries on the masters.
> Only one index is being processed/optimized.
>
> Also, if I may add to my same question, how can I find the
> amount of memory that an index would use, theoretically?
> i.e.: Is there a formulae etc?
>
> Thanks
> Madu
>
>
>
You may use one index at a time, but both indexes are active and loaded all
its terms in memory. Memory consumption will be certainly more.
Regards
Aditya
http://www.findbestopensource.com
On Fri, May 14, 2010 at 10:28 AM, Maduranga Kannangara <
mkannang...@infomedia.com.au> wrote:
> Hi
>
> We u
Hello Eric,
Certainly it is possible. I would strongly advice to have field which
differentiates the record type (RECORD_TYPE:"CAR" / "PROPERTY").
>>In general I was also wondering how Solr developers implement websites
that
uses tag filters.For example, a user clicks on "Hard drives" then get ta
I downloaded the 5883_Code.zip file but not able to extract the complete
contents.
Regards
Aditya
www.findbestopensource.com
On Tue, Apr 27, 2010 at 12:45 AM, Johan Cwiklinski <
johan.cwiklin...@ajlsm.com> wrote:
> Hello,
>
> Le 26/04/2010 20:53, findbestopensource a écrit :
&
I am able to successfully download the code. It is of 360 MB and took lot of
time to download.
https://www.packtpub.com/solr-1-4-enterprise-search-server/book
Select the "download the code" link and provide your email id, Download link
will be sent via email.
Regards
Aditya
www.findbestopensource.
Hello Dennis
>>If the load goes up, then queries are sent to the cloud at a certain
point.
My advice is to do load balancing between local and cloud. Your local
system seems to be capable as it is a dedicated host. Another option is to
do indexing in local and sync it with cloud. Cloud will be on
Thank you Dave and Michael for your feedback.
We are currently in beta and we will fix these issues sooner.
Regards
Aditya
www.findbestopensource.com
On Tue, Apr 20, 2010 at 3:01 PM, Michael Kuhlmann <
michael.kuhlm...@zalando.de> wrote:
> Nice site. Really!
>
> In addition to Dave:
> How do
37 matches
Mail list logo