Hi Luis,
First to ans you questions,
-Do you index at the same time as users launches queries to twitter index?
Yes, currently I am using the same server for indexing and querying due to
lack of resources to extent to another server.
-Are you using the autocommit option?
No, I am committing on
Thanks Chris,
This makes sense, at any time we show users a trend graph for all the tweets
relevant for them in the last 15 days. So I guess keeping a shards for the
last 15-20 days data would be a good option and all the other data moved to
different shards each with 2 months data.
I have no ide
: how do i configure the solr folder to specific directory when deploy to
: servlet container.
https://wiki.apache.org/solr/SolrInstall
-Hoss
On Fri, Sep 16, 2011 at 6:53 PM, Burton-West, Tom wrote:
> Hello,
>
> The TieredMergePolicy has become the default with Solr 3.3, but the
> configuration in the example uses the mergeFactor setting which applys to the
> LogByteSizeMergePolicy.
>
> How is the mergeFactor interpreted by the Tiered
Hello,
The TieredMergePolicy has become the default with Solr 3.3, but the
configuration in the example uses the mergeFactor setting which applys to the
LogByteSizeMergePolicy.
How is the mergeFactor interpreted by the TieredMergePolicy?
Is there an example somewhere showing how to configure t
Hello,
The TieredMergePolicy has become the default with Solr 3.3, but the
configuration in the example uses the mergeFactor setting which applys to the
LogByteSizeMergePolicy.
How is the mergeFactor interpreted by the TieredMergePolicy?
Is there an example somewhere showing how to configure th
Hello,
The TieredMergePolicy has become the default with Solr 3.3, but the
configuration in the example uses the mergeFactor setting which applys to the
LogByteSizeMergePolicy.
How is the mergeFactor interpreted by the TieredMergePolicy?
Is there an example somewhere showing how to configure th
Hello.
Facet queries are slower than others specially when you are working with a
69G index. I would like to know more about the context where occurs the Out
of memory exception: is it during an indexation? Do you index at the same
time as users launches queries to twitter index? Are you using the
: Has anyone ever had to create large mock/dummy datasets for test
: environments or for POCs/Demos to convince folks that Solr was the
: wave of the future? Any tips would be greatly appreciated. I suppose
: it sounds a lot like crawling even though it started out as innocent
: DIH usage.
the be
On 9/16/2011 4:58 PM, Brandon Fish wrote:
> Hi Yury,
>
> You could try checking out the details command of the replication handler:
> http://slave_host:port/solr/replication?command=details
> which has information such as "isReplicating".
How reliable is isReplicating? Is it updated on unexpected
Hi all,
I'm checking how to do soft commits with the new version of Solr. I'm using
EmbeddedSolrServer to add documents to my index. How can I perform a soft
commit using this class? Is it possible? Or should I use the trunk?
http://wiki.apache.org/solr/NearRealtimeSearch
http://lucene.apache.org
I just started to test solr in Chinese content. In the admin page, I could use
*:* and saw the correct result. However, when I tried some specific search, I
came to no results.
I don't know what the problem is. Any thought?
By the way, I have a few cores running in English without encoding
Adding missing link to the issue I mentioned:
https://issues.apache.org/jira/browse/SOLR-1855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12851462#action_12851462
2011/9/16 Yury Kats
> Let's say I'm forcing a replication of a core using "fetchindex" com
Hi Yury,
You could try checking out the details command of the replication handler:
http://slave_host:port/solr/replication?command=details
which has information such as "isReplicating".
You could also look at the script attached to this issue which shows a
thorough check of a slaves replication
Let's say I'm forcing a replication of a core using "fetchindex" command.
No new content is being added to the master.
I can check whether replication has finished by periodically querying
master and slave for their "indexversion" and comparing the two.
But what's the best way to check if replica
I have solr issues where I keep running out of memory. I am working on
solving the memory issues (this will take a long time), but in the meantime,
I'm trying to be notified when the error occurs. I saw with the jvm I can
pass the -XX:OnOutOfMemoryError= flag and pass a script to run. Every time
t
: Actually I am storing twitter streaming data into the core, so the rate of
: index is about 12tweets(docs)/second. The same solr contains 3 other cores
...
: . At any given time I dont need data more than past 15 days, unless
: someone queries for it explicetly. How can this be a
: What is the best way to get a float field value from docID?
: I tried the following code but when it runs throws an exception For input
: string: "`??eI" at line float lat = Float.parseFloat(tlat);
the most most straight foward level of abstraction to use (if you want to
leverage the field ca
Hello.
You can also develop an application by yourself that uses Solrj to retrieve all
the documents from your índex, process and add all the new information (fields)
desired and the index them into another Solr index. Its easy.
Goodbye!
El 16/09/2011, a las 17:39, "Olson, Ron" escribió:
: I am commiting but not doing replication now. Mine sort order also includes
: last login timestamp. the new profiles are being reflected in my SOLR admin
: & db. but its not listed on my website.
please be *specific*
what do you mean when you say "the new profiles are being reflected in my
SO
: I'm wondering if it would be possible for you (or someone else) to fix the
: issue with years from 1 to 999 being formatted incorrectly, and then creating
: a new ticket for the issue with negative years?
you're totally right ... i got so frustrated trying to find a "perfect"
solution to the e
BTW - ignore the xpath typo, it should read:
On Fri, Sep 16, 2011 at 12:59 PM, B B wrote:
> Has anyone successfully setup DIH URLDataSource to paginate imports
> using $nextUrl & $hasMore ?:
>
> http://wiki.apache.org/solr/DataImportHandler#Configuration_in_data-config.xml-1
>
> It's not
Has anyone successfully setup DIH URLDataSource to paginate imports
using $nextUrl & $hasMore ?:
http://wiki.apache.org/solr/DataImportHandler#Configuration_in_data-config.xml-1
It's not working for me, the data imports on the first page, but no
subsequent calls to the data source are processed.
: It is pretty obvious from this that the "sdouble" schema fieldtype is
: not setup to parse out group-separators from a number.
correct. the numeric (and date) field types are all designed to deal with
conversion of the canonical string represetantion.
: 1) Then my question is which type pf s
Hello Folks,
Surprisingly, the value from the following raw data gives me a NFE
(Number Format Exception) when running the DIH (Data Import Handler):
$1,000.00
The error logs look like:
Caused by: org.apache.solr.common.SolrException: Error while creating
field 'price{type=sdouble,properties=inde
: I see that in this use case this makes most sense - thanks.
:
: But why are fieldNorms in general very corse?
they are a floating point value encoded into a single byte...
https://lucene.apache.org/java/3_4_0/scoring.html#Score%20Boosting
https://lucene.apache.org/java/3_4_0/api/core/org/apac
On Sep 16, 2011, at 8:39 AM, Olson, Ron wrote:
> Is there any way to re-parse an existing index, adding the new copyTo field,
> without having to basically start all over again with DIH?
No.
Indexes cannot be modified. You must reinsert the documents with a new schema.
wunder
--
Walter Underw
Hi all-
I have an 11 gig index that I realize I need to add another field to, but not
from the actual query using DIH, but via copyTo.
Is there any way to re-parse an existing index, adding the new copyTo field,
without having to basically start all over again with DIH?
Thanks,
Ron
DISCLAIME
probably stupid question,
Which Directory implementation should be the best suited for index
mounted on ramfs/tmpfs? I guess plain old FSDirectory, (or mmap/nio?)
probably stupid question,
Which Directory implementation should be the best suited for index
mounted on ramfs/tmpfs? I guess plain old FSDirectory, (or mmap/nio?)
Hello,
Iif I have a request with
fq=NAME_ANALYZED:tour eiffel
and I have different answers like
Restaurant la tour Eiffel
Hotel la tour Eiffel
Tour Eiffel
...
Is there a way to get answers with NAME_ANALYZED beginning with "tour
Eiffel" first?
Thanks,
Elisabeth
thanks iroxxx
but how can l add quartz like scheduling to solr dih ,is there any changes
required in anyof the configuration files please specify the procedure.
--
View this message in context:
http://lucene.472066.n3.nabble.com/add-quartz-like-scheduling-cabalities-to-solr-DIH-tp3341141p334179
Cores are essentially independent indexes. If you look at your
directory structure you'll see something like
/cores/core0/data/index
/cores/core1/data/index
Sharing the same index across multiple cores doesn't make
sense from a Solr perspective. And the redundancy comment
is kind of odd too. If yo
Let's quickly review the difference between indexed and stored.
Indexing a field means you can search on it.
Stored means an entirely separate copy of the raw input is placed
in a particular file in your index (*.fdt to be exact).
These operations are entirely orthogonal. In effect when you
sear
Hi Scott, now your queries are going to be created by a QueryParser. you
have a couple of options here, most common are LuceneQueryParser,
DismaxQueryParser and ExtendedDismaxQueryParser, but there are others. The
QueryParser will be creating all those queries you mentiones, for example,
if you are
What version of Solr are you using? Do you have any custom
code? I know some JIRAs have mentioned, you could try
searching the Solr and Lucene JIRAs at:
https://issues.apache.org/jira/browse/SOLR
But without a lot more background, it's hard to say. You might
review:
http://wiki.apache.org/solr/Us
Uhm, you're putting a lot of index into not very much memory. I
really think you're
going to have to shard your index across several machines to get past this
problem. Simply increasing the size of your caches is still limited by the
physical memory you're working with.
You really have to put a p
At a guess because the documents in question are very similar.
Best
Erick
2011/9/14 Daniel Pötzinger :
>>
>> fieldNorms are very corse. In my opinion, if you have a
>> "weighting" you want to use to affect score sort, it's better to index
>> that weight as a numeric field, and explicitly factor
Dear all
I am trying to index a data using json object. For normal single field
it is easy to make things work for me. I have a data to be stored in format
like
Solr Version 3.4,
java implementation
Document 1:
{ "id" : "B34KPNJ0998",
"project_name_t":"Manual",
> i am unable to do the scheduling for solr DIH,and anyone
> specifies the
> procedure for how to add quartz like scheduling cabalities
> to solr -DIH
> .Please let it me know is it possible to add quartz-like
> scheduling to solr
> DIH.
I think most of the people use Operating Stytems' cron-job c
I could get this number by using
group.ngroups=true&group.limit=0
but doing grouping for this seems like an overkill
Would you advise using JIRA SOLR-1814 ?
- Original Message -
From: swiss knife
Sent: 09/15/11 12:43 PM
To: solr-user@lucene.apache.org
Subject: Distinct elements in a
I was wondering if there is *any* article on the web that provides me with
implementation details and some sort of analysis on Sweet Spot Similarity?
Google shows me all the JIRA commits and comments but no article about
actual implementation. What are the various configs that could be done. What
a
On Thu, 2011-09-15 at 22:54 +0200, Pulkit Singhal wrote:
> Has anyone ever had to create large mock/dummy datasets for test
> environments or for POCs/Demos to convince folks that Solr was the
> wave of the future?
Yes, but I did it badly. The problem is that real data are not random so
any simple
> I was wondering if there is *any*
> article on the web that provides me with
> implementation details and some sort of analysis on Sweet
> Spot Similarity?
> Google shows me all the JIRA commits and comments but no
> article about
> actual implementation. What are the various configs that
> could
Hi
Scott Smith, il 16/09/2011 02:30, ha scritto:
I've been using lucene for a number of years. We've now decided to move to
SOLR. I have a couple of questions.
1. I'm used to creating Boolean queries, filter queries, term queries,
etc. for lucene. Am I right in thinking that for SOL
45 matches
Mail list logo