Re: Multiple Facet prefixes on the same facet field in one request?

2009-12-11 Thread Shalin Shekhar Mangar
On Thu, Dec 10, 2009 at 9:47 PM, Robert Purdy r...@shoptoit.ca wrote:

 Hey there Shalin,

 After scrolling through the Jira Issues I found,
 http://issues.apache.org/jira/browse/SOLR-1387, discussing this exact
 thing.Is  anyone currently working on a patch for this issue still? If not
 has the syntax in the last post in the Jira issue been approved so a patch
 can be made?


SOLR-1387 is not final. There's also SOLR-1351. A local param based syntax
looks like the right way to go. However, we have not reached consensus yet.
You are welcome to take them forward.

-- 
Regards,
Shalin Shekhar Mangar.


Re: Custom Field sample?

2009-12-11 Thread Antonio Zippo
I need to add theese features to each document

Document1
---
Argument1, positive
Argument2, positive
Argument3, neutral
Argument4, positive
Argument5, negative
Argument6, negative

Document2
---
Argument1, negative
Argument2, positive
Argument3, negative
Argument6, negative
Argument7, neutral

where the argument name is dynamic
using a relational database I could use a master detail structure, but in solr?
I thought about a Map or Pair field








Da: Grant Ingersoll gsing...@apache.org
A: solr-user@lucene.apache.org
Inviato: Gio 10 dicembre 2009, 19:47:55
Oggetto: Re: Custom Field sample?

Can you perhaps give a little more info on what problem you are trying to 
solve?  FWIW, there are a lot of examples of custom FieldTypes in the Solr code.


On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote:

 Hi all,
 
 could you help me to create a custom field?
 
 I need to create a field structured like a Map
 is it possible? how to define if the search string is on key or value (or 
 both)?
 
 A way could be to create a char separated multivalued string field... but it 
 isn't the best way. and with facets is the worst way
 
 could you give me a custom field sample?
 
 
 Thanks in advance,  
  Revenge
 
 

--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
Solr/Lucene:
http://www.lucidimagination.com/search


  

SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server'

2009-12-11 Thread regany

hi, I've (hopefully correctly) install the solr php extension.

But I'm receiving the following error when trying to run my test script:

SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server'

Any ideas how to figure out why it's giving the error??

regan


?php

/* Domain name of the Solr server */
define('SOLR_SERVER_HOSTNAME', 'localhost');

define('SOLR_SERVER_PATH', '/solr/core0');

/* Whether or not to run in secure mode */
define('SOLR_SECURE', false );

/* HTTP Port to connection */
define('SOLR_SERVER_PORT', ((SOLR_SECURE) ? 8443 : 8983));

$options = array(
'hostname' = SOLR_SERVER_HOSTNAME
,'port' = SOLR_SERVER_PORT
,'path' = SOLR_SERVER_PATH

);

$client = new SolrClient($options);
$query = new SolrQuery();
$query-setQuery('apple');
$query-setStart(0);
$query-setRows(50);
$query_response = $client-Query($query);
print_r($query_response);
$respose = $query_response-getResponse();
print_r($response);

?


-- 
View this message in context: 
http://old.nabble.com/SolrClient%3A%3Aquery%28%29%3A-Solr-HTTP-Error-%3A-%27Couldn%27t-connect-to-server%27-tp26742899p26742899.html
Sent from the Solr - User mailing list archive at Nabble.com.



Sol server is not set up ??

2009-12-11 Thread regany

Hello!

I'm trying to successfully build/install the PHP Solr Extension, but am
running into an error when doing a make test - the following 4 tests fail,
the other 17 pass. The Solr server is definately running because I can
access it via the admin URL. Anyone know what else may be causing the make
test to think teh solr server is not set up???

regan

=
Running selected tests.
TEST 1/21 [tests/solrclient_001.phpt]
SKIP SolrClient::addDocument() - Sending a single document to the Solr
server [tests/solrclient_001.phpt] reason: Solr server is not set up
TEST 2/21 [tests/solrclient_002.phpt]
SKIP SolrClient::addDocuments() - sending multiple documents to the Solr
server [tests/solrclient_002.phpt] reason: Solr server is not set up
TEST 3/21 [tests/solrclient_003.phpt]
SKIP SolrClient::addDocuments() - sending a cloned document
[tests/solrclient_003.phpt] reason: Solr server is not set up
TEST 4/21 [tests/solrclient_004.phpt]
SKIP SolrClient::query() - Sending a chained query request
[tests/solrclient_004.phpt] reason: Solr server is not set up
-- 
View this message in context: 
http://old.nabble.com/Sol-server-is-not-set-uptp26743824p26743824.html
Sent from the Solr - User mailing list archive at Nabble.com.



Using facets to narrow results with multiword field

2009-12-11 Thread Tomasz Kępski

Hi,

I'm trying to prepare narrow you search functionality using facets. I 
do have some products and would like to use a brand as a narrow filter.


I did prepare in schema 2 fileds:

   fieldType name=brand_string class=solr.TextField 
sortMissingLast=true

omitNorms=true positionIncrementGap=100
analyzer
   tokenizer class=solr.KeywordTokenizerFactory/
   filter class=solr.TrimFilterFactory /
 /analyzer
   /fieldType

   fieldType name=lower_string class=solr.TextField 
sortMissingLast=true

omitNorms=true positionIncrementGap=100
analyzer
tokenizer class=solr.KeywordTokenizerFactory/
   filter class=solr.LowerCaseFilterFactory /
   filter class=solr.TrimFilterFactory /
 /analyzer
   /fieldType


  field name=brand type=brand_string indexed=true stored=true 
default=

none/
  field name=lbrand type=lower_string indexed=true 
stored=false defaul

t=none/

copyField source=brand dest=lbrand/

I'm using facet.field=lbrand and do get good results for eg: Geomax, 
GeoMax, GEOMAX  all of them falls into geomax. But when I'm filtering 
I do get strange results:


brand:geomax  gives numFound=0
lbrand:geomax  gives numFound=57 (GEOMAX, GeoMag, Geomag)

How should I redefine brand to let narrow work correctly?

Tomek


Re: Using facets to narrow results with multiword field

2009-12-11 Thread Tomasz Kępski

Correction:

I'm using facet.field=lbrand and do get good results for eg: Geomag, 
GeoMag, GEOMAG  all of them falls into geomag. But when I'm filtering 
I do get strange results:


brand:geomag  gives numFound=0
lbrand:geomag  gives numFound=57 (GEOMAG, GeoMag, Geomag)

How should I redefine brand to let narrow work correctly?


Of course all of the words are the same (only case is different)

TK


Re: Sol server is not set up ??

2009-12-11 Thread Israel Ekpo
On Fri, Dec 11, 2009 at 7:54 AM, regany re...@newzealand.co.nz wrote:


 Hello!

 I'm trying to successfully build/install the PHP Solr Extension, but am
 running into an error when doing a make test - the following 4 tests
 fail,
 the other 17 pass. The Solr server is definately running because I can
 access it via the admin URL. Anyone know what else may be causing the make
 test to think teh solr server is not set up???

 regan

 =
 Running selected tests.
 TEST 1/21 [tests/solrclient_001.phpt]
 SKIP SolrClient::addDocument() - Sending a single document to the Solr
 server [tests/solrclient_001.phpt] reason: Solr server is not set up
 TEST 2/21 [tests/solrclient_002.phpt]
 SKIP SolrClient::addDocuments() - sending multiple documents to the Solr
 server [tests/solrclient_002.phpt] reason: Solr server is not set up
 TEST 3/21 [tests/solrclient_003.phpt]
 SKIP SolrClient::addDocuments() - sending a cloned document
 [tests/solrclient_003.phpt] reason: Solr server is not set up
 TEST 4/21 [tests/solrclient_004.phpt]
 SKIP SolrClient::query() - Sending a chained query request
 [tests/solrclient_004.phpt] reason: Solr server is not set up
 --
 View this message in context:
 http://old.nabble.com/Sol-server-is-not-set-uptp26743824p26743824.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Hi Regan,

This is Israel, the author of the PHP extension.

There is nothing wrong with your Solr server, it is just a configuration
that you have to change in the test_config.php file before running the make
test command.

In the tests/test_config.php file you will have to change the value of *
SOLR_SERVER_CONFIGURED* from *false* to* true*.

You can the contents of the file here in the repository

http://svn.php.net/viewvc/pecl/solr/trunk/tests/test.config.php?revision=290120view=markup

You also have to specify the correct values for the host name and port
numbers.

I am going to make some changes to the README files, the test scripts other
documentations to make sure that this part is clear (why some tests may be
skipped). These changes should be available in the next update release early
next week.

So, please make these changes and try again. It should not be skipped this
time.

Also, I would like to know the version of the Solr extension, the PHP
version and the operating system you are using.

Please let me know if you need any help.

Sincerely,
Israel Ekpo

-- 
Good Enough is not good enough.
To give anything less than your best is to sacrifice the gift.
Quality First. Measure Twice. Cut Once.
http://www.israelekpo.com/


Re: SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server'

2009-12-11 Thread Israel Ekpo
On Fri, Dec 11, 2009 at 6:49 AM, regany re...@newzealand.co.nz wrote:


 hi, I've (hopefully correctly) install the solr php extension.

 But I'm receiving the following error when trying to run my test script:

 SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server'

 Any ideas how to figure out why it's giving the error??

 regan


 ?php

 /* Domain name of the Solr server */
 define('SOLR_SERVER_HOSTNAME', 'localhost');

 define('SOLR_SERVER_PATH', '/solr/core0');

 /* Whether or not to run in secure mode */
 define('SOLR_SECURE', false );

 /* HTTP Port to connection */
 define('SOLR_SERVER_PORT', ((SOLR_SECURE) ? 8443 : 8983));

 $options = array(
'hostname' = SOLR_SERVER_HOSTNAME
,'port' = SOLR_SERVER_PORT
,'path' = SOLR_SERVER_PATH

 );

 $client = new SolrClient($options);
 $query = new SolrQuery();
 $query-setQuery('apple');
 $query-setStart(0);
 $query-setRows(50);
 $query_response = $client-Query($query);
 print_r($query_response);
 $respose = $query_response-getResponse();
 print_r($response);

 ?


 --
 View this message in context:
 http://old.nabble.com/SolrClient%3A%3Aquery%28%29%3A-Solr-HTTP-Error-%3A-%27Couldn%27t-connect-to-server%27-tp26742899p26742899.html
 Sent from the Solr - User mailing list archive at Nabble.com.



Hi Regan,

I have the following questions:

0. What version of Apache Solr are you using? 1.3, 1.4, nightly builds?

1. What version of PHP are you using and on what operating system?

2. What version of the Solr extension are you using?

3. Which servlet container are you using for Solr? (Jetty, Tomcat, Glass
fish etc)

4. What is the hostname and port numbers and path to Solr? Is your port
number 8080 or 8983

All please let me know what the output of $client-getDebug() is. This
usually contains very detailed errors of what is happening during the
connection.

I would be happy to help you troubleshoot any errors you are having.


-- 
Good Enough is not good enough.
To give anything less than your best is to sacrifice the gift.
Quality First. Measure Twice. Cut Once.
http://www.israelekpo.com/


Simple Wildcard Search Question

2009-12-11 Thread QBasti

Hey there,

I'm pretty new to SOLR and I tried to understand the essentials of
searching, analyzing etc..
As far as I understand, when I index the string name01 it gets split into
name as well as 01. So when I'm searching for either name or 01 oder
name01, i'm getting the right result.
But since, i think, SOLR stores the string also under name01, why don't a
get any result when searching for name0* oder name0?

thanks for any help
QBasti
-- 
View this message in context: 
http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26747482.html
Sent from the Solr - User mailing list archive at Nabble.com.



RE: Simple Wildcard Search Question

2009-12-11 Thread Ankit Bhatnagar
Firstly wild card queries are not analysed by Solr.

Also there is a difference between stored and indexed fields.


Ankit

-Original Message-
From: QBasti [mailto:sebastian.f...@gmail.com] 
Sent: Friday, December 11, 2009 11:40 AM
To: solr-user@lucene.apache.org
Subject: Simple Wildcard Search Question


Hey there,

I'm pretty new to SOLR and I tried to understand the essentials of
searching, analyzing etc..
As far as I understand, when I index the string name01 it gets split into
name as well as 01. So when I'm searching for either name or 01 oder
name01, i'm getting the right result.
But since, i think, SOLR stores the string also under name01, why don't a
get any result when searching for name0* oder name0?

thanks for any help
QBasti
-- 
View this message in context: 
http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26747482.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Multiple Facet prefixes on the same facet field in one request?

2009-12-11 Thread Robert Purdy
Well I looked at SOLR-1351 and applied the patch and that seems to do 
everything that I need. The local param syntax definitely the way to go.

Robert,

On 2009-12-11, at 2:53 AM, Shalin Shekhar Mangar wrote:

 On Thu, Dec 10, 2009 at 9:47 PM, Robert Purdy r...@shoptoit.ca wrote:
 
 Hey there Shalin,
 
 After scrolling through the Jira Issues I found,
 http://issues.apache.org/jira/browse/SOLR-1387, discussing this exact
 thing.Is  anyone currently working on a patch for this issue still? If not
 has the syntax in the last post in the Jira issue been approved so a patch
 can be made?
 
 
 SOLR-1387 is not final. There's also SOLR-1351. A local param based syntax
 looks like the right way to go. However, we have not reached consensus yet.
 You are welcome to take them forward.
 
 -- 
 Regards,
 Shalin Shekhar Mangar.




RE: SolrPlugin Guidance

2009-12-11 Thread Vauthrin, Laurent
It looks like the SolrQueryParser constructor accepts an analyzer as a
parameter.  That seems to do the trick.  Although feel free to respond
anyway if you have a comment on the approach :)

-Original Message-
From:
solr-user-return-30215-laurent.vauthrin=disney@lucene.apache.org
[mailto:solr-user-return-30215-laurent.vauthrin=disney@lucene.apache
.org] On Behalf Of Vauthrin, Laurent
Sent: Thursday, December 10, 2009 11:44 AM
To: solr-user@lucene.apache.org
Subject: RE: SolrPlugin Guidance

Ok, looks like I may not be taking the right approach here.  I'm running
a problem.

Let's say a user is looking for all files in any directory 'foo' with a
directory description 'bar' 

q:+directory_name:foo +directory_description:bar

Our QParser plugin will perform queries against directory documents and
return any file document that has the matching directory id(s).  So the
plugin transforms the query to something like 

q:+(directory_id:4 directory:10) +directory_id:(4)

Note: directory_id is only in file documents.  The query above assumes
that two directories had the name 'foo' but only one had the description
'bar'

Currently the parser plugin is doing the lookup queries via the standard
request handler.  The problem with this approach is that the look up
queries are going to be analyzed twice.  This only seems to be a problem
because we're using stemming.  For example, stemming 'franchise' gives
'franchis' and stemming it again gives 'franchi'.  The second stemming
will cause the query not to match anymore.

So basically my questions are:
1. Should I not be passing my lookup queries back to the request
handler, but instead to some lower level component?  If so, which
component would be good to look at?
2. Is there a way to tell the SolrQueryParser not to analyze or a
different way to run the query so that they query analysis won't happen?

Thanks again,
Laurent Vauthrin

-Original Message-
From:
solr-user-return-30170-laurent.vauthrin=disney@lucene.apache.org
[mailto:solr-user-return-30170-laurent.vauthrin=disney@lucene.apache
.org] On Behalf Of Vauthrin, Laurent
Sent: Wednesday, December 09, 2009 2:53 PM
To: solr-user@lucene.apache.org
Subject: RE: SolrPlugin Guidance

Thanks for the response.  I went ahead and gave it a shot.  In my case,
the directory name may not be unique so if I get multiple ids back then
I create a BooleanQuery (Occur.SHOULD) to substitute the directory name
query.  This seems to work at the moment so hopefully that's the right
approach. 

Thanks,
Laurent Vauthrin


-Original Message-
From:
solr-user-return-30054-laurent.vauthrin=disney@lucene.apache.org
[mailto:solr-user-return-30054-laurent.vauthrin=disney@lucene.apache
.org] On Behalf Of Chris Hostetter
Sent: Monday, December 07, 2009 12:17 PM
To: solr-user@lucene.apache.org
Subject: RE: SolrPlugin Guidance


: e.g. For the following query that looks for a file in a directory:
: q=+directory_name:myDirectory +file_name:myFile
: 
: We'd need to decompose the query into the following two queries:
: 1. q=+directory_name:myDirectoryfl=directory_id
: 2. q=+file_name:myFile +directory_id:(results from query #1)
: 
: I guess I'm looking for the following feedback:
: - Does this sound crazy?  

it's a little crazy, but not absurd.

: - Is the QParser the right place for this logic?  If so, can I get a 
: little more guidance on how to decompose the queries there (filter 
: queries maybe)?

a QParser could work. (and in general, if you can solve something with a

QParser that's probably for the best, since it allows the most reuse).
but 
exactly how to do it depends on how many results you expect from your 
first query:  if you are going to structure things so they have to 
uniquely id a directory, and you'll have a singleID, then this is 
something that could easily make sense in a QParser (you are essentailly

just rewriting part of the query from string to id -- you just happen to

be using solr as a lookup table for those strings).

but if you plan to support any arbitrary N directories, then you may 
need something more complicated ... straight filter queries won't help 
much because you'll want the union instead of hte intersection, so for 
every directoryId you find, use it as a query to get a DocSet and then 
maintain a running union of all those DocSets to use as your final
filter 
(hmm... that may not actually be possible with the QParser API ... i 
haven't look at ti in a while, but for an approach like this you may
beed 
to subclass QueryComponent instead)




-Hoss



Re: Simple Wildcard Search Question

2009-12-11 Thread Erick Erickson
In addition to Ankit's comments, SOLR manipulates the
input for each field according to the rules that govern
the tokenizers and filters that you define in your schema.

This page is very valuable, although it takes a while to
get comfortable with it...

http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters

So your question is hard to answer unless you tell us what
your analyzers (index and query) for the field in question
are

Best
Erick



On Fri, Dec 11, 2009 at 11:39 AM, QBasti sebastian.f...@gmail.com wrote:


 Hey there,

 I'm pretty new to SOLR and I tried to understand the essentials of
 searching, analyzing etc..
 As far as I understand, when I index the string name01 it gets split into
 name as well as 01. So when I'm searching for either name or 01
 oder
 name01, i'm getting the right result.
 But since, i think, SOLR stores the string also under name01, why don't a
 get any result when searching for name0* oder name0?

 thanks for any help
 QBasti
 --
 View this message in context:
 http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26747482.html
 Sent from the Solr - User mailing list archive at Nabble.com.




RE: full-text indexing XML files

2009-12-11 Thread Feroze Daud
CDATA didn’t work either.It still complained about the input doc not being in 
correct format.

-Original Message-
From: Lance Norskog [mailto:goks...@gmail.com] 
Sent: Thursday, December 10, 2009 7:43 PM
To: solr-user@lucene.apache.org
Subject: Re: full-text indexing XML files

Or CDATA (much easier to work with).

On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar
shalinman...@gmail.com wrote:
 On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com wrote:

 Hi!



 I am trying to full text index an XML file. For various reasons, I
 cannot use Tika or other technology to parse the XML file. The
 requirement is to full-text index the XML file, including Tags and
 everything.



 So, I created a input index spec like this:



 add

 doc

 field name=id1001/field

 field name=nameNASA Advanced Research Labs/field

 field name=address1010 Main Street, Chattanooga, FL 32212/field

 field name=contentlistingid1001/id name  NASA Advanced
 Research Labs / name address1010 main street, chattanooga, FL
 32212/address/listing/field

 /doc

 /add



 You need to XML encode the value of the content field.

 --
 Regards,
 Shalin Shekhar Mangar.




-- 
Lance Norskog
goks...@gmail.com


RE: full-text indexing XML files

2009-12-11 Thread Feroze Daud
Yeah, xml tags as well. Essentially we want to full-text index the file,
without the need for stemming the tokens.

Will the SOLR analyzer be able to tokenize the document correctly if it
does not have any whitespaces (besides those required by XML syntax)?

-Original Message-
From: Walter Underwood [mailto:wun...@wunderwood.org] 
Sent: Thursday, December 10, 2009 8:00 PM
To: solr-user@lucene.apache.org
Subject: Re: full-text indexing XML files

What kind of searches do you want to do? Do you want to do searches that
match the XML tags?

wunder

On Dec 10, 2009, at 7:43 PM, Lance Norskog wrote:

 Or CDATA (much easier to work with).
 
 On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar
 shalinman...@gmail.com wrote:
 On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com
wrote:
 
 Hi!
 
 
 
 I am trying to full text index an XML file. For various reasons, I
 cannot use Tika or other technology to parse the XML file. The
 requirement is to full-text index the XML file, including Tags and
 everything.
 
 
 
 So, I created a input index spec like this:
 
 
 
 add
 
 doc
 
 field name=id1001/field
 
 field name=nameNASA Advanced Research Labs/field
 
 field name=address1010 Main Street, Chattanooga, FL
32212/field
 
 field name=contentlistingid1001/id name  NASA Advanced
 Research Labs / name address1010 main street, chattanooga, FL
 32212/address/listing/field
 
 /doc
 
 /add
 
 
 
 You need to XML encode the value of the content field.
 
 --
 Regards,
 Shalin Shekhar Mangar.
 
 
 
 
 -- 
 Lance Norskog
 goks...@gmail.com
 



Re: Simple Wildcard Search Question

2009-12-11 Thread QBasti

Hey,

well, the field is stored as well as indexed, and these are the filters that
are used while indexing:
W
-- 
View this message in context: 
http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26748588.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: full-text indexing XML files

2009-12-11 Thread Walter Underwood
If you really want to do XML-senstive search, it could be a lot of work in 
Solr. Lucene is a flat data model, so hierarchy requires a lot of mapping to 
the schema or fancy, slow queries.

There are engines that are designed for XML indexing and search, using XQuery, 
so consider whether that might be less work overall.

XML engines are less mature than Lucene and Solr, so there is a big performance 
and scalability gap between the best free engines (eXist) and the best 
commercial engines (Mark Logic, where I work).

wunder
Walter Underwood
Lead Engineer, Mark Logic

On Dec 11, 2009, at 9:42 AM, Feroze Daud wrote:

 Yeah, xml tags as well. Essentially we want to full-text index the file,
 without the need for stemming the tokens.
 
 Will the SOLR analyzer be able to tokenize the document correctly if it
 does not have any whitespaces (besides those required by XML syntax)?
 
 -Original Message-
 From: Walter Underwood [mailto:wun...@wunderwood.org] 
 Sent: Thursday, December 10, 2009 8:00 PM
 To: solr-user@lucene.apache.org
 Subject: Re: full-text indexing XML files
 
 What kind of searches do you want to do? Do you want to do searches that
 match the XML tags?
 
 wunder
 
 On Dec 10, 2009, at 7:43 PM, Lance Norskog wrote:
 
 Or CDATA (much easier to work with).
 
 On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar
 shalinman...@gmail.com wrote:
 On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com
 wrote:
 
 Hi!
 
 
 
 I am trying to full text index an XML file. For various reasons, I
 cannot use Tika or other technology to parse the XML file. The
 requirement is to full-text index the XML file, including Tags and
 everything.
 
 
 
 So, I created a input index spec like this:
 
 
 
 add
 
 doc
 
 field name=id1001/field
 
 field name=nameNASA Advanced Research Labs/field
 
 field name=address1010 Main Street, Chattanooga, FL
 32212/field
 
 field name=contentlistingid1001/id name  NASA Advanced
 Research Labs / name address1010 main street, chattanooga, FL
 32212/address/listing/field
 
 /doc
 
 /add
 
 
 
 You need to XML encode the value of the content field.
 
 --
 Regards,
 Shalin Shekhar Mangar.
 
 
 
 
 -- 
 Lance Norskog
 goks...@gmail.com
 
 



Definitive version of acts_as_solr

2009-12-11 Thread Jacob Elder
What versions of acts_as_solr are you all using?

There appears to be about a dozen forks on GitHub, including my own.
http://acts-as-solr.rubyforge.org/ has a notice that the official site is
now http://acts_as_solr.railsfreaks.com/, but *don't click that
link*because it's just a mess of pop-up ads now. It would be great to
get some
consolidation and agreement from the community.

-- 
Jacob Elder


Re: Simple Wildcard Search Question

2009-12-11 Thread Erick Erickson
Nothing came through showing the filters

Need analyzer definitions BOTH for indexing and querying, the correlation
between the two is important.

Storing and indexing are orthogonal. When you *index* a field, you are
putting the tokens that come from the input stream into the inverted index,
possibly applying transformations such as lowercasing, splitting various
ways, etc.. The result of *all* these transformations is what is searched
against.

When you *store* a field, the raw data with *no* processing
is put in the index. This data is never used for searching.

So, say you index and store a field. You can think of it as producing
two fields, the searchable tokens and the raw copy of the
data.

HTH
Erick

On Fri, Dec 11, 2009 at 12:48 PM, QBasti sebastian.f...@gmail.com wrote:


 Hey,

 well, the field is stored as well as indexed, and these are the filters
 that
 are used while indexing:
 W
 --
 View this message in context:
 http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26748588.html
 Sent from the Solr - User mailing list archive at Nabble.com.




List shards in the admin UI?

2009-12-11 Thread Jason Rutherglen
Seems like an ease of use thing to be able to click to shards from the admin UI?


Request Assistance with DIH

2009-12-11 Thread Robbin
I've been trying to use the DIH with oracle and would love it if someone 
could give me some pointers.  I put the ojdbc14.jar in both the Tomcat 
lib and solr home/lib.  I created a dataimport.xml and enabled it in 
the solrconfig.xml.  I go to the http://solr 
server/solr/admin/dataimport.jsp.  This all seems to be fine, but I get 
the default page response and doesn't look like the connection to the 
oracle server is even attempted.


I'm using the Solr 1.4 release on Nov 10. 

Do I need an oracle client on the server?  I thought having the ojdbc 
jar should be sufficient.  Any help or configuration examples for 
setting this up would be much appreciated.


Thanks
Robbin


Re: UI for solr core admin?

2009-12-11 Thread Chris Hostetter

: Well sure, Solr is no webapp framework.  But you can do some things with the
: ShowFileRequestHandler, like have static textual content (like CSS and

While i think it's definitely important to have some basic admin 
functionality available in the war, at a certain point we should really 
just focus on making sure there's a good HTTP/XMl/JSON based API for 
everything, and perhaps distribute the Admin console as a seperate webapp 
(or perhaps not even a webapp, maybe just a collection of HTML files that 
use AJAX to do everything)

ShowFileRequestHandler can take you pretty far, but eventually you either 
have to say put this big hunk of stuff in your solrcofig.xml or it won't 
work or you have to have the admin tool inspect hte core to get a lot of 
data it needs to build up the admin pages -- so let's just expose all that 
data via XML/JSON and (ie: registry.jsp on steriods) and then any external 
tool (built by us in velocity, built by someone else in ruby, built by 
several differnet people to integrate into several differnet tools) can 
use it to get the metadata it needs to drive tool behavior.

I had some notes along htese lines once, ... ah yes ...

   http://wiki.apache.org/solr/MakeSolrMoreSelfService

...some of that stuff has already come to pass, and the rest is pretty out 
of date with how we do things now, but the idea of having a clean API to 
discover what handlers a given solr port exposes, and what params that 
instance of those handlers say they accept is still a really good idea.

in the context of this discussion it's a little differnet because we're 
taling specificly about hte CoreAdminHandler -- but hte principle is hte 
same.



-Hoss



Request Assistance with DIH

2009-12-11 Thread Robbin

I've been trying to use the DIH with oracle and would love it if someone
could give me some pointers.  I put the ojdbc14.jar in both the Tomcat
lib and solr home/lib.  I created a dataimport.xml and enabled it in
the solrconfig.xml.  I go to the http://solr
server/solr/admin/dataimport.jsp.  This all seems to be fine, but I get
the default page response and doesn't look like the connection to the
oracle server is even attempted.

I'm using the Solr 1.4 release on Nov 10.

Do I need an oracle client on the server?  I thought having the ojdbc
jar should be sufficient.  Any help or configuration examples for
setting this up would be much appreciated.

Thanks
Robbin



using q= , adding fq=

2009-12-11 Thread Fer-Bj

We're running a 14M documents index. For each document we have:
   field name=id type=sint indexed=true  
stored=true
required=true / 
   field name=title  type=text_ngram indexed=true
stored=trueomitNorms=true/
   field name=cat_id type=sint indexed=true  
stored=true/
   field name=geo_id type=sint indexed=true  
stored=true/
   field name=body   type=text indexed=true  
stored=false
omitNorms=true/
   field name=modified_datetime  type=date indexed=true 
stored=true/
(and a few other fields).

Our most usual query is something like this:
q=cat_id:xxx AND geo_id:sort=id desc   where cat_id = which category
(cars,sports,toys,etc) the item belongs to, and geo_id = which city/district
the item belongs to.
So this query will return a list of documents posted in category xxx, region
yyy. 
Sorted by ID DESC, to get the newest first.

There are 2 questions I'd like to ask:

1) adding something like:  q=cat_id:xxxfq=geo_id= would boost
performance?

2) we do find problems when we ask for a page=large offset!  ie: 
q=cat_id:xxx and geo_id:yyystart=544545
(note that we limit docs to 50 max per resultset).
When start is 500 or more, Qtime is =5 seconds while the avg qtime is
100 ms

Any help or tips would be appreciated!

Thanks,



-- 
View this message in context: 
http://old.nabble.com/using-q%3D--%2C-adding-fq%3D-tp26753938p26753938.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Request Assistance with DIH

2009-12-11 Thread Joel Nylund

add ?command=full-import to your url

http://localhost:8983/solr/dataimport?command=full-import

thanks
Joel

On Dec 11, 2009, at 7:45 PM, Robbin wrote:

I've been trying to use the DIH with oracle and would love it if  
someone could give me some pointers.  I put the ojdbc14.jar in both  
the Tomcat lib and solr home/lib.  I created a dataimport.xml and  
enabled it in the solrconfig.xml.  I go to the http://solr server/ 
solr/admin/dataimport.jsp.  This all seems to be fine, but I get the  
default page response and doesn't look like the connection to the  
oracle server is even attempted.


I'm using the Solr 1.4 release on Nov 10.
Do I need an oracle client on the server?  I thought having the  
ojdbc jar should be sufficient.  Any help or configuration examples  
for setting this up would be much appreciated.


Thanks
Robbin




auto-starting Solr on OS X ?

2009-12-11 Thread regany

hello!

does anyone know how you go about setting up Solr so it auto starts after a
reboot etc on OS X?

thanks,

regan


-- 
View this message in context: 
http://old.nabble.com/auto-starting-Solr-on-OS-X---tp26753997p26753997.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: auto-starting Solr on OS X ?

2009-12-11 Thread Erick Erickson
See http://www.macosxhints.com/article.php?story=20041105070509783

In general, crontab is what you want, probably with the @Reboot
frequency

HTH
Erick

On Fri, Dec 11, 2009 at 10:04 PM, regany re...@newzealand.co.nz wrote:


 hello!

 does anyone know how you go about setting up Solr so it auto starts after a
 reboot etc on OS X?

 thanks,

 regan


 --
 View this message in context:
 http://old.nabble.com/auto-starting-Solr-on-OS-X---tp26753997p26753997.html
 Sent from the Solr - User mailing list archive at Nabble.com.




Re: List shards in the admin UI?

2009-12-11 Thread Erik Hatcher

On Dec 11, 2009, at 11:57 PM, Jason Rutherglen wrote:
Seems like an ease of use thing to be able to click to shards from  
the admin UI?


Where would the shards list come from?

In many cases, it makes sense for shards to be dynamically controlled  
per request though, so it may not be baked into config.


Erik



Re: Custom Field sample?

2009-12-11 Thread Lance Norskog
If 'Argument#' is unique, you can just use as a wildcard field name:
Argument1_s:positive is a string field.

If you want to find out which Argument fields a document has, you have
to fetch all fields for a document.

On Fri, Dec 11, 2009 at 3:05 AM, Antonio Zippo reven...@yahoo.it wrote:
 I need to add theese features to each document

 Document1
 ---
 Argument1, positive
 Argument2, positive
 Argument3, neutral
 Argument4, positive
 Argument5, negative
 Argument6, negative

 Document2
 ---
 Argument1, negative
 Argument2, positive
 Argument3, negative
 Argument6, negative
 Argument7, neutral

 where the argument name is dynamic
 using a relational database I could use a master detail structure, but in 
 solr?
 I thought about a Map or Pair field







 
 Da: Grant Ingersoll gsing...@apache.org
 A: solr-user@lucene.apache.org
 Inviato: Gio 10 dicembre 2009, 19:47:55
 Oggetto: Re: Custom Field sample?

 Can you perhaps give a little more info on what problem you are trying to 
 solve?  FWIW, there are a lot of examples of custom FieldTypes in the Solr 
 code.


 On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote:

 Hi all,

 could you help me to create a custom field?

 I need to create a field structured like a Map
 is it possible? how to define if the search string is on key or value (or 
 both)?

 A way could be to create a char separated multivalued string field... but it 
 isn't the best way. and with facets is the worst way

 could you give me a custom field sample?


 Thanks in advance,
  Revenge



 --
 Grant Ingersoll
 http://www.lucidimagination.com/

 Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
 Solr/Lucene:
 http://www.lucidimagination.com/search






-- 
Lance Norskog
goks...@gmail.com


Re: Definitive version of acts_as_solr

2009-12-11 Thread Erik Hatcher

Jacob,

Yeah, it really is a big mess with acts_as_solr.  I blame a few  
things: 1) me - for not wrangling this thing into Solr's client/ruby  
area from the start, 2) Rubyists!  (Apache's svn isn't an appealing  
home) and 3) the git craze promoting fork-mania.


And take a look at this ridiculous bug (in at least one fork) of  
acts_as_solr: http://groups.google.com/group/acts_as_solr/browse_thread/thread/4e1dee03a46c6415 



So... what do we do?

Can we get the best version of acts_as_solr donated to Solr's svn?   
Who can rightfully say where the IP of that comes from so we can  
ensure it's provenance is clear and cool by all who touched the  
code?   Or maybe we just start from scratch with the simplest thing  
that can possibly work (like my acts_as_solr_lite (page 13 of http://code4lib.org/files/solr-ruby.pdf) 
 and iterate from there in Apache's svn?


Erik


On Dec 11, 2009, at 7:52 PM, Jacob Elder wrote:


What versions of acts_as_solr are you all using?

There appears to be about a dozen forks on GitHub, including my own.
http://acts-as-solr.rubyforge.org/ has a notice that the official  
site is

now http://acts_as_solr.railsfreaks.com/, but *don't click that
link*because it's just a mess of pop-up ads now. It would be great to
get some
consolidation and agreement from the community.

--
Jacob Elder




Re: full-text indexing XML files

2009-12-11 Thread Lance Norskog
Please post a small sample file that has this problem with CDATA.

On Fri, Dec 11, 2009 at 9:41 AM, Feroze Daud fero...@zillow.com wrote:
 CDATA didn’t work either.It still complained about the input doc not being in 
 correct format.

 -Original Message-
 From: Lance Norskog [mailto:goks...@gmail.com]
 Sent: Thursday, December 10, 2009 7:43 PM
 To: solr-user@lucene.apache.org
 Subject: Re: full-text indexing XML files

 Or CDATA (much easier to work with).

 On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar
 shalinman...@gmail.com wrote:
 On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com wrote:

 Hi!



 I am trying to full text index an XML file. For various reasons, I
 cannot use Tika or other technology to parse the XML file. The
 requirement is to full-text index the XML file, including Tags and
 everything.



 So, I created a input index spec like this:



 add

 doc

 field name=id1001/field

 field name=nameNASA Advanced Research Labs/field

 field name=address1010 Main Street, Chattanooga, FL 32212/field

 field name=contentlistingid1001/id name  NASA Advanced
 Research Labs / name address1010 main street, chattanooga, FL
 32212/address/listing/field

 /doc

 /add



 You need to XML encode the value of the content field.

 --
 Regards,
 Shalin Shekhar Mangar.




 --
 Lance Norskog
 goks...@gmail.com




-- 
Lance Norskog
goks...@gmail.com


Re: Custom Field sample?

2009-12-11 Thread Noble Paul നോബിള്‍ नोब्ळ्
how exactly do you wish to query these documents?

On Fri, Dec 11, 2009 at 4:35 PM, Antonio Zippo reven...@yahoo.it wrote:
 I need to add theese features to each document

 Document1
 ---
 Argument1, positive
 Argument2, positive
 Argument3, neutral
 Argument4, positive
 Argument5, negative
 Argument6, negative

 Document2
 ---
 Argument1, negative
 Argument2, positive
 Argument3, negative
 Argument6, negative
 Argument7, neutral

 where the argument name is dynamic
 using a relational database I could use a master detail structure, but in 
 solr?
 I thought about a Map or Pair field







 
 Da: Grant Ingersoll gsing...@apache.org
 A: solr-user@lucene.apache.org
 Inviato: Gio 10 dicembre 2009, 19:47:55
 Oggetto: Re: Custom Field sample?

 Can you perhaps give a little more info on what problem you are trying to 
 solve?  FWIW, there are a lot of examples of custom FieldTypes in the Solr 
 code.


 On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote:

 Hi all,

 could you help me to create a custom field?

 I need to create a field structured like a Map
 is it possible? how to define if the search string is on key or value (or 
 both)?

 A way could be to create a char separated multivalued string field... but it 
 isn't the best way. and with facets is the worst way

 could you give me a custom field sample?


 Thanks in advance,
  Revenge



 --
 Grant Ingersoll
 http://www.lucidimagination.com/

 Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
 Solr/Lucene:
 http://www.lucidimagination.com/search






-- 
-
Noble Paul | Systems Architect| AOL | http://aol.com


Re: List shards in the admin UI?

2009-12-11 Thread Lance Norskog
The cores inside the Solr instance? That should be easy.

There are a lot of things that should be added to 'Full Interface'.

On Fri, Dec 11, 2009 at 7:31 PM, Erik Hatcher erik.hatc...@gmail.com wrote:
 On Dec 11, 2009, at 11:57 PM, Jason Rutherglen wrote:

 Seems like an ease of use thing to be able to click to shards from the
 admin UI?

 Where would the shards list come from?

 In many cases, it makes sense for shards to be dynamically controlled per
 request though, so it may not be baked into config.

        Erik





-- 
Lance Norskog
goks...@gmail.com


Auto update with deltaimport

2009-12-11 Thread Olala

Hi All!

I am developing a search engine using Solr, I was tested full-import and
delta-import command successfully.But now,I want to run delta-import
automatically with my schedule.So, can anyone help me???

Thanks  Regards,
-- 
View this message in context: 
http://old.nabble.com/Auto-update-with-deltaimport-tp26755386p26755386.html
Sent from the Solr - User mailing list archive at Nabble.com.