Re: Ranking result on the basis of field value irrespective of score

2008-10-24 Thread Shalin Shekhar Mangar
Create your languages field (or a copy) in such a way that sort gives you
the order you want. Then just query Solr with sort=language asc parameter.

2008/10/23 Amit [EMAIL PROTECTED]

 Hi All,



 How  we can do the ranking on the basis of specific field value
 irrespective
 of score in solr?



 For example:

   Let say field language which content values
 like German,English,French,Chines,Arabic.

  So I want English language document come first irrespective of
 score.



 Thanks in advance any kind reply.



 Regards,

  Amit





 No virus found in this outgoing message.
 Checked by AVG.
 Version: 7.5.549 / Virus Database: 270.8.2/1740 - Release Date: 22-10-2008
 19:24




-- 
Regards,
Shalin Shekhar Mangar.


RE: One document inserted but nothing showing up ? SOLR 1.3

2008-10-24 Thread sunnyfr

Even that doesn't work,
How can I check properly, I did insert one document but I can't get it back
???


Feak, Todd wrote:
 
 Unless q=ALL is a special query I don't know about, the only reason you
 would get results is if ALL showed up in the default field of the single
 document that was inserted/updated.
 
 You could try a query of *:* instead. Don't forget to URL encode if you
 are doing this via URL.
 
 -Todd
 
 
 -Original Message-
 From: sunnyfr [mailto:[EMAIL PROTECTED] 
 Sent: Thursday, October 23, 2008 9:17 AM
 To: solr-user@lucene.apache.org
 Subject: One document inserted but nothing showing up ? SOLR 1.3
 
 
 Hi 
 
 Can somebody help me ?
 How can I see all my documents, I just did a full import :
 str name=
 Indexing completed. Added/Updated: 1 documents. Deleted 0 documents.
 /str
 
 and when I do :8180/solr/video/select/?q=ALL, I've no result ?
 response
 −
 lst name=responseHeader
 int name=status0/int
 int name=QTime0/int
 −
 lst name=params
 str name=qALL/str
 /lst
 /lst
 result name=response numFound=0 start=0/
 /response
 
 Thanks a lot,
 
 -- 
 View this message in context:
 http://www.nabble.com/One-document-inserted-but-nothing-showing-up---SOLR-1.3-tp20134357p20134357.html
 Sent from the Solr - User mailing list archive at Nabble.com.
 
 
 
 

-- 
View this message in context: 
http://www.nabble.com/One-document-inserted-but-nothing-showing-up---SOLR-1.3-tp20134357p20145343.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: How to search a DataImportHandler solr index

2008-10-24 Thread Nick80

Hi Paul,

thanks for the answer but unfortunately it doesn't work. I have the
following:

entity name=campaign
field name=id column=id / 
field name=campaign_name column=campaign_name /

entity name=banner
   field name=banner_type column=banner_type /
   
entity name=size
field name=size column=size /
/entity
 /entity
 /entity

I have defined banner_type and size as:

field name=banner_type type=string indexed=true stored=true
multiValued=true omitNorms=true termVectors=true /

Now when I do a search with Solr, I get:

result name=response numFound=1 start=0
doc
  str name=id1/str
  str name=campaign_nameCampaign Name/str

  arr name=banner_type
 strflash/str
 strgif/str
  /arr

  arr name=size
 str50x50/str
 str100x100/str
 str50x50/str
 str100x100/str
  /arr
/doc
/result

While I was expecting that the size tags were inside the banner_type tags,
something like:

  arr name=banner_type
 strflash
  arr name=size
 str50x50/str
 str100x100/str
 /arr
 /str
 strgif
 arr name=size
 str50x50/str
 str100x100/str
 /arr
 /str
  /arr

Am I doing something wrong or is it just not possible? Because with the
output it generates now I cann't accurately find a campaign that has a flash
banner of size 50x50 for example, because the size 50x50 could as well be
from a gif banner. With the nested structure, I think it would be possible.
At least if Solr can search this type of structure. Any tips are welcome.
Thanks.

Kind regards,

Nick
-- 
View this message in context: 
http://www.nabble.com/How-to-search-a-DataImportHandler-solr-index-tp20120698p20145974.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: How to search a DataImportHandler solr index

2008-10-24 Thread Noble Paul നോബിള്‍ नोब्ळ्
probably u can paste your data-config.xml with the queries etc

--Noble

On Fri, Oct 24, 2008 at 1:33 PM, Nick80 [EMAIL PROTECTED] wrote:

 Hi Paul,

 thanks for the answer but unfortunately it doesn't work. I have the
 following:

 entity name=campaign
field name=id column=id /
field name=campaign_name column=campaign_name /

entity name=banner
   field name=banner_type column=banner_type /

entity name=size
field name=size column=size /
/entity
 /entity
  /entity

 I have defined banner_type and size as:

 field name=banner_type type=string indexed=true stored=true
 multiValued=true omitNorms=true termVectors=true /

 Now when I do a search with Solr, I get:

 result name=response numFound=1 start=0
 doc
  str name=id1/str
  str name=campaign_nameCampaign Name/str

  arr name=banner_type
 strflash/str
 strgif/str
  /arr

  arr name=size
 str50x50/str
 str100x100/str
 str50x50/str
 str100x100/str
  /arr
 /doc
 /result

 While I was expecting that the size tags were inside the banner_type tags,
 something like:

  arr name=banner_type
 strflash
  arr name=size
 str50x50/str
 str100x100/str
 /arr
 /str
 strgif
 arr name=size
 str50x50/str
 str100x100/str
 /arr
 /str
  /arr

 Am I doing something wrong or is it just not possible? Because with the
 output it generates now I cann't accurately find a campaign that has a flash
 banner of size 50x50 for example, because the size 50x50 could as well be
 from a gif banner. With the nested structure, I think it would be possible.
 At least if Solr can search this type of structure. Any tips are welcome.
 Thanks.

 Kind regards,

 Nick
 --
 View this message in context: 
 http://www.nabble.com/How-to-search-a-DataImportHandler-solr-index-tp20120698p20145974.html
 Sent from the Solr - User mailing list archive at Nabble.com.





-- 
--Noble Paul


Re: delta-import for XML files, Solr statistics

2008-10-24 Thread Noble Paul നോബിള്‍ नोब्ळ्
On Fri, Oct 24, 2008 at 5:14 PM,  [EMAIL PROTECTED] wrote:
 Hello,

 I have some questions about DataImportHandler and Solr statistics...


 1.)
 I'm using the DataImportHandler for creating my Lucene index from XML files:

 ###
 $ cat data-config.xml
 dataConfig
  dataSource type=FileDataSource /
  document
   entity name=xmlFile
processor=FileListEntityProcessor
baseDir=/tmp/files
fileName=myDoc_.*\.xml
newerThan='NOW-30DAYS'
recursive=false
rootEntity=false
dataSource=null
entity name=myDoc
  url=${xmlFile.fileAbsolutePath}
  processor=XPathEntityProcessor
  forEach=/myDoc
  ...
 /dataConfig
 ###

 No problems with this configuration - All works fine for full-imports, but...

 === What means 'rootEntity=false' and 'dataSource=null'?

It is a menace caused by 'sensible defaults'

An entity directly under the document is a root entity. That means
that for each row emitted by the root entity one document is created
in Solr/Lucene . but as in this case we do not wish to make one
document per file. we wish to make one document per row emitted by the
entity 'myDoc' .Because the entity 'xmlFile' is not has
rootEntity=false the entity directly under it becomes a root entity
automatically and each row emitted by that becomes a document.

In most of the cases there is only one datasource (A JdbcDataSource)
and all entities just use them . So it is an overkill to ask them to
write the datSource. So we have chosen to implicitly assign the
datasource with no name to that entity. But in case of
FileListEntityProcessor a datasource is not necessary . But it won't
hurt even if you do not put dataSource=null  . It just means that we
won't create a DataSource instance for that.





 2.)
 The documentation from DataImportHandler describes the index update process 
 for SQL databases only...

 My scenario:
 - My application creates, deletes and modifies files from /tmp/files every 
 night.
 - delta-import / DataImportHandler should mirror _all_ this changes to my 
 lucene index (= create, delete, update documents).
The only Entityprocessor which supports delta is SqlEntityProcessor.
The XPathEntityProcessor has not implemented it , because we do not
know of a consistent way of finding deltas for XML. So ,
unfortunately,no delta support for XML. But that said you can
implement those methods in XPathEntityProcessor . The methods are
explained in EntityProcessor.java. if you have questions specific to
this I can help.Probably we can contribute it back

 === Is this possible with delta-import / DataImportHandler?
 === If not: Do you have any suggestions on how to do this?



 3.)
 My scenario:
 - /tmp/files contains 682 'myDoc_.*\.xml' XML files.
 - Each XML file contains 12 XML elements (e.g. titlefoo/title).
 - DataImportHandler transfer only 5 from this 12 elements to the lucene index.


 I don't understand the output from 'solr/dataimport' (= status):

 ###
 response
  ...
  lst name=statusMessages
  str name=Total Requests made to DataSource0/str
  str name=Total Rows Fetched1363/str
  str name=Total Documents Skipped0/str
  str name=Full Dump Started2008-10-24 13:19:03/str
  str name=
Indexing completed. Added/Updated: 681 documents. Deleted 0 documents.
  /str
  str name=Committed2008-10-24 13:19:05/str
  str name=Optimized2008-10-24 13:19:05/str
  str name=Time taken 0:0:2.648/str
  /lst
 ...
 /response

 === What is Total Rows Fetched rsp. what is a row in a XML file? An 
 element? Why 1363?
 === Why shows the Added/Updated counter 681 and not 682?

rows fethed makes a lot of sense with SqlEntityProcessor. It is the
no:of rows fetched from DB . It is the cumulative no:of rows given out
by all entitiies put together. in your case it will be the total files
+ total rows emitted from the xml
Added updated is the no:of docs . How do you know the number is not accurate?



 4.)
 And my last questions about Solr statistics/informations...

 === Is it possible to get informations (number of indexed documents, stored 
 values from documents etc.) from the current lucene index?
 === The admin webinterface shows 'numDocs' and 'maxDoc' in 
 'statistics/core'. Is 'numDocs' = number of indexed documents? What means 
 'maxDocs'?


 Thanks a lot!
 gisto
 --
 GMX Kostenlose Spiele: Einfach online spielen und Spaß haben mit Pastry 
 Passion!
 http://games.entertainment.gmx.net/de/entertainment/games/free/puzzle/6169196




-- 
--Noble Paul


Deadlock problem on searcher at warm up.

2008-10-24 Thread Jérôme Etévé
Hi everyone,

 I'm implementing a search component inherited from SearchComponent .

 This component has to build a data structure from the index. Like in
the SpellChecker, I trigger this building by giving a special argument
at query time  (from the process method) and I'm using the searcher I
get like this:

RefCountedSolrIndexSearcher search = rb.req.getCore()
.getSearcher();
...
search.decref();

I included this component at the end of the chain in my search handler.

What I'd like to do is to trigger this building for a first time at
solr startup so I don't need to artificially trigger it for a first
time.

I though it'd be ok to trigger this the very first time the process
method is called by doing something like that:

 private boolean firstTime= true ;

 public void process(ResponseBuilder rb) throws IOException {
if ( firstTime ){
firstTime = false ;
buildMyStuff(rb) ;
}
 }


The problem is that my method buildMyStuff hangs when calling
rb.req.getCore().getSearcher() ; ,
and I believe this is happening when the warm up queries are executed.

Furthermore, any regular queries on a solr instance like this would
hang and wait forever.

I there any way I can get around this problem, or is there a better
way to buildMyStuff a first time when solr is started up?

Cheers,

Jerome.

-- 
Jerome Eteve.

Chat with me live at http://www.eteve.net

[EMAIL PROTECTED]


Re: delta-import for XML files, Solr statistics

2008-10-24 Thread gistolero
Thanks for your very fast response :-)


  2.)
  The documentation from DataImportHandler describes the index update
 process for SQL databases only...
 
  My scenario:
  - My application creates, deletes and modifies files from /tmp/files
 every night.
  - delta-import / DataImportHandler should mirror _all_ this changes to
 my lucene index (= create, delete, update documents).
 The only Entityprocessor which supports delta is SqlEntityProcessor.
 The XPathEntityProcessor has not implemented it , because we do not
 know of a consistent way of finding deltas for XML. So ,
 unfortunately,no delta support for XML. But that said you can
 implement those methods in XPathEntityProcessor . The methods are
 explained in EntityProcessor.java. if you have questions specific to
 this I can help.Probably we can contribute it back
 
  === Is this possible with delta-import / DataImportHandler?
  === If not: Do you have any suggestions on how to do this?

Ok so, at the moment I have to do a full-import to update my index. What 
happens with (user) queries while full-import is running? Does Solr block this 
queries the import is finished? Which configuration options control this 
behavior? 



  My scenario:
  - /tmp/files contains 682 'myDoc_.*\.xml' XML files.
  - Each XML file contains 12 XML elements (e.g. titlefoo/title).
  - DataImportHandler transfer only 5 from this 12 elements to the lucene
 index.
 
 
  I don't understand the output from 'solr/dataimport' (= status):
 
  ###
  response
   ...
   lst name=statusMessages
   str name=Total Requests made to DataSource0/str
   str name=Total Rows Fetched1363/str
   str name=Total Documents Skipped0/str
   str name=Full Dump Started2008-10-24 13:19:03/str
   str name=
 Indexing completed. Added/Updated: 681 documents. Deleted 0
 documents.
   /str
   str name=Committed2008-10-24 13:19:05/str
   str name=Optimized2008-10-24 13:19:05/str
   str name=Time taken 0:0:2.648/str
   /lst
  ...
  /response
 
  === Why shows the Added/Updated counter 681 and not 682?
 
 Added updated is the no:of docs . How do you know the number is not
 accurate?


/tmp/files$ ls myDoc_*.xml | wc -l
682

But Added/Updated shows 681. Does this mean that one file has an XML error? 
But the statistic says Total Documents Skipped = 0?!

 

  4.)
  And my last questions about Solr statistics/informations...
 
  === Is it possible to get informations (number of indexed documents,
 stored values from documents etc.) from the current lucene index?
  === The admin webinterface shows 'numDocs' and 'maxDoc' in
 'statistics/core'. Is 'numDocs' = number of indexed documents? What means 
 'maxDocs'?

Do you have answers for this questions too?

Bye,
Simon
-- 
Der GMX SmartSurfer hilft bis zu 70% Ihrer Onlinekosten zu sparen! 
Ideal für Modem und ISDN: http://www.gmx.net/de/go/smartsurfer


Re: delta-import for XML files, Solr statistics

2008-10-24 Thread Akshay
On Fri, Oct 24, 2008 at 6:07 PM, [EMAIL PROTECTED] wrote:

 Thanks for your very fast response :-)


   2.)
   The documentation from DataImportHandler describes the index update
  process for SQL databases only...
  
   My scenario:
   - My application creates, deletes and modifies files from /tmp/files
  every night.
   - delta-import / DataImportHandler should mirror _all_ this changes
 to
  my lucene index (= create, delete, update documents).
  The only Entityprocessor which supports delta is SqlEntityProcessor.
  The XPathEntityProcessor has not implemented it , because we do not
  know of a consistent way of finding deltas for XML. So ,
  unfortunately,no delta support for XML. But that said you can
  implement those methods in XPathEntityProcessor . The methods are
  explained in EntityProcessor.java. if you have questions specific to
  this I can help.Probably we can contribute it back
  
   === Is this possible with delta-import / DataImportHandler?
   === If not: Do you have any suggestions on how to do this?

 Ok so, at the moment I have to do a full-import to update my index. What
 happens with (user) queries while full-import is running? Does Solr block
 this queries the import is finished? Which configuration options control
 this behavior?


No queries to SOLR  are not blocked during full import.





   My scenario:
   - /tmp/files contains 682 'myDoc_.*\.xml' XML files.
   - Each XML file contains 12 XML elements (e.g. titlefoo/title).
   - DataImportHandler transfer only 5 from this 12 elements to the lucene
  index.
  
  
   I don't understand the output from 'solr/dataimport' (= status):
  
   ###
   response
...
lst name=statusMessages
str name=Total Requests made to DataSource0/str
str name=Total Rows Fetched1363/str
str name=Total Documents Skipped0/str
str name=Full Dump Started2008-10-24 13:19:03/str
str name=
  Indexing completed. Added/Updated: 681 documents. Deleted 0
  documents.
/str
str name=Committed2008-10-24 13:19:05/str
str name=Optimized2008-10-24 13:19:05/str
str name=Time taken 0:0:2.648/str
/lst
   ...
   /response
  
   === Why shows the Added/Updated counter 681 and not 682?
 
  Added updated is the no:of docs . How do you know the number is not
  accurate?


 /tmp/files$ ls myDoc_*.xml | wc -l
 682

 But Added/Updated shows 681. Does this mean that one file has an XML
 error? But the statistic says Total Documents Skipped = 0?!


It might be the case that somewhere there is a extra line in one of the XML
files, a line like ?xml version=1.0 encoding=utf-8? or something.






   4.)
   And my last questions about Solr statistics/informations...
  
   === Is it possible to get informations (number of indexed documents,
  stored values from documents etc.) from the current lucene index?
   === The admin webinterface shows 'numDocs' and 'maxDoc' in
  'statistics/core'. Is 'numDocs' = number of indexed documents? What means
 'maxDocs'?

 Do you have answers for this questions too?

 Bye,
 Simon
 --
 Der GMX SmartSurfer hilft bis zu 70% Ihrer Onlinekosten zu sparen!
 Ideal für Modem und ISDN: http://www.gmx.net/de/go/smartsurfer




-- 
Regards,
Akshay Ukey.


Re: How to search a DataImportHandler solr index

2008-10-24 Thread Nick80

Hi,

below is a simplified copy of my data-config file:

dataConfig
dataSource type=JdbcDataSource driver=com.mysql.jdbc.Driver
url=jdbc:mysql://localhost/campaign user=root password=/
document name=campaigns
entity name=campaign query=SELECT * FROM campaigns WHERE 
deleted_at IS
NULL
field column=id name=id /
  field column=name name=campaign_name /

  entity name=banner query=SELECT * FROM banners WHERE
campaign_id=${campaign.id}
field name=banner_type column=banner_type /
entity name=size query=SELECT * FROM banner_sizes WHERE
banner_id=${banner.id}
field name=size column=size /
  /entity
  /entity
/entity
/document
/dataConfig

I have defined the following fields in schema.xml:

field name=id type=string indexed=true stored=true /
field name=campaign_name type=string indexed=true stored=true /
field name=banner_type type=string indexed=true stored=true
multiValued=true omitNorms=true termVectors=true /
field name=size type=string indexed=true stored=true
multiValued=true omitNorms=true termVectors=true /

Hope that makes it a bit clearer. Thanks.

Kind regards,

Nick
-- 
View this message in context: 
http://www.nabble.com/How-to-search-a-DataImportHandler-solr-index-tp20120698p20149960.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: One document inserted but nothing showing up ? SOLR 1.3

2008-10-24 Thread Jérôme Etévé
Hi there,

Are you sure you did a commit after your insertion ?

On Fri, Oct 24, 2008 at 8:11 AM, sunnyfr [EMAIL PROTECTED] wrote:

 Even that doesn't work,
 How can I check properly, I did insert one document but I can't get it back
 ???


 Feak, Todd wrote:

 Unless q=ALL is a special query I don't know about, the only reason you
 would get results is if ALL showed up in the default field of the single
 document that was inserted/updated.

 You could try a query of *:* instead. Don't forget to URL encode if you
 are doing this via URL.

 -Todd


 -Original Message-
 From: sunnyfr [mailto:[EMAIL PROTECTED]
 Sent: Thursday, October 23, 2008 9:17 AM
 To: solr-user@lucene.apache.org
 Subject: One document inserted but nothing showing up ? SOLR 1.3


 Hi

 Can somebody help me ?
 How can I see all my documents, I just did a full import :
 str name=
 Indexing completed. Added/Updated: 1 documents. Deleted 0 documents.
 /str

 and when I do :8180/solr/video/select/?q=ALL, I've no result ?
 response
 −
 lst name=responseHeader
 int name=status0/int
 int name=QTime0/int
 −
 lst name=params
 str name=qALL/str
 /lst
 /lst
 result name=response numFound=0 start=0/
 /response

 Thanks a lot,

 --
 View this message in context:
 http://www.nabble.com/One-document-inserted-but-nothing-showing-up---SOLR-1.3-tp20134357p20134357.html
 Sent from the Solr - User mailing list archive at Nabble.com.





 --
 View this message in context: 
 http://www.nabble.com/One-document-inserted-but-nothing-showing-up---SOLR-1.3-tp20134357p20145343.html
 Sent from the Solr - User mailing list archive at Nabble.com.





-- 
Jerome Eteve.

Chat with me live at http://www.eteve.net

[EMAIL PROTECTED]


Re: Deadlock problem on searcher at warm up.

2008-10-24 Thread Yonik Seeley
On Fri, Oct 24, 2008 at 8:21 AM, Jérôme Etévé [EMAIL PROTECTED] wrote:
 I though it'd be ok to trigger this the very first time the process
 method is called by doing something like that:

  private boolean firstTime= true ;

  public void process(ResponseBuilder rb) throws IOException {
if ( firstTime ){
firstTime = false ;
buildMyStuff(rb) ;
}
  }


 The problem is that my method buildMyStuff hangs when calling
 rb.req.getCore().getSearcher() ; ,
 and I believe this is happening when the warm up queries are executed.

getSearcher() can wait for a searcher to be registered.
getNewestSearcher() can be used from places like inform(), but if you
are already in process()
then the one you should use is the one bound to the request (the
SolrQueryRequest object) - rb.req.getSearcher()

-Yonik


customizing results in StandardQueryHandler

2008-10-24 Thread Manepalli, Kalyan
Hi,
In my usecase, I query a set of fields. Then based on the
results, I want to output a customized set of fields. Can I do this
without using a search component?
E:g. I query for fields f1, f2, f3, f4. Now based on some conditions, I
want to output just f1, f3, f4 (the list of final fields may vary). 

How do I rewrite the resultant xml optimally?
Any thoughts on this will be helpful

Thanks,
Kalyan


Re: Deadlock problem on searcher at warm up.

2008-10-24 Thread Jérôme Etévé
Great, it works now.

Thanks !

J

On Fri, Oct 24, 2008 at 4:45 PM, Yonik Seeley [EMAIL PROTECTED] wrote:
 On Fri, Oct 24, 2008 at 8:21 AM, Jérôme Etévé [EMAIL PROTECTED] wrote:
 I though it'd be ok to trigger this the very first time the process
 method is called by doing something like that:

  private boolean firstTime= true ;

  public void process(ResponseBuilder rb) throws IOException {
if ( firstTime ){
firstTime = false ;
buildMyStuff(rb) ;
}
  }


 The problem is that my method buildMyStuff hangs when calling
 rb.req.getCore().getSearcher() ; ,
 and I believe this is happening when the warm up queries are executed.

 getSearcher() can wait for a searcher to be registered.
 getNewestSearcher() can be used from places like inform(), but if you
 are already in process()
 then the one you should use is the one bound to the request (the
 SolrQueryRequest object) - rb.req.getSearcher()

 -Yonik




-- 
Jerome Eteve.

Chat with me live at http://www.eteve.net

[EMAIL PROTECTED]


Re: customizing results in StandardQueryHandler

2008-10-24 Thread Ryan McKinley

isn't this just: fl=f1,f3,f4  etc

or am I missing something?


On Oct 24, 2008, at 12:26 PM, Manepalli, Kalyan wrote:


Hi,
In my usecase, I query a set of fields. Then based on the
results, I want to output a customized set of fields. Can I do this
without using a search component?
E:g. I query for fields f1, f2, f3, f4. Now based on some  
conditions, I

want to output just f1, f3, f4 (the list of final fields may vary).

How do I rewrite the resultant xml optimally?
Any thoughts on this will be helpful

Thanks,
Kalyan




RE: customizing results in StandardQueryHandler

2008-10-24 Thread Manepalli, Kalyan
Ryan,
Actually, what I need is: I always query for a set of fields say
(f1, f2, f3 .. f6). Now once I get the results, based on some logic, I
need to generate the XML which is customized and contains only fields
say (f2, f3, and some new data). 
So the fl will always be (f1 ... f6)



Thanks,
Kalyan Manepalli

-Original Message-
From: Ryan McKinley [mailto:[EMAIL PROTECTED] 
Sent: Friday, October 24, 2008 1:25 PM
To: solr-user@lucene.apache.org
Subject: Re: customizing results in StandardQueryHandler

isn't this just: fl=f1,f3,f4  etc

or am I missing something?


On Oct 24, 2008, at 12:26 PM, Manepalli, Kalyan wrote:

 Hi,
   In my usecase, I query a set of fields. Then based on the
 results, I want to output a customized set of fields. Can I do this
 without using a search component?
 E:g. I query for fields f1, f2, f3, f4. Now based on some  
 conditions, I
 want to output just f1, f3, f4 (the list of final fields may vary).

 How do I rewrite the resultant xml optimally?
 Any thoughts on this will be helpful

 Thanks,
 Kalyan



RE: customizing results in StandardQueryHandler

2008-10-24 Thread Norskog, Lance
Ah!  This will let you post-process result sets with an XSL script:

http://wiki.apache.org/solr/XsltResponseWriter 

-Original Message-
From: Manepalli, Kalyan [mailto:[EMAIL PROTECTED] 
Sent: Friday, October 24, 2008 11:44 AM
To: solr-user@lucene.apache.org
Subject: RE: customizing results in StandardQueryHandler

Ryan,
Actually, what I need is: I always query for a set of fields say
(f1, f2, f3 .. f6). Now once I get the results, based on some logic, I
need to generate the XML which is customized and contains only fields
say (f2, f3, and some new data). 
So the fl will always be (f1 ... f6)



Thanks,
Kalyan Manepalli

-Original Message-
From: Ryan McKinley [mailto:[EMAIL PROTECTED]
Sent: Friday, October 24, 2008 1:25 PM
To: solr-user@lucene.apache.org
Subject: Re: customizing results in StandardQueryHandler

isn't this just: fl=f1,f3,f4  etc

or am I missing something?


On Oct 24, 2008, at 12:26 PM, Manepalli, Kalyan wrote:

 Hi,
   In my usecase, I query a set of fields. Then based on the
results, I 
 want to output a customized set of fields. Can I do this without using

 a search component?
 E:g. I query for fields f1, f2, f3, f4. Now based on some conditions, 
 I want to output just f1, f3, f4 (the list of final fields may vary).

 How do I rewrite the resultant xml optimally?
 Any thoughts on this will be helpful

 Thanks,
 Kalyan



Re: customizing results in StandardQueryHandler

2008-10-24 Thread Ryan McKinley

I'm still not following...

Does the logic depend on the result?  That is, are you asking for doc  
A to have f1,f2 and doc B to have f1,f4?  If that is your question,  
then no -- as is, the ResponseWriter will write the same fields for  
every document.


When you say (f2, f3, and some new data) -- where does the new  
data come from?  Are you trying to augment the results of a document  
with something else?  If so, take a look at the local solr patch in:

 https://issues.apache.org/jira/browse/LUCENE-1387
that adds a distance calculation to each matched document.



On Oct 24, 2008, at 2:43 PM, Manepalli, Kalyan wrote:


Ryan,
Actually, what I need is: I always query for a set of fields say
(f1, f2, f3 .. f6). Now once I get the results, based on some logic, I
need to generate the XML which is customized and contains only fields
say (f2, f3, and some new data).
So the fl will always be (f1 ... f6)



Thanks,
Kalyan Manepalli

-Original Message-
From: Ryan McKinley [mailto:[EMAIL PROTECTED]
Sent: Friday, October 24, 2008 1:25 PM
To: solr-user@lucene.apache.org
Subject: Re: customizing results in StandardQueryHandler

isn't this just: fl=f1,f3,f4  etc

or am I missing something?


On Oct 24, 2008, at 12:26 PM, Manepalli, Kalyan wrote:


Hi,
In my usecase, I query a set of fields. Then based on the
results, I want to output a customized set of fields. Can I do this
without using a search component?
E:g. I query for fields f1, f2, f3, f4. Now based on some
conditions, I
want to output just f1, f3, f4 (the list of final fields may vary).

How do I rewrite the resultant xml optimally?
Any thoughts on this will be helpful

Thanks,
Kalyan






Re: timeouts

2008-10-24 Thread Grant Ingersoll
What is happening at the time of the freeze?  Are you indexing?   
Searching?  For either of these options, what are the operations that  
are happening?  Is Tomcat and the Solr admin accessible while this  
happens?  Are there any exceptions in the logs?  How much memory are  
you giving it versus MySQL?  I assume you're hitting both, right?  Do  
you have enough memory to do that?  Typically DBs like as much mem as  
they can get, as does Solr (for the most part).


That being said, 100K records is next to nothing, so I doubt Solr is  
taking that much memory.  Please describe a bit more what you are  
doing and seeing.



On Oct 24, 2008, at 8:58 AM, Simon Collins wrote:


Hi

We’re running solr on a win 2k3 box under tomcat with about 100,000  
records. When doing large updates of records via solr sharp, solr  
completely freezes and doesn’t come back until we restart tomcat.


This has only started happening since putting mysql on the same box  
(as a source of the data to update from).


Are there any known issues with running solr and mysql on the same  
box? When it’s frozen, the cpu usage is around 1-2% not exactly out  
of resources!


Am i best using something else instead of tomcat? We’re still  
trialling solr (presently, used for our main site search www.shoe-shop.com 
 and search and navigation for our microsites ). It’s an excellent  
search product, but I don’t want to fork out on new hardware for it  
just yet – until i know how more about the performance and which  
environment i’m best to go for (win/linux).


If anyone has any suggestions/needs more info, i’d be extremely  
grateful.


Thanks
Simon
Simon Collins
Systems Analyst

Telephone:  01904 606 867
Fax Number: 01904 528 791

shoe-shop.com ltd
Catherine House
Northminster Business Park
Upper Poppleton, YORK
YO26 6QU

www.shoe-shop.com


This message (and any associated files) is intended only for the use  
of the individual or entity to which it is addressed and may contain  
information that is confidential, subject to copyright or  
constitutes a trade secret. If you are not the intended recipient  
you are hereby notified that any dissemination, copying or  
distribution of this message, or files associated with this message,  
is strictly prohibited. If you have received this message in error,  
please notify us immediately by replying to the message and deleting  
it from your computer. Messages sent to and from us may be monitored.


Internet communications cannot be guaranteed to be secure or error- 
free as information could be intercepted, corrupted, lost,  
destroyed, arrive late or incomplete, or contain viruses. Therefore,  
we do not accept responsibility for any errors or omissions that are  
present in this message, or any attachment, that have arisen as a  
result of e-mail transmission. If verification is required, please  
request a hard-copy version. Any views or opinions presented are  
solely those of the author and do not necessarily represent those of  
the company. (PAVD001)


Shoe-shop.com Limited is a company registered in England and Wales  
with company number 03817232. Vat Registration GB 734 256 241.  
Registered Office Catherine House, Northminster Business Park, Upper  
Poppleton, YORK, YO26 6QU.













This message has been scanned for malware by SurfControl plc. 
www.surfcontrol.com



--
Grant Ingersoll
Lucene Boot Camp Training Nov. 3-4, 2008, ApacheCon US New Orleans.
http://www.lucenebootcamp.com


Lucene Helpful Hints:
http://wiki.apache.org/lucene-java/BasicsOfPerformance
http://wiki.apache.org/lucene-java/LuceneFAQ











Entity extraction?

2008-10-24 Thread Charlie Jackson
During a recent sales pitch to my company by FAST, they mentioned entity
extraction. I'd never heard of it before, but they described it as
basically recognizing people/places/things in documents being indexed
and then being able to do faceting on this data at query time. Does
anything like this already exist in SOLR? If not, I'm not opposed to
developing it myself, but I could use some pointers on where to start.

 

Thanks,

- Charlie



RE: customizing results in StandardQueryHandler

2008-10-24 Thread Manepalli, Kalyan
Populating the field at the index time is a good option. Is there any
other way to do it in search time

Thanks,
Kalyan Manepalli

-Original Message-
From: Ryan McKinley [mailto:[EMAIL PROTECTED] 
Sent: Friday, October 24, 2008 2:14 PM
To: solr-user@lucene.apache.org
Subject: Re: customizing results in StandardQueryHandler

I'm still not following...

Does the logic depend on the result?  That is, are you asking for doc  
A to have f1,f2 and doc B to have f1,f4?  If that is your question,  
then no -- as is, the ResponseWriter will write the same fields for  
every document.

When you say (f2, f3, and some new data) -- where does the new  
data come from?  Are you trying to augment the results of a document  
with something else?  If so, take a look at the local solr patch in:
  https://issues.apache.org/jira/browse/LUCENE-1387
that adds a distance calculation to each matched document.



On Oct 24, 2008, at 2:43 PM, Manepalli, Kalyan wrote:

 Ryan,
   Actually, what I need is: I always query for a set of fields say
 (f1, f2, f3 .. f6). Now once I get the results, based on some logic, I
 need to generate the XML which is customized and contains only fields
 say (f2, f3, and some new data).
 So the fl will always be (f1 ... f6)



 Thanks,
 Kalyan Manepalli

 -Original Message-
 From: Ryan McKinley [mailto:[EMAIL PROTECTED]
 Sent: Friday, October 24, 2008 1:25 PM
 To: solr-user@lucene.apache.org
 Subject: Re: customizing results in StandardQueryHandler

 isn't this just: fl=f1,f3,f4  etc

 or am I missing something?


 On Oct 24, 2008, at 12:26 PM, Manepalli, Kalyan wrote:

 Hi,
  In my usecase, I query a set of fields. Then based on the
 results, I want to output a customized set of fields. Can I do this
 without using a search component?
 E:g. I query for fields f1, f2, f3, f4. Now based on some
 conditions, I
 want to output just f1, f3, f4 (the list of final fields may vary).

 How do I rewrite the resultant xml optimally?
 Any thoughts on this will be helpful

 Thanks,
 Kalyan




Solr 1.3 stack overflow when accessing solr/admin page

2008-10-24 Thread Dadasheva, Olga
Hi,

I  run Solr 1.3 in Weblogic 10.3 Java 6;
I have a single core application deployed to the same server as Solr
1.3. 
I am not using solrj.

I do not have any problems with search and update - they work fine. 

However I am getting stack overflow when I try to access solr/admin
page.


java.lang.StackOverflowError javax.servlet.ServletException:
java.lang.StackOverflowError at
weblogic.servlet.internal.RequestDispatcherImpl.forward(RequestDispatche
rImpl.java:266) at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.j
ava:273) at
weblogic.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:
42) at
weblogic.servlet.internal.RequestDispatcherImpl.invokeServlet(RequestDis
patcherImpl.java:503) at 
...

I am pretty sure it is my configuration, and I would appreciate any
pointers at what I should be checking.


Thank you very much,

-Olga


Re: Solr 1.3 stack overflow when accessing solr/admin page

2008-10-24 Thread Ryan McKinley
are you running the packaged .war directly?  or something custom?  Did  
it ever work?


Is anyone else running successfully on weblogic?


On Oct 24, 2008, at 5:10 PM, Dadasheva, Olga wrote:


Hi,

I  run Solr 1.3 in Weblogic 10.3 Java 6;
I have a single core application deployed to the same server as Solr
1.3.
I am not using solrj.

I do not have any problems with search and update - they work fine.

However I am getting stack overflow when I try to access solr/admin
page.


java.lang.StackOverflowError javax.servlet.ServletException:
java.lang.StackOverflowError at
weblogic 
.servlet.internal.RequestDispatcherImpl.forward(RequestDispatche

rImpl.java:266) at
org 
.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.j

ava:273) at
weblogic 
.servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:

42) at
weblogic 
.servlet.internal.RequestDispatcherImpl.invokeServlet(RequestDis

patcherImpl.java:503) at
...

I am pretty sure it is my configuration, and I would appreciate any
pointers at what I should be checking.


Thank you very much,

-Olga




Re: Entity extraction?

2008-10-24 Thread Rogerio Pereira
You can find more about this topic in this book availabe at amazon:
http://www.amazon.com/Building-Search-Applications-Lucene-Lingpipe/dp/0615204252/

2008/10/24 Rafael Rossini [EMAIL PROTECTED]

 Solr can do a simple facet seach like FAST, but the entity extraction
 demands other tecnologies. I do not know how FAST does it but at the
 company
 I´m working on (www.cortex-intelligence.com), we use a mix of statistical
 and language-specific tasks to recognize and categorize entities in the
 text. Ling Pipe is another tool (free) that does that too. In case you
 would
 like to see a simple demo: http://www.cortex-intelligence.com/tech/

 Rossini


 On Fri, Oct 24, 2008 at 6:18 PM, Charlie Jackson 
 [EMAIL PROTECTED]
  wrote:

  During a recent sales pitch to my company by FAST, they mentioned entity
  extraction. I'd never heard of it before, but they described it as
  basically recognizing people/places/things in documents being indexed
  and then being able to do faceting on this data at query time. Does
  anything like this already exist in SOLR? If not, I'm not opposed to
  developing it myself, but I could use some pointers on where to start.
 
 
 
  Thanks,
 
  - Charlie
 
 




-- 
Regards,

Rogério (_rogerio_)

[Blog: http://faces.eti.br]  [Sandbox: http://bmobile.dyndns.org]  [Twitter:
http://twitter.com/ararog]

Faça a diferença! Ajude o seu país a crescer, não retenha conhecimento,
distribua e aprenda mais.
(http://faces.eti.br/2006/10/30/conhecimento-e-amadurecimento)


Re: Entity extraction?

2008-10-24 Thread Ryan McKinley

This is not something solr does currently...

It sounds like something that should be added to Mahout:
http://lucene.apache.org/mahout/


On Oct 24, 2008, at 4:18 PM, Charlie Jackson wrote:

During a recent sales pitch to my company by FAST, they mentioned  
entity

extraction. I'd never heard of it before, but they described it as
basically recognizing people/places/things in documents being indexed
and then being able to do faceting on this data at query time. Does
anything like this already exist in SOLR? If not, I'm not opposed to
developing it myself, but I could use some pointers on where to start.



Thanks,

- Charlie





RE: Entity extraction?

2008-10-24 Thread Charlie Jackson
Thanks for the replies, guys, that gives me a good place to start looking. 

- Charlie

-Original Message-
From: Rogerio Pereira [mailto:[EMAIL PROTECTED] 
Sent: Friday, October 24, 2008 5:14 PM
To: solr-user@lucene.apache.org
Subject: Re: Entity extraction?

You can find more about this topic in this book availabe at amazon:
http://www.amazon.com/Building-Search-Applications-Lucene-Lingpipe/dp/0615204252/

2008/10/24 Rafael Rossini [EMAIL PROTECTED]

 Solr can do a simple facet seach like FAST, but the entity extraction
 demands other tecnologies. I do not know how FAST does it but at the
 company
 I´m working on (www.cortex-intelligence.com), we use a mix of statistical
 and language-specific tasks to recognize and categorize entities in the
 text. Ling Pipe is another tool (free) that does that too. In case you
 would
 like to see a simple demo: http://www.cortex-intelligence.com/tech/

 Rossini


 On Fri, Oct 24, 2008 at 6:18 PM, Charlie Jackson 
 [EMAIL PROTECTED]
  wrote:

  During a recent sales pitch to my company by FAST, they mentioned entity
  extraction. I'd never heard of it before, but they described it as
  basically recognizing people/places/things in documents being indexed
  and then being able to do faceting on this data at query time. Does
  anything like this already exist in SOLR? If not, I'm not opposed to
  developing it myself, but I could use some pointers on where to start.
 
 
 
  Thanks,
 
  - Charlie
 
 




-- 
Regards,

Rogério (_rogerio_)

[Blog: http://faces.eti.br]  [Sandbox: http://bmobile.dyndns.org]  [Twitter:
http://twitter.com/ararog]

Faça a diferença! Ajude o seu país a crescer, não retenha conhecimento,
distribua e aprenda mais.
(http://faces.eti.br/2006/10/30/conhecimento-e-amadurecimento)



Re: Entity extraction?

2008-10-24 Thread Rogerio Pereira
I agree Ryan and I would like see a completly integration between solr,
nutch, tika and mahout in the future.

2008/10/24 Ryan McKinley [EMAIL PROTECTED]

 This is not something solr does currently...

 It sounds like something that should be added to Mahout:
 http://lucene.apache.org/mahout/



 On Oct 24, 2008, at 4:18 PM, Charlie Jackson wrote:

  During a recent sales pitch to my company by FAST, they mentioned entity
 extraction. I'd never heard of it before, but they described it as
 basically recognizing people/places/things in documents being indexed
 and then being able to do faceting on this data at query time. Does
 anything like this already exist in SOLR? If not, I'm not opposed to
 developing it myself, but I could use some pointers on where to start.



 Thanks,

 - Charlie





-- 
Regards,

Rogério (_rogerio_)

[Blog: http://faces.eti.br]  [Sandbox: http://bmobile.dyndns.org]  [Twitter:
http://twitter.com/ararog]

Faça a diferença! Ajude o seu país a crescer, não retenha conhecimento,
distribua e aprenda mais.
(http://faces.eti.br/2006/10/30/conhecimento-e-amadurecimento)


Lucene project subprojects news RSS feed?

2008-10-24 Thread David Smiley @MITRE.org

On the main lucene web page: http://lucene.apache.org/index.html
There is a list of news items spanning all the lucene subprojects.  Does
anyone know if there is an RSS feed or announce mailing list that has this
information?

~ David Smiley
-- 
View this message in context: 
http://www.nabble.com/Lucene-project---subprojects-news-RSS-feed--tp20158991p20158991.html
Sent from the Solr - User mailing list archive at Nabble.com.



RE: Solr 1.3 stack overflow when accessing solr/admin page

2008-10-24 Thread Dadasheva, Olga
I am running the packaged war.
 
It used to work 1 year ago when I briefly looked into using Solr instead
of Lucene - at that time I just built the trunk and it worked from the
first plug (Solr 1.3 dev around Sep 9th 2007). 


-Original Message-
From: Ryan McKinley [mailto:[EMAIL PROTECTED] 
Sent: Friday, October 24, 2008 5:28 PM
To: solr-user@lucene.apache.org
Subject: Re: Solr 1.3 stack overflow when accessing solr/admin page

are you running the packaged .war directly?  or something custom?  Did
it ever work?

Is anyone else running successfully on weblogic?


On Oct 24, 2008, at 5:10 PM, Dadasheva, Olga wrote:

 Hi,

 I  run Solr 1.3 in Weblogic 10.3 Java 6; I have a single core 
 application deployed to the same server as Solr 1.3.
 I am not using solrj.

 I do not have any problems with search and update - they work fine.

 However I am getting stack overflow when I try to access solr/admin 
 page.


 java.lang.StackOverflowError javax.servlet.ServletException:
 java.lang.StackOverflowError at
 weblogic
 .servlet.internal.RequestDispatcherImpl.forward(RequestDispatche
 rImpl.java:266) at
 org
 .apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.j
 ava:273) at
 weblogic
 .servlet.internal.FilterChainImpl.doFilter(FilterChainImpl.java:
 42) at
 weblogic
 .servlet.internal.RequestDispatcherImpl.invokeServlet(RequestDis
 patcherImpl.java:503) at
 ...

 I am pretty sure it is my configuration, and I would appreciate any 
 pointers at what I should be checking.


 Thank you very much,

 -Olga



Index partitioning

2008-10-24 Thread mahendra mahendra
Hi,
 
I want to partition my index based on category information. Also, while 
indexing I want to store particular category data to corresponding index 
partition. In the same way I need to search for category information on 
corresponding partition..
 
I found some information on wiki link 
http://wiki.apache.org/solr/IndexPartitioning. But it couldn't help much in 
implementation.
 
Could some one give example with how I can configure, index data 
into particular partitions and search on particular partition.
 
Would appreciate for any help!!


Thanks  Regards,
Mahendra