Re: Storing MYSQL DATETIME field in solr as String

2013-12-16 Thread manju16832003
OK here are the scenarios I tried.

*Scenario - 1: *

dih.xml (aka data-config.xml)

entity dataSource=solr name=listing query=...
transformer=DateFormatTransformer

field name=quot;publish_datequot; column=quot;publish_datequot;
xpath=quot;/RDF/item/datequot;
dateTimeFormat=quot;lt;b-MM-dd'T'HH:mm:ss* locale=en/

schema.xml
field name=publish_date type=date indexed=true stored=true 
multiValued=false default=NOW /

It does not work. Throws an following exception 
WARN  org.apache.solr.handler.dataimport.DateFormatTransformer  – Could not
parse a Date field
java.text.ParseException: Unparseable date: 2013-12-05 15:40:03.0

*Scenario - 2*
dih.xml (aka data-config.xml)

entity dataSource=solr name=listing query=...
transformer=DateFormatTransformer

field name=quot;publish_datequot; column=quot;publish_datequot;
xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd
HH:mm:ss* locale=en/ 

schema.xml
field name=publish_date type=date indexed=true stored=true 
multiValued=false default=NOW /

Output is 
date name=publish_date2012-12-05T07:38:27Z/date

*Scenario - 3*
dih.xml (aka data-config.xml)

entity dataSource=solr name=listing query=...
transformer=DateFormatTransformer

field name=quot;publish_datequot; column=quot;publish_datequot;
xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd
HH:mm:ss* locale=en/ 

schema.xml
field name=publish_date type=string indexed=true stored=true 
multiValued=false default=NOW / 

Output is 
  str name=publish_dateThu Dec 05 15:38:27 MYT 2013/str

My expectation was *2012-12-05 07:38:27* (MySQL DateTime format)



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Storing-MYSQL-DATETIME-field-in-solr-as-String-tp4106836p4106854.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Solr Profiler

2013-12-16 Thread manju16832003
Hi NewRelic is good enough to monitor the Solr.

Are you using Solarium or SolrJ client to connect to Solr?.

We have used Solarium and able to monitor each calls and gather most of the
info.





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-Profiler-tp4106231p4106859.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Backwards compatibility issue

2013-12-16 Thread Karel Vervaeke
Thanks for your feedback. Upgrading SolrJ is going to be a bit difficult
(because it's tied in with a framework which is only partially owned by
us), so we'll just have to find another way to deal with it.

Karel


On Fri, Dec 13, 2013 at 7:28 PM, Shawn Heisey s...@elyograg.org wrote:

 On 12/13/2013 7:40 AM, Karel Vervaeke wrote:
  IIUC A Solrj 4.4 client should be able to speak to a Solr 4.6 cloud
  instance.
  Assuming this is correct there's an issue because of a change in
  clusterstate.json:
 
  In a 4.4 cloud server clusterstate.json would contain `router:
 implicit`
  In a 4.6 cloud server clusterstate.json contains `router:
  {name:implicit}`.
 
  Because of this a 4.4 solrj client will look up a router (in
  DocRouter.java) using the key `{name:implicit}` instead of just
  `implicit`.
  Is there a way around this that does not involve upgrading the solrj
 client?

 Although the committers do try really really hard to maintain
 cross-version compatibility between SolrJ and Solr versions, sometimes
 changes must be made that break that compatibility, either to support
 critical new functionality or because of bugs or situations that were
 not anticipated.  SolrCloud technology is relatively new, so things are
 changing quickly to accommodate new functionality and fix stability
 problems.

 SOLR-4221 made this change (released in version 4.5.0) so that more than
 one property could be assigned to 'router' because more information is
 sometimes needed.

 https://issues.apache.org/jira/browse/SOLR-4221

 One of the commits made on this issue was to allow a newer SolrJ to talk
 to an older cluster with the previous clusterstate router format.  If
 for some reason you need to run different SolrJ and Solr versions, it is
 almost always better to have a newer SolrJ than Sol

 Thanks,
 Shawn




Heap size and Solr 4.3

2013-12-16 Thread Marcello Lorenzi

Hi All,
we have deployed on our production environment a new Solr 4.3 instance 
(2 nodes with SolrCloud) but this morning one node gone on outofmemory 
status and we have noticed that the JVM uses a lot of Old Gen space 
during the normal lifecycle.


What are the items that improve this high usage of Heap?

Thanks,
Marcello


Re: Poor performance on distributed search

2013-12-16 Thread ku3ia
Yonik Seeley-2-2 wrote
 On Wed, Dec 28, 2011 at 5:47 AM, ku3ia lt;

 demesg@

 gt; wrote:
 So, based on p.2) and on my previous researches, I conclude, that the
 more
 documents I want to retrieve, the slower is search and main problem is
 the
 cycle in writeDocs method. Am I right? Can you advice something in this
 situation?
 
 For the first phase in a distributed search, Solr must return the top
 N ids (in your case 200).  It currently does this by loading stored
 fields, which is slow.  A better approach is to store the id field
 as a column stride field.
 
 https://issues.apache.org/jira/browse/SOLR-2753
 
 -Yonik
 http://www.lucidimagination.com

Hi all again.
I have upgraded Solr version to 4.6.0 and want to test DocValues fields in
action.
My schema is: 4 shards of 48M of docs each. All shards are on the same
machine. I have turned on DV for id field in my schema.xml:

field name=RecordID type=uid indexed=true stored=true
required=true docValues=true/
...
uniqueKeyRecordID/uniqueKey
...
fieldType name=uid class=solr.StrField sortMissingLast=true
docValuesFormat=Disk/

I had created 30 queries like,
q=(a OR b) OR (c OR d e OR f), where a through f are random words

These queries were curled each other, not concurrently. After that I turned
off doc values, rebuild all four shards and runs queries again.

At results, a haven't seen any difference between stored ID field and
docValues and one more moment, when I set stored=false and docValues=true
for id field I had an error when running distributed search:

ERROR - 2013-12-12 16:59:45.508; org.apache.solr.common.SolrException;
java.lang.NullPointerException
at
org.apache.solr.handler.component.QueryComponent.mergeIds(QueryComponent.java:1004)
at
org.apache.solr.handler.component.QueryComponent.handleRegularResponses(QueryComponent.java:750)
at
org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:729)
at
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:312)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1863)
at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:710)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:413)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)
at
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
at
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
at
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
at
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
at
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
at
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
at
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
at
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
at
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
at org.eclipse.jetty.server.Server.handle(Server.java:370)
at
org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
at
org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:949)
at
org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1011)
at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:644)
at
org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
at
org.eclipse.jetty.server.AsyncHttpConnection.handle(AsyncHttpConnection.java:82)
at
org.eclipse.jetty.io.nio.SelectChannelEndPoint.handle(SelectChannelEndPoint.java:668)
at
org.eclipse.jetty.io.nio.SelectChannelEndPoint$1.run(SelectChannelEndPoint.java:52)
at
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
at
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
at java.lang.Thread.run(Thread.java:744)

This error is on QueryComponent on merge Ids:
resultIds.put(shardDoc.id.toString(), shardDoc);

So seems it try to use stored id, not docvalue? I tried different codecs:
Direct, Disk, Memory, Lucene45. Results are nearby the 

OutOfMemoryError in RamUsageEstimator in solr 4.6

2013-12-16 Thread Torben Greulich
Hi,
we get a OutOfMemoryError in RamUsageEstimator and are a little bit
confused about the error.
We are using solr 4.6 and are confused about the Lucene42DocValuesProducer.
We checked current solr code and found that Lucene42NormsFormat will be
returned as NormFormat in Lucene46Codec and so the Lucene42DocValuesProducer
will be returned. Is there any special reason not to have a higher version
of NormsFormat?

This is our StackTrace:

SolrException|null:java.lang.OutOfMemoryError: Java heap space

at
org.apache.lucene.util.RamUsageEstimator$IdentityHashSet.allocateBuffers(RamUsageEstimator.java:762)

at
org.apache.lucene.util.RamUsageEstimator$IdentityHashSet.expandAndRehash(RamUsageEstimator.java:736)

at
org.apache.lucene.util.RamUsageEstimator$IdentityHashSet.add(RamUsageEstimator.java:678)

at
org.apache.lucene.util.RamUsageEstimator.measureObjectSize(RamUsageEstimator.java:437)

at
org.apache.lucene.util.RamUsageEstimator.sizeOf(RamUsageEstimator.java:350)

at
org.apache.lucene.codecs.lucene42.Lucene42DocValuesProducer.ramBytesUsed(Lucene42DocValuesProducer.java:194)

at
org.apache.lucene.index.SegmentCoreReaders.ramBytesUsed(SegmentCoreReaders.java:195)

at
org.apache.lucene.index.SegmentReader.ramBytesUsed(SegmentReader.java:558)

at
org.apache.solr.handler.admin.LukeRequestHandler.getIndexHeapUsed(LukeRequestHandler.java:579)

at
org.apache.solr.handler.admin.LukeRequestHandler.getIndexInfo(LukeRequestHandler.java:558)

at
org.apache.solr.handler.admin.CoreAdminHandler.getCoreStatus(CoreAdminHandler.java:1044)

at
org.apache.solr.handler.admin.CoreAdminHandler.handleStatusAction(CoreAdminHandler.java:685)

at
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestBody(CoreAdminHandler.java:167)

at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)

at
org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:662)

at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:248)

at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)

at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243)

at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210)

...

Please contact me, if there are further questions.

Best regards Torben Greulich

-- 
Torben Greulich · Entwickler
Backend Development
shopping24 internet group

Wandsbeker Straße 3-7 · 22172 Hamburg
Telefon: +49 (0) 40 6461  · Fax: +49 (0) 40 6461 7879
torben.greul...@s24.com · www.s24.com
AG Hamburg HRB 63371
vertreten durch Dr. Björn Schäfers und Martin Mildner


How to boost documents ?

2013-12-16 Thread Anca Kopetz



Hi,

How to boost documents that contain all search terms in several of its fields ?

Below you cand find a simplified example : 

The query with Min should match:
q=beautiful Christmas treemm=2qf=title^12 description^2

There are two offers that match the query : 
offer1 {title:Christmas tree, description:a joy for children}

offer2 {title:Christmas tree, description:beautiful for holidays}}

The first offer ranks before the second, despite of the fact that the second one contains all the search terms. I tried to play with the boosts of
qf, but the results vary a lot.

Is there a way to add a boost on all search fields, the same way we do with pf on one field :
pf=title:2^3.0 ?

Thank you,
Anca
-- 






Anca Kopetz Software engineer

E anca.kop...@kelkoo.comY!Messenger kelkooancak
T 33 (0)4 56 09 07 55 

A 4/6 Rue des Meridiens 38130 Echirolles











Kelkoo SAS
Société par Actions Simplifiée
Au capital de € 4.168.964,30
Siège social : 8, rue du Sentier 75002 Paris
425 093 069 RCS Paris

Ce message et les pièces jointes sont confidentiels et établis à l'attention exclusive de leurs destinataires. Si vous n'êtes pas le destinataire de ce message, merci de le détruire et d'en avertir l'expéditeur.





indexing from bowser

2013-12-16 Thread Nutan
how to index pdf,doc files from browser?

this query is used for indexing :
curl
http://localhost:8080/solr/document/update/extract?literal.id=12commit=true;
-Fmyfile=@C:\solr\document\src\test1\Coding.pdf

but i need to index from browser, as we do for delete:
http://localhost:8080/solr/document/update?stream.body=deletequeryid:3/query/deletecommit=true

When i try to index using this:
http://localhost:8080/solr/document/update/extract?literal.id=12commit=true;
-Fmyfile=@C:\solr\document\src\test1\Coding.pdf

the document does not get indexed.

Is there any query to index from browser,using html url??



--
View this message in context: 
http://lucene.472066.n3.nabble.com/indexing-from-bowser-tp4106889.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: indexing from bowser

2013-12-16 Thread Koji Sekiguchi

Hi,

(13/12/16 19:46), Nutan wrote:

how to index pdf,doc files from browser?


I think you can index from browser.

If you said that


this query is used for indexing :
curl
http://localhost:8080/solr/document/update/extract?literal.id=12commit=true;
-Fmyfile=@C:\solr\document\src\test1\Coding.pdf


curl works for you but


When i try to index using this:
http://localhost:8080/solr/document/update/extract?literal.id=12commit=true;
-Fmyfile=@C:\solr\document\src\test1\Coding.pdf

the document does not get indexed.


browser doesn't work for you, why don't you look into Solr log and
compare the logs between when you using curl and browser?

koji
--
http://soleami.com/blog/automatically-acquiring-synonym-knowledge-from-wikipedia.html


Re: indexing from bowser

2013-12-16 Thread Gora Mohanty
On 16 December 2013 16:30, Koji Sekiguchi k...@r.email.ne.jp wrote:

 Hi,

 (13/12/16 19:46), Nutan wrote:

 how to index pdf,doc files from browser?


 I think you can index from browser.

 If you said that

 this query is used for indexing :
 curl
 http://localhost:8080/solr/document/update/extract?literal.id=12commit=true;
 -Fmyfile=@C:\solr\document\src\test1\Coding.pdf


 curl works for you but

 When i try to index using this:
 http://localhost:8080/solr/document/update/extract?literal.id=12commit=true;
 -Fmyfile=@C:\solr\document\src\test1\Coding.pdf

 the document does not get indexed.


 browser doesn't work for you, why don't you look into Solr log and
 compare the logs between when you using curl and browser?

No, this will not work by loading the URL into a browser, as with the
-F option, curl is emulating a POST for a form filled in by the user.
One must use curl, or a similar tool, for this.

Regards,
Gora


High load from solr 4.5

2013-12-16 Thread Gastone Penzo
I tried to install some versions of solr, from solr 4 to solr 4.6 (the last
release) in my production environment. The configuration of the machine is

- Debian 7 64bit
- java opendjk 7
- CPU 8 core
- 16Gb ram

but i notice that the machine load grows if i install the version 4.4 or
above. Why?
the configurations (solrconfig and schema) are the same. The indexes are
the same.
I don't use solrcloud. I run only one instance and i give to solr 6 gb or
ram.
The logs are disable.

What the difference? thank you




-- 

*Gastone Penzo*


Re: indexing from bowser

2013-12-16 Thread Nutan
ok thanks,
but is there any other way where -F is not used?

I am creating a api in vc++ and to link to solr i am using libcurl,for this
to work the string is the url,
eg: 
curl_easy_setopt(curl,
CURLOPT_URL,http://localhost:8080/solr/document/select?q=*%3A*wt=jsonindent=truefl=id;);
 

so for indexing i need a url.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/indexing-from-bowser-tp4106889p4106900.html
Sent from the Solr - User mailing list archive at Nabble.com.


RE: SOLR DIH - Sub Entity with different datasource not working

2013-12-16 Thread Lokn
Thanks for the reply.

The error shows it is not able to execute the query.
In my case, if you see my config file I am joining the entities between two
different datasources..

i.e., Entity1 - Datasource1
 --Subentity - DataSource2

My doubt is, can we join the entities in two different datasources. Is this
fine..





--
View this message in context: 
http://lucene.472066.n3.nabble.com/SOLR-DIH-Sub-Entity-with-different-datasource-not-working-tp4106550p4106906.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SOLR DIH - Sub Entity with different datasource not working

2013-12-16 Thread Karan Ahuja
hi

i dont know if this can be done

but to avoid this you can create a new table with the results and index
that new table :)
you can then delete the table as well after indexing ...

:)

tc cheers
karan



On Mon, Dec 16, 2013 at 5:42 PM, Lokn nlokesh...@gmail.com wrote:

 Thanks for the reply.

 The error shows it is not able to execute the query.
 In my case, if you see my config file I am joining the entities between two
 different datasources..

 i.e., Entity1 - Datasource1
  --Subentity - DataSource2

 My doubt is, can we join the entities in two different datasources. Is this
 fine..





 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/SOLR-DIH-Sub-Entity-with-different-datasource-not-working-tp4106550p4106906.html
 Sent from the Solr - User mailing list archive at Nabble.com.



Re: How to boost documents ?

2013-12-16 Thread Ahmet Arslan
Hi Anca,

Can you try following URL?



q=beautiful Christmas treemm=2qf=title^12 
description^2defType=dismaxbf=map(query($qq),0,0,0,100.0)qq={!dismax 
qf='title description' mm=100%}beautiful Christmas tree

Modified from Jan's solution. See his original post [1] to a similar discussion.
[1] http://search-lucene.com/m/nK6t9j1fuc2





On Monday, December 16, 2013 12:19 PM, Anca Kopetz anca.kop...@kelkoo.com 
wrote:

Hi,

How to boost documents that contain all search terms in several of its fields ?

Below you cand find a simplified example : 

The query with Min should match:
q=beautiful Christmas treemm=2qf=title^12 description^2

There are two offers that match the query : 
offer1 {title:Christmas tree, description:a joy for children}

offer2 {title:Christmas tree, description:beautiful for holidays}}

The first offer ranks before the second, despite of the fact that the second 
one contains all the search terms. I tried to play with the boosts of qf, but 
the results vary a lot.

Is there a way to add a boost on all search fields, the same way we do with pf 
on one field : pf=title:2^3.0 ?

Thank you,
Anca

-- 

Anca Kopetz Software engineer

eanca.kop...@kelkoo.com  Y!Messenger kelkooancak
T +33 (0)4 56 09 07 55    
A 4/6 Rue des Meridiens 38130 Echirolles   

Kelkoo SAS
Société par Actions Simplifiée
Au capital de € 4.168.964,30
Siège social : 8, rue du Sentier 75002 Paris
425 093 069 RCS Paris

Ce message et les pièces jointes sont confidentiels et établis à l'attention 
exclusive de leurs destinataires. Si vous n'êtes pas le destinataire de ce 
message, merci de le détruire et d'en avertir l'expéditeur.


Re: uniquekey generation in solr

2013-12-16 Thread gpssolr2020

one of the field used in unique key generation is in date format and other
are in string format. So when we remove this date field ,it is working fine.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/uniquekey-generation-in-solr-tp4106766p4106919.html
Sent from the Solr - User mailing list archive at Nabble.com.


External File Field

2013-12-16 Thread Mukundaraman valakumaresan
Hi

Is there a way to find ,whether the External File Fields mentioned in the
schema.xml is being used or whether Solr reads the value of those external
fields.

I am not sure how to use an External field, can I request the value of an
External File Field in the field list or can I use in my custom function as
shown below:

http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,adjlocalityhttp://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,property_price_min,lanorm(%22%22,luxury_amenities_c,10),adjlocality
,custfunc(adjlocality).

I didn't get any output for the external fields in the queries that I
provide.

Thanks  Regards
Mukund


Re: External File Field

2013-12-16 Thread Ahmet Arslan
Hi,

You can request the value of the that field via fl=*,field(adjlocality)

See more about it : 
https://cwiki.apache.org/confluence/display/solr/Working+with+External+Files+and+Processes

Actually you can search it too with frange query parser. {!frange l=0 
u=0}field(adjlocality)




On Monday, December 16, 2013 5:05 PM, Mukundaraman valakumaresan 
muk...@8kmiles.com wrote:
Hi

Is there a way to find ,whether the External File Fields mentioned in the
schema.xml is being used or whether Solr reads the value of those external
fields.

I am not sure how to use an External field, can I request the value of an
External File Field in the field list or can I use in my custom function as
shown below:

http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,adjlocalityhttp://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,property_price_min,lanorm(%22%22,luxury_amenities_c,10),adjlocality
,custfunc(adjlocality).

I didn't get any output for the external fields in the queries that I
provide.

Thanks  Regards
Mukund


RE: How can you move a shard from one SolrCloud node to another?

2013-12-16 Thread Tim Potter
Hi Chris,

The easiest approach is to just create a new core on the new machine that 
references the collection and shard you want to migrate. For example, say you 
split shard1 of a collection named cloud, which results in having: shard1_0 
and shard1_1. Now let's say you want to migrate shard 1_0 over to the new 
machine. 

First, fire off a q=*:*distrib=false query to the shard you're migrating so 
that you know how many docs it has (which will be used to verify the migration 
was clean below).

Next, bring up the new machine in cloud mode (-zkHost=?) and then go to the 
admin console on that server. Nav to the core admin page and create a new core, 
specifying the collection and shard1_0 in the form; note: the form leads you to 
believe you need to create the directory on the local system but you actually 
don't need to worry about doing that as the config will get pulled from ZK and 
the directory will get created on the fly (at least that's what happened in my 
env using branch_4x). 

When the new core initializes, it will use good ol' snapshot replication to 
pull the index from the leader. Verify the new core is happy by executing the 
q=*:*distrib=false  query again. Once you're satisfied, you can unload the 
core you migrated.

Btw ... you can do all this with the core admin API instead of the Web UI if 
you want to script it.

Cheers,

Timothy Potter
Sr. Software Engineer, LucidWorks
www.lucidworks.com


From: cwhi chris.whi...@gmail.com
Sent: Sunday, December 15, 2013 3:43 PM
To: solr-user@lucene.apache.org
Subject: How can you move a shard from one SolrCloud node to another?

Let's say I want to rebalance a SolrCloud collection.  I call SPLITSHARD to
split an existing shard, and then I'd like to move one of the subshards to a
new machine so the index is more balanced.  Can this be done?  If not, how
do you rebalance an existing SolrCloud collection?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/How-can-you-move-a-shard-from-one-SolrCloud-node-to-another-tp4106815.html
Sent from the Solr - User mailing list archive at Nabble.com.

Re: How can you move a shard from one SolrCloud node to another?

2013-12-16 Thread Yago Riveiro
Tim, 

Can you explain how the replication snapshot is done using the coreAdminAPI? 

-- 
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)


On Monday, December 16, 2013 at 4:23 PM, Tim Potter wrote:

 Hi Chris,
 
 The easiest approach is to just create a new core on the new machine that 
 references the collection and shard you want to migrate. For example, say you 
 split shard1 of a collection named cloud, which results in having: shard1_0 
 and shard1_1. Now let's say you want to migrate shard 1_0 over to the new 
 machine. 
 
 First, fire off a q=*:*distrib=false query to the shard you're migrating so 
 that you know how many docs it has (which will be used to verify the 
 migration was clean below).
 
 Next, bring up the new machine in cloud mode (-zkHost=?) and then go to the 
 admin console on that server. Nav to the core admin page and create a new 
 core, specifying the collection and shard1_0 in the form; note: the form 
 leads you to believe you need to create the directory on the local system but 
 you actually don't need to worry about doing that as the config will get 
 pulled from ZK and the directory will get created on the fly (at least that's 
 what happened in my env using branch_4x). 
 
 When the new core initializes, it will use good ol' snapshot replication to 
 pull the index from the leader. Verify the new core is happy by executing the 
 q=*:*distrib=false query again. Once you're satisfied, you can unload the 
 core you migrated.
 
 Btw ... you can do all this with the core admin API instead of the Web UI if 
 you want to script it.
 
 Cheers,
 
 Timothy Potter
 Sr. Software Engineer, LucidWorks
 www.lucidworks.com (http://www.lucidworks.com)
 
 
 From: cwhi chris.whi...@gmail.com (mailto:chris.whi...@gmail.com)
 Sent: Sunday, December 15, 2013 3:43 PM
 To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org)
 Subject: How can you move a shard from one SolrCloud node to another?
 
 Let's say I want to rebalance a SolrCloud collection. I call SPLITSHARD to
 split an existing shard, and then I'd like to move one of the subshards to a
 new machine so the index is more balanced. Can this be done? If not, how
 do you rebalance an existing SolrCloud collection?
 
 
 
 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/How-can-you-move-a-shard-from-one-SolrCloud-node-to-another-tp4106815.html
 Sent from the Solr - User mailing list archive at Nabble.com 
 (http://Nabble.com).
 
 




Re: indexing from bowser

2013-12-16 Thread Gora Mohanty
On 16 December 2013 16:50, Nutan nutanshinde1...@gmail.com wrote:

 ok thanks,
 but is there any other way where -F is not used?

 I am creating a api in vc++ and to link to solr i am using libcurl,for this
 to work the string is the url,
 eg:
 curl_easy_setopt(curl,
 CURLOPT_URL,http://localhost:8080/solr/document/select?q=*%3A*wt=jsonindent=truefl=id;);
[...]

If you can do a post using curl from the command line, you can
add --libcurl file to get a libcurl example written to file. Please
see the curl man page for details.

Regards,
Gora


Re: OutOfMemoryError in RamUsageEstimator in solr 4.6

2013-12-16 Thread Shawn Heisey
On 12/16/2013 2:34 AM, Torben Greulich wrote:
 we get a OutOfMemoryError in RamUsageEstimator and are a little bit
 confused about the error.
 We are using solr 4.6 and are confused about the Lucene42DocValuesProducer.
 We checked current solr code and found that Lucene42NormsFormat will be
 returned as NormFormat in Lucene46Codec and so the Lucene42DocValuesProducer
 will be returned. Is there any special reason not to have a higher version
 of NormsFormat?

The format for that part of the index apparently hasn't changed since
Lucene 4.2.  Something else must have changed about the default index
format, or I expect that there would not be a 4.6 specific Codec object.
 From what I've seen, the verion number encoded in the class name
doesn't change until there's a new class needed for the default codec.

A java OutOfMemoryError means that you need to increase the max heap
when you start the program.  It looks like you're trying to get a core
status, so this is probably happening when using the admin UI.  To put
it quite simply, your Java max heap is not big enough to handle what
Solr has been asked to do.

http://wiki.apache.org/solr/SolrPerformanceProblems#Java_Heap

Thanks,
Shawn



Re: Poor performance on distributed search

2013-12-16 Thread ku3ia
Any ideas?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Poor-performance-on-distributed-search-tp3590028p4106968.html
Sent from the Solr - User mailing list archive at Nabble.com.


RE: How can you move a shard from one SolrCloud node to another?

2013-12-16 Thread Tim Potter
Hi Yago,

When you create a new core (via API or Web UI), you specify the collection name 
and shard id, in my example cloud and shard1_0. When the core initializes 
in SolrCloud mode, it recognizes that the collection exists and adds itself as 
a replica to the shard. Then the main replica recovery process kicks in; try 
PeerSync, realize too far out of date, try snapshot replication from leader. 
The following core API command led to the same result as using the UI:

curl -v 
http://localhost:8986/solr/admin/cores?action=CREATEcollection=cloudshard=shard1_0name=cloud_shard1_0_replica3;

The only trick here is you need to set the name of the core, which from what I 
can tell can be arbitrary but I chose to use the same naming standard as the 
other cores

Cheers,

Timothy Potter
Sr. Software Engineer, LucidWorks
www.lucidworks.com


From: Yago Riveiro yago.rive...@gmail.com
Sent: Monday, December 16, 2013 9:32 AM
To: solr-user@lucene.apache.org
Subject: Re: How can you move a shard from one SolrCloud node to another?

Tim,

Can you explain how the replication snapshot is done using the coreAdminAPI?

--
Yago Riveiro
Sent with Sparrow (http://www.sparrowmailapp.com/?sig)


On Monday, December 16, 2013 at 4:23 PM, Tim Potter wrote:

 Hi Chris,

 The easiest approach is to just create a new core on the new machine that 
 references the collection and shard you want to migrate. For example, say you 
 split shard1 of a collection named cloud, which results in having: shard1_0 
 and shard1_1. Now let's say you want to migrate shard 1_0 over to the new 
 machine.

 First, fire off a q=*:*distrib=false query to the shard you're migrating so 
 that you know how many docs it has (which will be used to verify the 
 migration was clean below).

 Next, bring up the new machine in cloud mode (-zkHost=?) and then go to the 
 admin console on that server. Nav to the core admin page and create a new 
 core, specifying the collection and shard1_0 in the form; note: the form 
 leads you to believe you need to create the directory on the local system but 
 you actually don't need to worry about doing that as the config will get 
 pulled from ZK and the directory will get created on the fly (at least that's 
 what happened in my env using branch_4x).

 When the new core initializes, it will use good ol' snapshot replication to 
 pull the index from the leader. Verify the new core is happy by executing the 
 q=*:*distrib=false query again. Once you're satisfied, you can unload the 
 core you migrated.

 Btw ... you can do all this with the core admin API instead of the Web UI if 
 you want to script it.

 Cheers,

 Timothy Potter
 Sr. Software Engineer, LucidWorks
 www.lucidworks.com (http://www.lucidworks.com)

 
 From: cwhi chris.whi...@gmail.com (mailto:chris.whi...@gmail.com)
 Sent: Sunday, December 15, 2013 3:43 PM
 To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org)
 Subject: How can you move a shard from one SolrCloud node to another?

 Let's say I want to rebalance a SolrCloud collection. I call SPLITSHARD to
 split an existing shard, and then I'd like to move one of the subshards to a
 new machine so the index is more balanced. Can this be done? If not, how
 do you rebalance an existing SolrCloud collection?



 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/How-can-you-move-a-shard-from-one-SolrCloud-node-to-another-tp4106815.html
 Sent from the Solr - User mailing list archive at Nabble.com 
 (http://Nabble.com).





RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

2013-12-16 Thread Trevor Handley
Also, I'm aware that there's two typos in my schema.xml attached. I forgot to 
remove the linebreak \ character from the two splitOnCaseChange sections.
This typo does not exist in the official schema.xml that solr is using.

-Original Message-
From: Trevor Handley [mailto:hand...@civicplus.com] 
Sent: Monday, December 16, 2013 12:24 PM
To: solr-user@lucene.apache.org
Subject: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Hello, I'm working with SolrCloud and trying to integrate the Suggester 
functionality http://wiki.apache.org/solr/Suggester.

I've configured the requestHandler and searchComponent sections of 
solrconfig.xml, and added a new fieldtype and field to schema.xml. These 
documents are attached to this message.

Background: This is a change I'm trying to make to a currently working/stable 
version of Solr 4.6 that has nearly 1.5 million unique documents in the index. 
The whole architecture is a single SolrCloud collection with 2 core shards that 
are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 
core exist on one physical server, and shard1_slice2 core and shard2_slice1 
core exist on a separate physical server.

When I try to restart solr with suggester enabled then I get java error 
java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

I tried a few different suggester classes but they all fail to load with the 
same message here.
I verified that my .jar files do contain that class and are in the correct lib 
directory using a script that searches .jar files for a class name:

[solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup 
/opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class
/opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class

And here's a listing of the jar files in my lib directory:
activation-1.1.jar
AlchemyAPIAnnotator-2.3.1.jar
apache-mime4j-core-0.7.2.jar
apache-mime4j-dom-0.7.2.jar
attributes-binder-1.2.0.jar
bcmail-jdk15-1.45.jar
bcprov-jdk15-1.45.jar
boilerpipe-1.1.0.jar
carrot2-mini-3.8.0.jar
commons-beanutils-1.7.0.jar
commons-collections-3.2.1.jar
commons-compress-1.4.1.jar
commons-digester-2.0.jar
dom4j-1.6.1.jar
fontbox-1.8.1.jar
hppc-0.5.2.jar
icu4j-49.1.jar
isoparser-1.0-RC-1.jar
jackson-core-asl-1.7.4.jar
jackson-mapper-asl-1.7.4.jar
jdom-1.0.jar
jempbox-1.8.1.jar
jetty-continuation-8.1.10.v20130312.jar
jetty-deploy-8.1.10.v20130312.jar
jetty-http-8.1.10.v20130312.jar
jetty-io-8.1.10.v20130312.jar
jetty-jmx-8.1.10.v20130312.jar
jetty-security-8.1.10.v20130312.jar
jetty-server-8.1.10.v20130312.jar
jetty-servlet-8.1.10.v20130312.jar
jetty-util-8.1.10.v20130312.jar
jetty-webapp-8.1.10.v20130312.jar
jetty-xml-8.1.10.v20130312.jar
jsonic-1.2.7.jar
juniversalchardet-1.0.3.jar
langdetect-1.1-20120112.jar
lucene-analyzers-common-4.6-SNAPSHOT.jar
lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar
lucene-analyzers-phonetic-4.6-SNAPSHOT.jar
lucene-codecs-4.6-SNAPSHOT.jar
lucene-core-4.6-SNAPSHOT.jar
lucene-grouping-4.6-SNAPSHOT.jar
lucene-highlighter-4.6-SNAPSHOT.jar
lucene-join-4.6-SNAPSHOT.jar
lucene-memory-4.6-SNAPSHOT.jar
lucene-misc-4.6-SNAPSHOT.jar
lucene-queries-4.6-SNAPSHOT.jar
lucene-queryparser-4.6-SNAPSHOT.jar
lucene-spatial-4.6-SNAPSHOT.jar
lucene-suggest-4.6-SNAPSHOT.jar
mahout-collections-1.0.jar
mahout-math-0.6.jar
mail-1.4.1.jar
metadata-extractor-2.6.2.jar
morfologik-fsa-1.7.1.jar
morfologik-polish-1.7.1.jar
morfologik-stemming-1.7.1.jar
netcdf-4.2-min.jar
OpenCalaisAnnotator-2.3.1.jar
pdfbox-1.8.1.jar
poi-3.9.jar
poi-ooxml-3.9.jar
poi-ooxml-schemas-3.9.jar
poi-scratchpad-3.9.jar
rome-0.9.jar
servlet-api-3.0.jar
simple-xml-2.7.jar
solr-analysis-extras-4.6.0.jar
solr-cell-4.6.0.jar
solr-clustering-4.6.0.jar
solr-core-4.6.0.jar
solr-dataimporthandler-4.6.0.jar
solr-dataimporthandler-extras-4.6.0.jar
solr-langid-4.6.0.jar
solr-solrj-4.6.0.jar
solr-test-framework-4.6.0.jar
solr-uima-4.6.0.jar
solr-velocity-4.6.0.jar
Tagger-2.3.1.jar
tagsoup-1.2.1.jar
tika-core-1.4.jar
tika-parsers-1.4.jar
uimaj-core-2.3.1.jar
velocity-1.7.jar
velocity-tools-2.0.jar
vorbis-java-core-0.1.jar
vorbis-java-tika-0.1.jar
WhitespaceTokenizer-2.3.1.jar
xercesImpl-2.9.1.jar
xmlbeans-2.3.0.jar
xz-1.0.jar

This is how I start solr with jetty:
java -Dbootstrap_confdir=/opt/solrcloud/zkBootstrapConfigs/ 
-Dcollection.configName=CP_Search 
-DzkHost=zookeeper-001:2181,zookeeper-002:2181,zookeeper-003:2181 
-Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.port=3000 
-Dcom.sun.management.jmxremote.ssl=false 
-Dcom.sun.management.jmxremote.authenticate=false -server -Xms3g -Xmx6g -Xmn2g 
-XX:ParallelGCThreads=20 -XX:SurvivorRatio=8 -XX:TargetSurvivorRatio=90 
-XX:MaxTenuringThreshold=31 -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 
-XX:+CMSParallelRemarkEnabled -XX:+DisableExplicitGC -Djetty.port=8501 
-Djetty.home=/opt/solrcloud/ -Dsolr.solr.home=/opt/solrcloud/ -DnumShards=2 
-jar 

PostingsSolrHighlighter

2013-12-16 Thread Josip Delic

Hi @all,

i am playing with the PostingsSolrHighlighter. I'm running solr 4.6.0 
and my configuration is from here:


https://lucene.apache.org/solr/4_6_0/solr-core/org/apache/solr/highlight/PostingsSolrHighlighter.html

Search query and result (not working):

http://pastebin.com/13Uan0ZF

Schema (not complete):

http://pastebin.com/JGa38UDT

Search query and result (working):

http://pastebin.com/4CP8XKnr

Solr config:

searchComponent class=solr.HighlightComponent name=highlight
  highlighting 
class=org.apache.solr.highlight.PostingsSolrHighlighter/ 



/searchComponent

So this is working just fine, but now i have some questions:

1.) With the old default highlighter component it was possible to search 
in searchable_text and to retrive highlighted text. This is 
essential, because we use copyfield to put almost everything to 
searchable_text (title, subtitle, description, ...)


2.) I can't get ellipsis working i tried hl.tag.ellipsis=..., 
f.text.hl.tag.ellipsis=..., configuring it in RequestHandler noting 
seems to work, maxAnalyzedChars is just cutting the sentence?


Kind Regards

Josip Delic



smime.p7s
Description: S/MIME Cryptographic Signature


RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

2013-12-16 Thread Tim Potter
There have been some recent refactorings in this area of the code. The 
following class name should work:

org.apache.solr.spelling.suggest.tst.TSTLookupFactory

Cheers,

Timothy Potter
Sr. Software Engineer, LucidWorks
www.lucidworks.com


From: Trevor Handley hand...@civicplus.com
Sent: Monday, December 16, 2013 11:27 AM
To: solr-user@lucene.apache.org
Subject: RE: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Also, I'm aware that there's two typos in my schema.xml attached. I forgot to 
remove the linebreak \ character from the two splitOnCaseChange sections.
This typo does not exist in the official schema.xml that solr is using.

-Original Message-
From: Trevor Handley [mailto:hand...@civicplus.com]
Sent: Monday, December 16, 2013 12:24 PM
To: solr-user@lucene.apache.org
Subject: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Hello, I'm working with SolrCloud and trying to integrate the Suggester 
functionality http://wiki.apache.org/solr/Suggester.

I've configured the requestHandler and searchComponent sections of 
solrconfig.xml, and added a new fieldtype and field to schema.xml. These 
documents are attached to this message.

Background: This is a change I'm trying to make to a currently working/stable 
version of Solr 4.6 that has nearly 1.5 million unique documents in the index. 
The whole architecture is a single SolrCloud collection with 2 core shards that 
are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 
core exist on one physical server, and shard1_slice2 core and shard2_slice1 
core exist on a separate physical server.

When I try to restart solr with suggester enabled then I get java error 
java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

I tried a few different suggester classes but they all fail to load with the 
same message here.
I verified that my .jar files do contain that class and are in the correct lib 
directory using a script that searches .jar files for a class name:

[solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup 
/opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class
/opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class

And here's a listing of the jar files in my lib directory:
activation-1.1.jar
AlchemyAPIAnnotator-2.3.1.jar
apache-mime4j-core-0.7.2.jar
apache-mime4j-dom-0.7.2.jar
attributes-binder-1.2.0.jar
bcmail-jdk15-1.45.jar
bcprov-jdk15-1.45.jar
boilerpipe-1.1.0.jar
carrot2-mini-3.8.0.jar
commons-beanutils-1.7.0.jar
commons-collections-3.2.1.jar
commons-compress-1.4.1.jar
commons-digester-2.0.jar
dom4j-1.6.1.jar
fontbox-1.8.1.jar
hppc-0.5.2.jar
icu4j-49.1.jar
isoparser-1.0-RC-1.jar
jackson-core-asl-1.7.4.jar
jackson-mapper-asl-1.7.4.jar
jdom-1.0.jar
jempbox-1.8.1.jar
jetty-continuation-8.1.10.v20130312.jar
jetty-deploy-8.1.10.v20130312.jar
jetty-http-8.1.10.v20130312.jar
jetty-io-8.1.10.v20130312.jar
jetty-jmx-8.1.10.v20130312.jar
jetty-security-8.1.10.v20130312.jar
jetty-server-8.1.10.v20130312.jar
jetty-servlet-8.1.10.v20130312.jar
jetty-util-8.1.10.v20130312.jar
jetty-webapp-8.1.10.v20130312.jar
jetty-xml-8.1.10.v20130312.jar
jsonic-1.2.7.jar
juniversalchardet-1.0.3.jar
langdetect-1.1-20120112.jar
lucene-analyzers-common-4.6-SNAPSHOT.jar
lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar
lucene-analyzers-phonetic-4.6-SNAPSHOT.jar
lucene-codecs-4.6-SNAPSHOT.jar
lucene-core-4.6-SNAPSHOT.jar
lucene-grouping-4.6-SNAPSHOT.jar
lucene-highlighter-4.6-SNAPSHOT.jar
lucene-join-4.6-SNAPSHOT.jar
lucene-memory-4.6-SNAPSHOT.jar
lucene-misc-4.6-SNAPSHOT.jar
lucene-queries-4.6-SNAPSHOT.jar
lucene-queryparser-4.6-SNAPSHOT.jar
lucene-spatial-4.6-SNAPSHOT.jar
lucene-suggest-4.6-SNAPSHOT.jar
mahout-collections-1.0.jar
mahout-math-0.6.jar
mail-1.4.1.jar
metadata-extractor-2.6.2.jar
morfologik-fsa-1.7.1.jar
morfologik-polish-1.7.1.jar
morfologik-stemming-1.7.1.jar
netcdf-4.2-min.jar
OpenCalaisAnnotator-2.3.1.jar
pdfbox-1.8.1.jar
poi-3.9.jar
poi-ooxml-3.9.jar
poi-ooxml-schemas-3.9.jar
poi-scratchpad-3.9.jar
rome-0.9.jar
servlet-api-3.0.jar
simple-xml-2.7.jar
solr-analysis-extras-4.6.0.jar
solr-cell-4.6.0.jar
solr-clustering-4.6.0.jar
solr-core-4.6.0.jar
solr-dataimporthandler-4.6.0.jar
solr-dataimporthandler-extras-4.6.0.jar
solr-langid-4.6.0.jar
solr-solrj-4.6.0.jar
solr-test-framework-4.6.0.jar
solr-uima-4.6.0.jar
solr-velocity-4.6.0.jar
Tagger-2.3.1.jar
tagsoup-1.2.1.jar
tika-core-1.4.jar
tika-parsers-1.4.jar
uimaj-core-2.3.1.jar
velocity-1.7.jar
velocity-tools-2.0.jar
vorbis-java-core-0.1.jar
vorbis-java-tika-0.1.jar
WhitespaceTokenizer-2.3.1.jar
xercesImpl-2.9.1.jar
xmlbeans-2.3.0.jar
xz-1.0.jar

This is how I start solr with jetty:
java -Dbootstrap_confdir=/opt/solrcloud/zkBootstrapConfigs/ 
-Dcollection.configName=CP_Search 

RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

2013-12-16 Thread Trevor Handley
Brilliant, thanks Timothy!

Changing the solrconfig.xml lookupImpl (not className) to the 
org.apache.solr.spelling.suggest.tst.TSTLookupFactory fixed this issue for me.

Thanks, Trevor

-Original Message-
From: Tim Potter [mailto:tim.pot...@lucidworks.com] 
Sent: Monday, December 16, 2013 12:32 PM
To: solr-user@lucene.apache.org
Subject: RE: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

There have been some recent refactorings in this area of the code. The 
following class name should work:

org.apache.solr.spelling.suggest.tst.TSTLookupFactory

Cheers,

Timothy Potter
Sr. Software Engineer, LucidWorks
www.lucidworks.com


From: Trevor Handley hand...@civicplus.com
Sent: Monday, December 16, 2013 11:27 AM
To: solr-user@lucene.apache.org
Subject: RE: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Also, I'm aware that there's two typos in my schema.xml attached. I forgot to 
remove the linebreak \ character from the two splitOnCaseChange sections.
This typo does not exist in the official schema.xml that solr is using.

-Original Message-
From: Trevor Handley [mailto:hand...@civicplus.com]
Sent: Monday, December 16, 2013 12:24 PM
To: solr-user@lucene.apache.org
Subject: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Hello, I'm working with SolrCloud and trying to integrate the Suggester 
functionality http://wiki.apache.org/solr/Suggester.

I've configured the requestHandler and searchComponent sections of 
solrconfig.xml, and added a new fieldtype and field to schema.xml. These 
documents are attached to this message.

Background: This is a change I'm trying to make to a currently working/stable 
version of Solr 4.6 that has nearly 1.5 million unique documents in the index. 
The whole architecture is a single SolrCloud collection with 2 core shards that 
are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 
core exist on one physical server, and shard1_slice2 core and shard2_slice1 
core exist on a separate physical server.

When I try to restart solr with suggester enabled then I get java error 
java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

I tried a few different suggester classes but they all fail to load with the 
same message here.
I verified that my .jar files do contain that class and are in the correct lib 
directory using a script that searches .jar files for a class name:

[solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup 
/opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class
/opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class

And here's a listing of the jar files in my lib directory:
activation-1.1.jar
AlchemyAPIAnnotator-2.3.1.jar
apache-mime4j-core-0.7.2.jar
apache-mime4j-dom-0.7.2.jar
attributes-binder-1.2.0.jar
bcmail-jdk15-1.45.jar
bcprov-jdk15-1.45.jar
boilerpipe-1.1.0.jar
carrot2-mini-3.8.0.jar
commons-beanutils-1.7.0.jar
commons-collections-3.2.1.jar
commons-compress-1.4.1.jar
commons-digester-2.0.jar
dom4j-1.6.1.jar
fontbox-1.8.1.jar
hppc-0.5.2.jar
icu4j-49.1.jar
isoparser-1.0-RC-1.jar
jackson-core-asl-1.7.4.jar
jackson-mapper-asl-1.7.4.jar
jdom-1.0.jar
jempbox-1.8.1.jar
jetty-continuation-8.1.10.v20130312.jar
jetty-deploy-8.1.10.v20130312.jar
jetty-http-8.1.10.v20130312.jar
jetty-io-8.1.10.v20130312.jar
jetty-jmx-8.1.10.v20130312.jar
jetty-security-8.1.10.v20130312.jar
jetty-server-8.1.10.v20130312.jar
jetty-servlet-8.1.10.v20130312.jar
jetty-util-8.1.10.v20130312.jar
jetty-webapp-8.1.10.v20130312.jar
jetty-xml-8.1.10.v20130312.jar
jsonic-1.2.7.jar
juniversalchardet-1.0.3.jar
langdetect-1.1-20120112.jar
lucene-analyzers-common-4.6-SNAPSHOT.jar
lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar
lucene-analyzers-phonetic-4.6-SNAPSHOT.jar
lucene-codecs-4.6-SNAPSHOT.jar
lucene-core-4.6-SNAPSHOT.jar
lucene-grouping-4.6-SNAPSHOT.jar
lucene-highlighter-4.6-SNAPSHOT.jar
lucene-join-4.6-SNAPSHOT.jar
lucene-memory-4.6-SNAPSHOT.jar
lucene-misc-4.6-SNAPSHOT.jar
lucene-queries-4.6-SNAPSHOT.jar
lucene-queryparser-4.6-SNAPSHOT.jar
lucene-spatial-4.6-SNAPSHOT.jar
lucene-suggest-4.6-SNAPSHOT.jar
mahout-collections-1.0.jar
mahout-math-0.6.jar
mail-1.4.1.jar
metadata-extractor-2.6.2.jar
morfologik-fsa-1.7.1.jar
morfologik-polish-1.7.1.jar
morfologik-stemming-1.7.1.jar
netcdf-4.2-min.jar
OpenCalaisAnnotator-2.3.1.jar
pdfbox-1.8.1.jar
poi-3.9.jar
poi-ooxml-3.9.jar
poi-ooxml-schemas-3.9.jar
poi-scratchpad-3.9.jar
rome-0.9.jar
servlet-api-3.0.jar
simple-xml-2.7.jar
solr-analysis-extras-4.6.0.jar
solr-cell-4.6.0.jar
solr-clustering-4.6.0.jar
solr-core-4.6.0.jar
solr-dataimporthandler-4.6.0.jar
solr-dataimporthandler-extras-4.6.0.jar
solr-langid-4.6.0.jar
solr-solrj-4.6.0.jar
solr-test-framework-4.6.0.jar
solr-uima-4.6.0.jar
solr-velocity-4.6.0.jar
Tagger-2.3.1.jar

RE: SolrCloud Suggester java ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

2013-12-16 Thread Tim Potter
Awesome ... I'll update the Wiki to reflect the new class names.

Timothy Potter
Sr. Software Engineer, LucidWorks
www.lucidworks.com


From: Trevor Handley hand...@civicplus.com
Sent: Monday, December 16, 2013 11:44 AM
To: solr-user@lucene.apache.org
Subject: RE: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Brilliant, thanks Timothy!

Changing the solrconfig.xml lookupImpl (not className) to the 
org.apache.solr.spelling.suggest.tst.TSTLookupFactory fixed this issue for me.

Thanks, Trevor

-Original Message-
From: Tim Potter [mailto:tim.pot...@lucidworks.com]
Sent: Monday, December 16, 2013 12:32 PM
To: solr-user@lucene.apache.org
Subject: RE: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

There have been some recent refactorings in this area of the code. The 
following class name should work:

org.apache.solr.spelling.suggest.tst.TSTLookupFactory

Cheers,

Timothy Potter
Sr. Software Engineer, LucidWorks
www.lucidworks.com


From: Trevor Handley hand...@civicplus.com
Sent: Monday, December 16, 2013 11:27 AM
To: solr-user@lucene.apache.org
Subject: RE: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Also, I'm aware that there's two typos in my schema.xml attached. I forgot to 
remove the linebreak \ character from the two splitOnCaseChange sections.
This typo does not exist in the official schema.xml that solr is using.

-Original Message-
From: Trevor Handley [mailto:hand...@civicplus.com]
Sent: Monday, December 16, 2013 12:24 PM
To: solr-user@lucene.apache.org
Subject: SolrCloud Suggester java ClassNotFoundException: 
org.apache.solr.suggest.tst.TSTLookup

Hello, I'm working with SolrCloud and trying to integrate the Suggester 
functionality http://wiki.apache.org/solr/Suggester.

I've configured the requestHandler and searchComponent sections of 
solrconfig.xml, and added a new fieldtype and field to schema.xml. These 
documents are attached to this message.

Background: This is a change I'm trying to make to a currently working/stable 
version of Solr 4.6 that has nearly 1.5 million unique documents in the index. 
The whole architecture is a single SolrCloud collection with 2 core shards that 
are replicated for a total of 4 cores. The shard1_slice1 core and shard2_slice2 
core exist on one physical server, and shard1_slice2 core and shard2_slice1 
core exist on a separate physical server.

When I try to restart solr with suggester enabled then I get java error 
java.lang.ClassNotFoundException: org.apache.solr.suggest.tst.TSTLookup

I tried a few different suggester classes but they all fail to load with the 
same message here.
I verified that my .jar files do contain that class and are in the correct lib 
directory using a script that searches .jar files for a class name:

[solr@Searchnode-001 ~]$ ./findclass.sh /opt/solrcloud/lib/ TSTLookup 
/opt/solrcloud/lib/solr-core-4.6.0.jar:org/apache/solr/spelling/suggest/tst/TSTLookupFactory.class
/opt/solrcloud/lib/lucene-suggest-4.6-SNAPSHOT.jar:org/apache/lucene/search/suggest/tst/TSTLookup.class

And here's a listing of the jar files in my lib directory:
activation-1.1.jar
AlchemyAPIAnnotator-2.3.1.jar
apache-mime4j-core-0.7.2.jar
apache-mime4j-dom-0.7.2.jar
attributes-binder-1.2.0.jar
bcmail-jdk15-1.45.jar
bcprov-jdk15-1.45.jar
boilerpipe-1.1.0.jar
carrot2-mini-3.8.0.jar
commons-beanutils-1.7.0.jar
commons-collections-3.2.1.jar
commons-compress-1.4.1.jar
commons-digester-2.0.jar
dom4j-1.6.1.jar
fontbox-1.8.1.jar
hppc-0.5.2.jar
icu4j-49.1.jar
isoparser-1.0-RC-1.jar
jackson-core-asl-1.7.4.jar
jackson-mapper-asl-1.7.4.jar
jdom-1.0.jar
jempbox-1.8.1.jar
jetty-continuation-8.1.10.v20130312.jar
jetty-deploy-8.1.10.v20130312.jar
jetty-http-8.1.10.v20130312.jar
jetty-io-8.1.10.v20130312.jar
jetty-jmx-8.1.10.v20130312.jar
jetty-security-8.1.10.v20130312.jar
jetty-server-8.1.10.v20130312.jar
jetty-servlet-8.1.10.v20130312.jar
jetty-util-8.1.10.v20130312.jar
jetty-webapp-8.1.10.v20130312.jar
jetty-xml-8.1.10.v20130312.jar
jsonic-1.2.7.jar
juniversalchardet-1.0.3.jar
langdetect-1.1-20120112.jar
lucene-analyzers-common-4.6-SNAPSHOT.jar
lucene-analyzers-kuromoji-4.6-SNAPSHOT.jar
lucene-analyzers-phonetic-4.6-SNAPSHOT.jar
lucene-codecs-4.6-SNAPSHOT.jar
lucene-core-4.6-SNAPSHOT.jar
lucene-grouping-4.6-SNAPSHOT.jar
lucene-highlighter-4.6-SNAPSHOT.jar
lucene-join-4.6-SNAPSHOT.jar
lucene-memory-4.6-SNAPSHOT.jar
lucene-misc-4.6-SNAPSHOT.jar
lucene-queries-4.6-SNAPSHOT.jar
lucene-queryparser-4.6-SNAPSHOT.jar
lucene-spatial-4.6-SNAPSHOT.jar
lucene-suggest-4.6-SNAPSHOT.jar
mahout-collections-1.0.jar
mahout-math-0.6.jar
mail-1.4.1.jar
metadata-extractor-2.6.2.jar
morfologik-fsa-1.7.1.jar
morfologik-polish-1.7.1.jar
morfologik-stemming-1.7.1.jar
netcdf-4.2-min.jar
OpenCalaisAnnotator-2.3.1.jar
pdfbox-1.8.1.jar
poi-3.9.jar
poi-ooxml-3.9.jar

Re: OutOfMemoryError in RamUsageEstimator in solr 4.6

2013-12-16 Thread Torben Greulich
Hi Shawn,
thanks for your reply. But we don't think that this is really a OOM error,
because we already increased the heap to 64gb and the OOM occurs at a usage
of 30-40gb. So solr would allocate more than 20gb at once. this sounds a
little bit too much.

Furthermore we found Lucene45DocValuesProducer and wondered why it exists
and isn't used here. Lucene45DocValuesProducer.ramBytesUsed() also looks
different to Lucene42DocValuesProducer.ramBytesUsed()

Best Regards
Torben


2013/12/16 Shawn Heisey s...@elyograg.org

 On 12/16/2013 2:34 AM, Torben Greulich wrote:
  we get a OutOfMemoryError in RamUsageEstimator and are a little bit
  confused about the error.
  We are using solr 4.6 and are confused about the
 Lucene42DocValuesProducer.
  We checked current solr code and found that Lucene42NormsFormat will be
  returned as NormFormat in Lucene46Codec and so the
 Lucene42DocValuesProducer
  will be returned. Is there any special reason not to have a higher
 version
  of NormsFormat?

 The format for that part of the index apparently hasn't changed since
 Lucene 4.2.  Something else must have changed about the default index
 format, or I expect that there would not be a 4.6 specific Codec object.
  From what I've seen, the verion number encoded in the class name
 doesn't change until there's a new class needed for the default codec.

 A java OutOfMemoryError means that you need to increase the max heap
 when you start the program.  It looks like you're trying to get a core
 status, so this is probably happening when using the admin UI.  To put
 it quite simply, your Java max heap is not big enough to handle what
 Solr has been asked to do.

 http://wiki.apache.org/solr/SolrPerformanceProblems#Java_Heap

 Thanks,
 Shawn




-- 
Torben Greulich · Entwickler
Backend Development
shopping24 internet group

Wandsbeker Straße 3-7 · 22172 Hamburg
Telefon: +49 (0) 40 6461  · Fax: +49 (0) 40 6461 7879
torben.greul...@s24.com · www.s24.com
AG Hamburg HRB 63371
vertreten durch Dr. Björn Schäfers und Martin Mildner


Re: Possible parent/child query bug

2013-12-16 Thread Mikhail Khludnev
On Fri, Nov 22, 2013 at 6:18 PM, Neil Ireson n.ire...@sheffield.ac.ukwrote:


 If the child of” query matches both parent and child docs it returns the
 child documents but a spurious numFound.


follow up https://issues.apache.org/jira/browse/SOLR-5553


-- 
Sincerely yours
Mikhail Khludnev
Principal Engineer,
Grid Dynamics

http://www.griddynamics.com
 mkhlud...@griddynamics.com


Question about replication problem

2013-12-16 Thread Fred Drake
I have a multi-core master/slave configuration that's showing
unexpected replication behavior for one core; other cores are
replicating without problems.

The master is running Solr 4.1; one slave is running 4.1 under Tomcat,
and another (for testing) is running 4.6 under Jetty.  These are
exhibiting similar problems.

The core on the master is at generation 4; this is reported as both
the generation for searching and replication using the replication API
on the slave.

When the slave attempts to replicate, it fails to load the file list
required for replication, and logs the message:

INFO: Master's generation: 4
INFO: Slave's generation: 5
INFO: Starting replication process
SEVERE: No files to download for index generation: 4

(Timestamps removed for readability.)

I have verified that the slave returns empty results where the master
returns non-empty results; so this isn't a matter of reporting being
incorrect.

I'm not sure how to diagnose this; what can I check for to understand
the actual problem?

(I have Googled for relevant issues or mailing list traffic, but
nothing actually explained what's happening or how to correct it.)



  -Fred

-- 
Fred L. Drake, Jr.fdrake at gmail.com
A storm broke loose in my mind.  --Albert Einstein


Off-line search on mobile devices

2013-12-16 Thread Arcadius Ahouansou
Hello.

We are planning to offer search as an embedded functionality into
mobile/low-power devices.

The main requirement are:

- ability to index and search documents available on the mobile device,
- no need of internet access,
- lightweight, low footprint and fast

We are looking into various options.

As I understand it, Solr would be way too heavy for mobile devices.

Has anyone used Lucene/Solr for off-line search on mobile devices?

Are there better alternatives for off-line full-text search?

Many thanks.

Arcadius.


Re: Off-line search on mobile devices

2013-12-16 Thread Sameer Maggon
1. Which platform are you looking at? Android, iOS, other?

If you are on Android, you can directly use lucene to build an embedded
solution for search. Depending upon your need, that can offer a small
enough footprint. We've done some work around embedding lucene for a
specific application on Android, happy to brainstorm offline.

Thanks,
Sameer.
--
http://measuredsearch.com



On Mon, Dec 16, 2013 at 3:07 PM, Arcadius Ahouansou arcad...@menelic.comwrote:

 Hello.

 We are planning to offer search as an embedded functionality into
 mobile/low-power devices.

 The main requirement are:

 - ability to index and search documents available on the mobile device,
 - no need of internet access,
 - lightweight, low footprint and fast

 We are looking into various options.

 As I understand it, Solr would be way too heavy for mobile devices.

 Has anyone used Lucene/Solr for off-line search on mobile devices?

 Are there better alternatives for off-line full-text search?

 Many thanks.

 Arcadius.




-- 
Sameer Maggon
Founder, Measured Search
m: 310.344.7266
tw: @measuredsearch
w: http://www.measuredsearch.com


how to remove documents which have been deleted from the database

2013-12-16 Thread kaustubh147

Hi,

Glassfish 3.1.2.2 
Solr 4.5 
Zookeeper 3.4.5 

We have set up a SolrCloud with 4 Solr nodes and 3 zookeeper instances.
I have 5 cores with  1 shard/4 replica setup on each of them.


One of our core is very small, and it takes less than one minute to index.
We run full import on it every hour with commit= true and clear =false but
full import does not delete documents which are no more there in database.
if I run the import with clear =true it will delete the index first and
shows 0 records during the import is in progress.

For delta we will have to create a new table in database which will keep
track of deleted records. we dont want to go that route.

Is there a way, where, we can run a full import with clear= true and solr
only update the data after the import is completed.


Regards,
Kaustubh 





--
View this message in context: 
http://lucene.472066.n3.nabble.com/how-to-remove-documents-which-have-been-deleted-from-the-database-tp4107015.html
Sent from the Solr - User mailing list archive at Nabble.com.


zookeeper timeout issue

2013-12-16 Thread kaustubh147
Hi,

Following warning message is filling our application logs very rapidly. This
statement is printed every time application talks with zookeeper.


[#|2013-12-13T08:33:03.023-0800|INFO|glassfish3.1.2|javax.enterprise.system.std.com.sun.enterprise.server.logging|_ThreadID=64;_ThreadName=Thread-2;|2013-12-13
08:33:03,023 WARN   org.apache.zookeeper.ClientCnxn - Session
0x342e9013f2d0063 for server null, unexpected error, closing socket
connection and attempting reconnect
java.lang.NullPointerException: key can't be null
at java.lang.System.checkKey(System.java:771)
at java.lang.System.getProperty(System.java:647)
at
org.apache.zookeeper.client.ZooKeeperSaslClient.init(ZooKeeperSaslClient.java:133)
at
org.apache.zookeeper.ClientCnxn$SendThread.startConnect(ClientCnxn.java:943)
at org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:993)


I am getting following INFO on zookeeper logs



[myid:1] - INFO 
[NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:ZooKeeperServer@812] - Refusing
session request for client /IP1:15703 as it has seen zxid 0x1c1103 our
last zxid is 0x40348 client must try another server
2013-12-16 15:45:56,999 [myid:1] - INFO 
[NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1001] - Closed
socket connection for client /10.163.160.78:15703 (no session established
for client)


it could be related to this problem

https://issues.apache.org/jira/browse/ZOOKEEPER-1237
or 
http://mail-archives.apache.org/mod_mbox/zookeeper-user/201208.mbox/%3CCANLc_9Jwieyyig=yg1yvczaeobc8swwj3fqd4x993ryrpod...@mail.gmail.com%3E

Is there a solution to this problem or I will have to wait for next
zookeeper release.

Regards,
Kaustubh





--
View this message in context: 
http://lucene.472066.n3.nabble.com/zookeeper-timeout-issue-tp4107016.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: how to remove documents which have been deleted from the database

2013-12-16 Thread Shalin Shekhar Mangar
If your index is that small then use clean=true and let
DataImportHandler clear all documents at the start of the import.
Don't bother with delta tables for such a small index.

On Tue, Dec 17, 2013 at 6:10 AM, kaustubh147 kaustubh.j...@gmail.com wrote:

 Hi,

 Glassfish 3.1.2.2
 Solr 4.5
 Zookeeper 3.4.5

 We have set up a SolrCloud with 4 Solr nodes and 3 zookeeper instances.
 I have 5 cores with  1 shard/4 replica setup on each of them.


 One of our core is very small, and it takes less than one minute to index.
 We run full import on it every hour with commit= true and clear =false but
 full import does not delete documents which are no more there in database.
 if I run the import with clear =true it will delete the index first and
 shows 0 records during the import is in progress.

 For delta we will have to create a new table in database which will keep
 track of deleted records. we dont want to go that route.

 Is there a way, where, we can run a full import with clear= true and solr
 only update the data after the import is completed.


 Regards,
 Kaustubh





 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/how-to-remove-documents-which-have-been-deleted-from-the-database-tp4107015.html
 Sent from the Solr - User mailing list archive at Nabble.com.



-- 
Regards,
Shalin Shekhar Mangar.


Re: Storing MYSQL DATETIME field in solr as String

2013-12-16 Thread Shalin Shekhar Mangar
I'm sorry. I thought you wanted to parse a date stored as string into
a java.util.Date. Clearly, you are trying to go the other way round.

There's nothing in DIH which will convert a mysql date to a string in
specific format. You will need to write a custom transformer either in
javascript or in java to do this.

http://wiki.apache.org/solr/DIHCustomTransformer
http://wiki.apache.org/solr/DataImportHandler#ScriptTransformer

On Mon, Dec 16, 2013 at 1:34 PM, manju16832003 manju16832...@gmail.com wrote:
 OK here are the scenarios I tried.

 *Scenario - 1: *

 dih.xml (aka data-config.xml)

 entity dataSource=solr name=listing query=...
 transformer=DateFormatTransformer

 field name=quot;publish_datequot; column=quot;publish_datequot;
 xpath=quot;/RDF/item/datequot;
 dateTimeFormat=quot;lt;b-MM-dd'T'HH:mm:ss* locale=en/

 schema.xml
 field name=publish_date type=date indexed=true stored=true
 multiValued=false default=NOW /

 It does not work. Throws an following exception
 WARN  org.apache.solr.handler.dataimport.DateFormatTransformer  – Could not
 parse a Date field
 java.text.ParseException: Unparseable date: 2013-12-05 15:40:03.0

 *Scenario - 2*
 dih.xml (aka data-config.xml)

 entity dataSource=solr name=listing query=...
 transformer=DateFormatTransformer

 field name=quot;publish_datequot; column=quot;publish_datequot;
 xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd
 HH:mm:ss* locale=en/

 schema.xml
 field name=publish_date type=date indexed=true stored=true
 multiValued=false default=NOW /

 Output is
 date name=publish_date2012-12-05T07:38:27Z/date

 *Scenario - 3*
 dih.xml (aka data-config.xml)

 entity dataSource=solr name=listing query=...
 transformer=DateFormatTransformer

 field name=quot;publish_datequot; column=quot;publish_datequot;
 xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd
 HH:mm:ss* locale=en/

 schema.xml
 field name=publish_date type=string indexed=true stored=true
 multiValued=false default=NOW /

 Output is
   str name=publish_dateThu Dec 05 15:38:27 MYT 2013/str

 My expectation was *2012-12-05 07:38:27* (MySQL DateTime format)



 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Storing-MYSQL-DATETIME-field-in-solr-as-String-tp4106836p4106854.html
 Sent from the Solr - User mailing list archive at Nabble.com.



-- 
Regards,
Shalin Shekhar Mangar.


Cannot create core

2013-12-16 Thread YouPeng Yang
Hi

 I get an weird problem.
 I try to create a core within Solr4.6.

Firstly, on my solr web server tomcat, a log come out[1]:
Then lots of Overseer  Info  logs come out as [2]. And then the creation
failed.

I have also notice that there is a lot of qn in the Overseer  on the
zookeeper:

[zk: localhost:2181(CONNECTED) 45] ls /overseer/queue
[qn-38, qn-37, qn-39, qn-34, qn-33,
qn-36, qn-44, qn-35, qn-42, qn-43,
qn-30, qn-31, qn-40, qn-32, qn-41]


Why are there so many qn.? I do not know the reason to the core creation
failed.
Please help .



[1]
..
293699 [http-bio-8081-exec-1] INFO  org.apache.solr.cloud.ZkController  –
publishing core=reportCore_201310 state=down
293700 [http-bio-8081-exec-1] INFO  org.apache.solr.cloud.ZkController  –
numShards not found on descriptor - reading it from system property
293704 [http-bio-8081-exec-1] INFO  org.apache.solr.cloud.ZkController  –
waiting to find shard id in clusterstate for reportCore_201310


[2]---
136661 [Thread-2] INFO  org.apache.solr.common.cloud.ZkStateReader  –
Updating cloud state from ZooKeeper...
136663 [Thread-2] INFO  org.apache.solr.cloud.Overseer  – Update state
numShards=null message={
  operation:state,
  state:down,
  base_url:http://10.1.22.1:8080/solr;,
  core:reportCore_201310,
  roles:null,
  node_name:10.1.22.1:8080_solr,
  shard:reportCore_201310,
  shard_range:null,
  shard_state:active,
  shard_parent:null,
  collection:repCore,
  numShards:null,
  core_node_name:reportCore_201310}
136663 [Thread-2] ERROR org.apache.solr.cloud.Overseer  – Exception in
Overseer main queue loop
java.lang.NullPointerException
at
org.apache.solr.cloud.Overseer$ClusterStateUpdater.updateState(Overseer.java:360)
at
org.apache.solr.cloud.Overseer$ClusterStateUpdater.processMessage(Overseer.java:204)
at
org.apache.solr.cloud.Overseer$ClusterStateUpdater.run(Overseer.java:166)
at java.lang.Thread.run(Thread.java:662)


Solr cores across multiple machines

2013-12-16 Thread sivaprasad
Hi,

In my project, we are doing full index on dedicated machine and the index
will be copied to other search serving machine. For this, we are copying the
data folder from indexing machine to serving machine manually. Now, we
wanted to use Solr's SWAP configuration to do this job. Looks like the SWAP
will work between the cores. Based on our setup, any one has any idea how to
move the data from indexing machine to serving machine? Is there any other
alternatives?

Regards,
Siva



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-cores-across-multiple-machines-tp4107035.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Storing MYSQL DATETIME field in solr as String

2013-12-16 Thread Raymond Wiker
To me, the obvious way of doing this would be to CAST the DATETIME to
CHAR(n), or (probably better) use DATE_FORMAT().


On Tue, Dec 17, 2013 at 5:21 AM, Shalin Shekhar Mangar 
shalinman...@gmail.com wrote:

 I'm sorry. I thought you wanted to parse a date stored as string into
 a java.util.Date. Clearly, you are trying to go the other way round.

 There's nothing in DIH which will convert a mysql date to a string in
 specific format. You will need to write a custom transformer either in
 javascript or in java to do this.

 http://wiki.apache.org/solr/DIHCustomTransformer
 http://wiki.apache.org/solr/DataImportHandler#ScriptTransformer

 On Mon, Dec 16, 2013 at 1:34 PM, manju16832003 manju16832...@gmail.com
 wrote:
  OK here are the scenarios I tried.
 
  *Scenario - 1: *
 
  dih.xml (aka data-config.xml)
 
  entity dataSource=solr name=listing query=...
  transformer=DateFormatTransformer
 
  field name=quot;publish_datequot; column=quot;publish_datequot;
  xpath=quot;/RDF/item/datequot;
  dateTimeFormat=quot;lt;b-MM-dd'T'HH:mm:ss* locale=en/
 
  schema.xml
  field name=publish_date type=date indexed=true stored=true
  multiValued=false default=NOW /
 
  It does not work. Throws an following exception
  WARN  org.apache.solr.handler.dataimport.DateFormatTransformer  – Could
 not
  parse a Date field
  java.text.ParseException: Unparseable date: 2013-12-05 15:40:03.0
 
  *Scenario - 2*
  dih.xml (aka data-config.xml)
 
  entity dataSource=solr name=listing query=...
  transformer=DateFormatTransformer
 
  field name=quot;publish_datequot; column=quot;publish_datequot;
  xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd
  HH:mm:ss* locale=en/
 
  schema.xml
  field name=publish_date type=date indexed=true stored=true
  multiValued=false default=NOW /
 
  Output is
  date name=publish_date2012-12-05T07:38:27Z/date
 
  *Scenario - 3*
  dih.xml (aka data-config.xml)
 
  entity dataSource=solr name=listing query=...
  transformer=DateFormatTransformer
 
  field name=quot;publish_datequot; column=quot;publish_datequot;
  xpath=quot;/RDF/item/datequot; dateTimeFormat=quot;lt;b-MM-dd
  HH:mm:ss* locale=en/
 
  schema.xml
  field name=publish_date type=string indexed=true stored=true
  multiValued=false default=NOW /
 
  Output is
str name=publish_dateThu Dec 05 15:38:27 MYT 2013/str
 
  My expectation was *2012-12-05 07:38:27* (MySQL DateTime format)
 
 
 
  --
  View this message in context:
 http://lucene.472066.n3.nabble.com/Storing-MYSQL-DATETIME-field-in-solr-as-String-tp4106836p4106854.html
  Sent from the Solr - User mailing list archive at Nabble.com.



 --
 Regards,
 Shalin Shekhar Mangar.



Re: External File Field

2013-12-16 Thread Mukundaraman valakumaresan
Hi

http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,locality,property_price_min,field(adjlocality)q=*:*http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,locality,property_price_min,norm:pricenorm(1000,10,property_price_median),field(adjlocality)q=*:*

I did it, it is giving me a message underfined field.

I have placed my files at F:\solr\example\solr\collection1\data where the
Solr data is maintained.

Is there a way to check whether the files are read or not?

Thanks  Regards
Mukund



On Mon, Dec 16, 2013 at 8:48 PM, Ahmet Arslan iori...@yahoo.com wrote:

 Hi,

 You can request the value of the that field via fl=*,field(adjlocality)

 See more about it :
 https://cwiki.apache.org/confluence/display/solr/Working+with+External+Files+and+Processes

 Actually you can search it too with frange query parser. {!frange l=0
 u=0}field(adjlocality)




 On Monday, December 16, 2013 5:05 PM, Mukundaraman valakumaresan 
 muk...@8kmiles.com wrote:
 Hi

 Is there a way to find ,whether the External File Fields mentioned in the
 schema.xml is being used or whether Solr reads the value of those external
 fields.

 I am not sure how to use an External field, can I request the value of an
 External File Field in the field list or can I use in my custom function as
 shown below:


 http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,adjlocality
 
 http://localhost:8983/solr/collection1/select?wt=jsonomitHeader=onindent=truedefType=edismaxfl=id,property_price_max,property_price_min,lanorm(%22%22,luxury_amenities_c,10),adjlocality
 
 ,custfunc(adjlocality).

 I didn't get any output for the external fields in the queries that I
 provide.

 Thanks  Regards
 Mukund