Re: Replication Error

2014-01-03 Thread anand chandak

Thanks Daniel.


So, if I understand correctly the below exception is almost always 
caused because of merging segments ? Though I see different file names 
(for e.g download_av3.fdt in this case)  in the exception messages


[explicit-fetchindex-cmd] ERROR 
org.apache.solr.handler.ReplicationHandler- SnapPull failed 
:org.apache.solr.common.SolrException: Unable to download _av3.fdt 
completely . Downloaded 0!=497037



Thanks,

Anand


On 1/3/2014 11:04 PM, Daniel Collins wrote:

We see this a lot as well, my understanding is that recovery asks the
leader for a list of the files that it should download, then it downloads
them.  But if the leader has been merging segments whilst this is going on
(recovery is taking a reasonable period of time and you have an NRT system
where commits/merges are reasonably frequent), then the segments might
disappear during this recovery period, hence the replica can't download
them.

So its an error, but something the system can recover from, since it will
re-recover, which should pick up a (larger) segment next time, which is
less likely to be removed whilst recovery is going on.


On 3 January 2014 05:32, anand chandak  wrote:


Hi,

I am hitting this error on replication, can somebody please tell me what's
wrong here and what can be done to correct this error :


[explicit-fetchindex-cmd] ERROR org.apache.solr.handler.ReplicationHandler-
SnapPull failed :org.apache.solr.common.SolrException: Unable to download
_av3.fdt completely. Downloaded 0!=497037
   2 at org.apache.solr.handler.SnapPuller$
DirectoryFileFetcher.cleanup(SnapPuller.java:1268)
   3 at org.apache.solr.handler.SnapPuller$DirectoryFileFetcher.
fetchFile(SnapPuller.java:1148)
   4 at org.apache.solr.handler.SnapPuller.downloadIndexFiles(
SnapPuller.java:743)
   5 at org.apache.solr.handler.SnapPuller.fetchLatestIndex(
SnapPuller.java:407)
   6 at org.apache.solr.handler.ReplicationHandler.doFetch(
ReplicationHandler.java:319)
   7 at org.apache.solr.handler.ReplicationHandler$1.run(
ReplicationHandler.java:220)

Thanks,

Anand






Re: Tracking down the input that hits an analysis chain bug

2014-01-03 Thread Michael Sokolov
Have you considered using a custom UpdateProcessor to catch the 
exception and provide more context in the logs?


-Mike

On 01/03/2014 03:33 PM, Benson Margulies wrote:

Robert,

Yes, if the problem was not data-dependent, indeed I wouldn't need to
index anything. However, I've run a small mountain of data through our
tokenizer on my machine, and never seen the error, but my customer
gets these errors in the middle of a giant spew of data. As it
happens, I _was_ missing that call to clearAttributes(), (and the
usual implementation of end()), but I found and fixed that problem
precisely by creating a random data test case using checkRandomData().
Unfortunately, fixing that didn't make the customer's errors go away.

So I'm left needing to help them identify the data that provokes this,
because I've so far failed to come up with any.

--benson


On Fri, Jan 3, 2014 at 2:16 PM, Robert Muir  wrote:

This exception comes from OffsetAttributeImpl (e.g. you dont need to
index anything to reproduce it).

Maybe you have a missing clearAttributes() call (your tokenizer
'returns true' without calling that first)? This could explain it, if
something like a StopFilter is also present in the chain: basically
the offsets overflow.

the test stuff in BaseTokenStreamTestCase should be able to detect
this as well...

On Fri, Jan 3, 2014 at 1:56 PM, Benson Margulies  wrote:

Using Solr Cloud with 4.3.1.

We've got a problem with a tokenizer that manifests as calling
OffsetAtt.setOffsets() with invalid inputs. OK, so, we want to figure out
what input provokes our code into getting into this pickle.

The problem happens on SolrCloud nodes.

The problem manifests as this sort of thing:

Jan 3, 2014 6:05:33 PM org.apache.solr.common.SolrException log
SEVERE: java.lang.IllegalArgumentException: startOffset must be
non-negative, and endOffset must be >= startOffset,
startOffset=-1811581632,endOffset=-1811581632

How could we get a document ID so that we can tell which document was being
processed?




Re: SPLITSHARD failed to invoke SPLIT core admin command

2014-01-03 Thread sriram
Thanks Anshum. Will keep track of that. 

Also the shard is in construction state for long time. I see the size of the
index folder inside the newly created cores :

solr/collection1_shard1_0_replica1
solr/collection1_shard1_1_replica1

They are of almost similar size and more than half the size of the original
core: collection1. The size of the newly created cores also became constant
after the split happened and no longer changing. But still I see the shard
status as under construction. 

As mentioned before, I tried to remove the new shard1_0 and shard1_1 and it
created core.properties.unloaded with shardState=construction. Hence I think
since my cluster is hung up in construction state for the new shards for
longer time, it is not able to delete it.

Do you have any inputs on this?

Thanks,
V.Sriram



--
View this message in context: 
http://lucene.472066.n3.nabble.com/SPLITSHARD-failed-to-invoke-SPLIT-core-admin-command-tp4109431p4109444.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SPLITSHARD failed to invoke SPLIT core admin command

2014-01-03 Thread Anshum Gupta
As far as I remember you can only delete slices which are either
1. Inactive or
2. A part of a collection that uses implicit routing.

The delete shard API would not let you delete a slice directly. However,
you can perhaps retry the command. As far as I remember, it's built to
cleanup an earlier incomplete run on subsequent calls.

As for a requestParam that allows you to specify a higher timeout value, I
don't think that's in place yet. You can follow the SOLR Jira I mentioned
in my last response to track the support of async calls in Solr which
should fix this kind of issues.


On Fri, Jan 3, 2014 at 4:27 PM, sriram  wrote:

> Thanks a lot Anshum for your quick reply. I checked the clusterstate.json
> and
> there it says the two new shards as under construction.
>
> It shows :
> shard1 : active
> shard2 : active
> shard1_1 : construction
> shard1_0 : construction
>
> But the loggers don't indicate any other progress in that. Also when I try
> to delete the newly created shards, it is not able to delete ( which I
> think
> should be the case as it is still under construction)
>
> Is there any workaround to resolve this for now? like some reqestParam
> specifying a higher timeout value?
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/SPLITSHARD-failed-to-invoke-SPLIT-core-admin-command-tp4109431p4109441.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>



-- 

Anshum Gupta
http://www.anshumgupta.net


Re: SPLITSHARD failed to invoke SPLIT core admin command

2014-01-03 Thread sriram
Thanks a lot Anshum for your quick reply. I checked the clusterstate.json and
there it says the two new shards as under construction. 

It shows : 
shard1 : active
shard2 : active
shard1_1 : construction
shard1_0 : construction

But the loggers don't indicate any other progress in that. Also when I try
to delete the newly created shards, it is not able to delete ( which I think
should be the case as it is still under construction)

Is there any workaround to resolve this for now? like some reqestParam
specifying a higher timeout value?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/SPLITSHARD-failed-to-invoke-SPLIT-core-admin-command-tp4109431p4109441.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: how to include result ordinal in response

2014-01-03 Thread Jack Krupansky

Or just use the internal document ID: fl=*,[docid]

Granted, the docID may change if a segment merge occurs and earlier 
documents have been deleted, but it may be sufficient for your purposes.


-- Jack Krupansky

-Original Message- 
From: Upayavira

Sent: Friday, January 03, 2014 5:58 PM
To: solr-user@lucene.apache.org
Subject: Re: how to include result ordinal in response

On Fri, Jan 3, 2014, at 10:00 PM, Peter Keegan wrote:

Is there a simple way to output the result number (ordinal) with each
returned document using the 'fl' parameter? This would be useful when
visually comparing the results from 2 queries.


I'm not aware of a simple way.

If you're competent in Java, this could be a neat new DocTransformer
component. You'd say:

fl=*,[position]

and you'd get a new field in your search results.

Cruder ways would be to use XSLT to add it to an XML output, or a
velocity template, but the DocTransformer approach would create
something that could be of ongoing use.

Upayavira 



Re: SPLITSHARD failed to invoke SPLIT core admin command

2014-01-03 Thread Anshum Gupta
Also, here's another issue that would help with the timeouts
(SOLR-5477
).


On Fri, Jan 3, 2014 at 3:34 PM, Anshum Gupta  wrote:

> Hi Sriram,
>
> People hit timeout issues when trying to split larger collections/shards.
> This is because of the timing out of internal calls.
>
> About the UI showing everything as active, that's another open issue(
> SOLR-4777 ). You should
> instead check the clusterstate to see the state of the slices (both parent
> and child).
> What is the current state of the slices?
>
>
>
> On Fri, Jan 3, 2014 at 2:39 PM, sriram  wrote:
>
>> Hi,
>>
>> I am using solr cloud 4.6 with 2 shards. I have loaded close to 13Million
>> data in them. When I tried to give split shard, I am getting the following
>> errors in the log,
>>
>> INFO  - 2014-01-03 14:17:45.357;
>> org.apache.solr.cloud.DistributedQueue$LatchChildWatcher; Watcher fired on
>> path: /overseer/collection-queue-work/qnr-02 state: SyncConnected
>> type NodeDataChanged
>> INFO  - 2014-01-03 14:17:45.359;
>> org.apache.solr.cloud.DistributedQueue$LatchChildWatcher; Watcher fired on
>> path: /overseer/collection-queue-work state: SyncConnected type
>> NodeChildrenChanged
>> INFO  - 2014-01-03 14:17:45.360;
>> org.apache.solr.cloud.OverseerCollectionProcessor; Overseer Collection
>> Processor: Message id:/overseer/collection-queue-work/qn-02
>> complete,
>>
>> response:{success={null={responseHeader={status=0,QTime=3689},core=collection1_shard1_1_replica1},null={responseHeader={status=0,QTime=3694},core=collection1_shard1_0_replica1},null={responseHeader={status=0,QTime=2004}},null={responseHeader={status=0,QTime=2005}}},failure={null=org.apache.solr.client.solrj.SolrServerException:Timeout
>> occured while waiting response from server at:
>> http:///solr},Operation splitshard caused
>> exception:=org.apache.solr.common.SolrException: SPLITSHARD failed to
>> invoke
>> SPLIT core admin command,exception={msg=SPLITSHARD failed to invoke SPLIT
>> core admin command,rspCode=500}}
>> INFO  - 2014-01-03 14:17:45.365;
>> org.apache.solr.servlet.SolrDispatchFilter;
>> [admin] webapp=null path=/admin/collections
>> params={shard=shard1&action=SPLITSHARD&collection=collection1} status=500
>> QTime=15736
>> ERROR - 2014-01-03 14:17:45.366; org.apache.solr.common.SolrException;
>> null:org.apache.solr.common.SolrException: SPLITSHARD failed to invoke
>> SPLIT
>> core admin command
>> at
>>
>> org.apache.solr.handler.admin.CollectionsHandler.handleResponse(CollectionsHandler.java:200)
>> at
>>
>> org.apache.solr.handler.admin.CollectionsHandler.handleSplitShardAction(CollectionsHandler.java:422)
>> at
>>
>> org.apache.solr.handler.admin.CollectionsHandler.handleRequestBody(CollectionsHandler.java:158)
>> at
>>
>> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
>> at
>>
>> org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:662)
>> at
>>
>> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:257)
>> at
>>
>> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)
>> at
>>
>> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
>> at
>> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
>> at
>>
>> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
>> at
>>
>> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
>> at
>>
>> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
>> at
>>
>> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
>> at
>> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
>> at
>>
>> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
>> at
>>
>> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
>> at
>>
>> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
>> at
>>
>> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
>> at
>>
>> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
>> at
>>
>> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
>> at org.eclipse.jetty.server.Server.handle(Server.java:368)
>> at
>>
>> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
>> at
>>
>> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
>> at
>>
>> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java

Re: SPLITSHARD failed to invoke SPLIT core admin command

2014-01-03 Thread Anshum Gupta
Hi Sriram,

People hit timeout issues when trying to split larger collections/shards.
This is because of the timing out of internal calls.

About the UI showing everything as active, that's another open issue(
SOLR-4777 ). You should
instead check the clusterstate to see the state of the slices (both parent
and child).
What is the current state of the slices?



On Fri, Jan 3, 2014 at 2:39 PM, sriram  wrote:

> Hi,
>
> I am using solr cloud 4.6 with 2 shards. I have loaded close to 13Million
> data in them. When I tried to give split shard, I am getting the following
> errors in the log,
>
> INFO  - 2014-01-03 14:17:45.357;
> org.apache.solr.cloud.DistributedQueue$LatchChildWatcher; Watcher fired on
> path: /overseer/collection-queue-work/qnr-02 state: SyncConnected
> type NodeDataChanged
> INFO  - 2014-01-03 14:17:45.359;
> org.apache.solr.cloud.DistributedQueue$LatchChildWatcher; Watcher fired on
> path: /overseer/collection-queue-work state: SyncConnected type
> NodeChildrenChanged
> INFO  - 2014-01-03 14:17:45.360;
> org.apache.solr.cloud.OverseerCollectionProcessor; Overseer Collection
> Processor: Message id:/overseer/collection-queue-work/qn-02
> complete,
>
> response:{success={null={responseHeader={status=0,QTime=3689},core=collection1_shard1_1_replica1},null={responseHeader={status=0,QTime=3694},core=collection1_shard1_0_replica1},null={responseHeader={status=0,QTime=2004}},null={responseHeader={status=0,QTime=2005}}},failure={null=org.apache.solr.client.solrj.SolrServerException:Timeout
> occured while waiting response from server at:
> http:///solr},Operation splitshard caused
> exception:=org.apache.solr.common.SolrException: SPLITSHARD failed to
> invoke
> SPLIT core admin command,exception={msg=SPLITSHARD failed to invoke SPLIT
> core admin command,rspCode=500}}
> INFO  - 2014-01-03 14:17:45.365;
> org.apache.solr.servlet.SolrDispatchFilter;
> [admin] webapp=null path=/admin/collections
> params={shard=shard1&action=SPLITSHARD&collection=collection1} status=500
> QTime=15736
> ERROR - 2014-01-03 14:17:45.366; org.apache.solr.common.SolrException;
> null:org.apache.solr.common.SolrException: SPLITSHARD failed to invoke
> SPLIT
> core admin command
> at
>
> org.apache.solr.handler.admin.CollectionsHandler.handleResponse(CollectionsHandler.java:200)
> at
>
> org.apache.solr.handler.admin.CollectionsHandler.handleSplitShardAction(CollectionsHandler.java:422)
> at
>
> org.apache.solr.handler.admin.CollectionsHandler.handleRequestBody(CollectionsHandler.java:158)
> at
>
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:662)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:257)
> at
>
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)
> at
>
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
> at
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
> at
>
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
> at
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
> at
>
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
> at
>
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
> at
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
> at
>
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
> at
>
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
> at
>
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
> at
>
> org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
> at
>
> org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
> at
>
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
> at org.eclipse.jetty.server.Server.handle(Server.java:368)
> at
>
> org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
> at
>
> org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
> at
>
> org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942)
> at
>
> org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004)
> at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640)
> at
> org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
> a

Re: Exact match on KeywordTokenizer

2014-01-03 Thread Chris Hostetter

Can you show us examples of the types of data you are indexing, and the 
types of queries you want to match? (as well as examples of queries you 
*don't* want to match)

https://wiki.apache.org/solr/UsingMailingLists#Information_useful_for_searching_problems

Best guess, based on your problem description, is that you are indexing 
text like "Foo Bar" and then searching for things like "foOBaR" and 
you want those to match.

With your analyzer as it is, you will never get a match unless the client 
sending hte query string has already lowercased it, done any asciifolding 
needed, and always sends "-" instead of space characters.

i suspect what you really want is to have index & query analyzers that are 
the same (or at least better matches for exachother then what you have 
below)...


: Hi,
: 
: Is there a way to do an exact match search on a tokenized field?
: 
: I have a scenario which i need a field to be indexed and searchable
: regardless of the case or white spaces used. For this, I created a custom
: field type with the following configuration:
: 
: 
: 
: 
: 
: 
: 
: 
: 
:   
:   
: 
:   
: 
: 
: Even using KeywordTokenizerFactory on both index and query, all my searchs
: based on exact match stopped working.
: 
: Is there a way to search exact match like a string field and at the same
: time use customs tokenizers aplied to that field?
: 
: Thank's in advance
: 
: 
: 
*--*
: *"E conhecereis a verdade, e a verdade vos libertará." (João 8:32)*
: 
:  *andre.maldonado*@gmail.com 
:  (11) 9112-4227
: 
: 
: 
: 
:   
: 
:   
: 

-Hoss
http://www.lucidworks.com/

Re: how to include result ordinal in response

2014-01-03 Thread Upayavira
On Fri, Jan 3, 2014, at 10:00 PM, Peter Keegan wrote:
> Is there a simple way to output the result number (ordinal) with each
> returned document using the 'fl' parameter? This would be useful when
> visually comparing the results from 2 queries.

I'm not aware of a simple way.

If you're competent in Java, this could be a neat new DocTransformer
component. You'd say:

fl=*,[position]

and you'd get a new field in your search results.

Cruder ways would be to use XSLT to add it to an XML output, or a
velocity template, but the DocTransformer approach would create
something that could be of ongoing use.

Upayavira


SPLITSHARD failed to invoke SPLIT core admin command

2014-01-03 Thread sriram
Hi,

I am using solr cloud 4.6 with 2 shards. I have loaded close to 13Million
data in them. When I tried to give split shard, I am getting the following
errors in the log,

INFO  - 2014-01-03 14:17:45.357;
org.apache.solr.cloud.DistributedQueue$LatchChildWatcher; Watcher fired on
path: /overseer/collection-queue-work/qnr-02 state: SyncConnected
type NodeDataChanged
INFO  - 2014-01-03 14:17:45.359;
org.apache.solr.cloud.DistributedQueue$LatchChildWatcher; Watcher fired on
path: /overseer/collection-queue-work state: SyncConnected type
NodeChildrenChanged
INFO  - 2014-01-03 14:17:45.360;
org.apache.solr.cloud.OverseerCollectionProcessor; Overseer Collection
Processor: Message id:/overseer/collection-queue-work/qn-02
complete,
response:{success={null={responseHeader={status=0,QTime=3689},core=collection1_shard1_1_replica1},null={responseHeader={status=0,QTime=3694},core=collection1_shard1_0_replica1},null={responseHeader={status=0,QTime=2004}},null={responseHeader={status=0,QTime=2005}}},failure={null=org.apache.solr.client.solrj.SolrServerException:Timeout
occured while waiting response from server at:
http:///solr},Operation splitshard caused
exception:=org.apache.solr.common.SolrException: SPLITSHARD failed to invoke
SPLIT core admin command,exception={msg=SPLITSHARD failed to invoke SPLIT
core admin command,rspCode=500}}
INFO  - 2014-01-03 14:17:45.365; org.apache.solr.servlet.SolrDispatchFilter;
[admin] webapp=null path=/admin/collections
params={shard=shard1&action=SPLITSHARD&collection=collection1} status=500
QTime=15736
ERROR - 2014-01-03 14:17:45.366; org.apache.solr.common.SolrException;
null:org.apache.solr.common.SolrException: SPLITSHARD failed to invoke SPLIT
core admin command
at
org.apache.solr.handler.admin.CollectionsHandler.handleResponse(CollectionsHandler.java:200)
at
org.apache.solr.handler.admin.CollectionsHandler.handleSplitShardAction(CollectionsHandler.java:422)
at
org.apache.solr.handler.admin.CollectionsHandler.handleRequestBody(CollectionsHandler.java:158)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at
org.apache.solr.servlet.SolrDispatchFilter.handleAdminRequest(SolrDispatchFilter.java:662)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:257)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:197)
at
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1419)
at
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:455)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
at
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
at
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
at
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1075)
at
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:384)
at
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
at
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1009)
at
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
at
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:255)
at
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:154)
at
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
at org.eclipse.jetty.server.Server.handle(Server.java:368)
at
org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:489)
at
org.eclipse.jetty.server.BlockingHttpConnection.handleRequest(BlockingHttpConnection.java:53)
at
org.eclipse.jetty.server.AbstractHttpConnection.headerComplete(AbstractHttpConnection.java:942)
at
org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.headerComplete(AbstractHttpConnection.java:1004)
at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:640)
at
org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:235)
at
org.eclipse.jetty.server.BlockingHttpConnection.handle(BlockingHttpConnection.java:72)
at
org.eclipse.jetty.server.bio.SocketConnector$ConnectorEndPoint.run(SocketConnector.java:264)
at
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
at
org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
at java.lang.Thread.run(Thread.java:619)


Before this exception, I am seeing couple of timeouts : 

ERROR - 2014-01-03 14:17:45.350; org.apache.solr.common.SolrException;
Collection splitshard of splitshard
failed:org.apache.solr.common.SolrException: SPLITSHARD failed to invoke
SPLIT core ad

how to include result ordinal in response

2014-01-03 Thread Peter Keegan
Is there a simple way to output the result number (ordinal) with each
returned document using the 'fl' parameter? This would be useful when
visually comparing the results from 2 queries.

Thanks,
Peter


Re: SSLinitializationException after SOLR upgrade

2014-01-03 Thread Chris Hostetter

: We recently ungraded from solr 3.5 to solr4.4 . After the upgrade we are
: facing the following issue. Could anyone please suggest on what could be
: done.

You are aparently running Solr code from a JVM configured to use an SSL 
keystore which has a password configured on it, and you either have not 
specified hte password when running the JVM, or you are attempting ot use 
the wrong keystore password...

:  Caused by: java.io.IOException: Keystore was tampered with, or password
: was incorrect
...
: Caused by: java.security.UnrecoverableKeyException: Password verification
: failed

You are seeing this after upgrading because Solr (as of 4.2) now uses 
HttpClient in a way that respects system default HTTP settings (and you 
evidently have a setting fo the keystore but your corrisponding setting 
for the keystore password is wrong/missing)...

https://hc.apache.org/httpcomponents-client-ga/httpclient/apidocs/org/apache/http/impl/client/SystemDefaultHttpClient.html




-Hoss
http://www.lucidworks.com/


Re: Tracking down the input that hits an analysis chain bug

2014-01-03 Thread Benson Margulies
Robert,

Yes, if the problem was not data-dependent, indeed I wouldn't need to
index anything. However, I've run a small mountain of data through our
tokenizer on my machine, and never seen the error, but my customer
gets these errors in the middle of a giant spew of data. As it
happens, I _was_ missing that call to clearAttributes(), (and the
usual implementation of end()), but I found and fixed that problem
precisely by creating a random data test case using checkRandomData().
Unfortunately, fixing that didn't make the customer's errors go away.

So I'm left needing to help them identify the data that provokes this,
because I've so far failed to come up with any.

--benson


On Fri, Jan 3, 2014 at 2:16 PM, Robert Muir  wrote:
> This exception comes from OffsetAttributeImpl (e.g. you dont need to
> index anything to reproduce it).
>
> Maybe you have a missing clearAttributes() call (your tokenizer
> 'returns true' without calling that first)? This could explain it, if
> something like a StopFilter is also present in the chain: basically
> the offsets overflow.
>
> the test stuff in BaseTokenStreamTestCase should be able to detect
> this as well...
>
> On Fri, Jan 3, 2014 at 1:56 PM, Benson Margulies  wrote:
>> Using Solr Cloud with 4.3.1.
>>
>> We've got a problem with a tokenizer that manifests as calling
>> OffsetAtt.setOffsets() with invalid inputs. OK, so, we want to figure out
>> what input provokes our code into getting into this pickle.
>>
>> The problem happens on SolrCloud nodes.
>>
>> The problem manifests as this sort of thing:
>>
>> Jan 3, 2014 6:05:33 PM org.apache.solr.common.SolrException log
>> SEVERE: java.lang.IllegalArgumentException: startOffset must be
>> non-negative, and endOffset must be >= startOffset,
>> startOffset=-1811581632,endOffset=-1811581632
>>
>> How could we get a document ID so that we can tell which document was being
>> processed?


Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Nice to hear you (not me) solved the problem. You're welcome

Andrea
On 3 Jan 2014 21:19, "PeterKerk"  wrote:

> No need, you solved it!
> It was the id name, it had to be uppercase.
>
> btw the ; is still there in the query, but everything still works.
>
> Thanks!
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109425.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>


Re: Empty facets on Solr with MySQL

2014-01-03 Thread PeterKerk
No need, you solved it!
It was the id name, it had to be uppercase.

btw the ; is still there in the query, but everything still works.

Thanks!



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109425.html
Sent from the Solr - User mailing list archive at Nabble.com.


SSLinitializationException after SOLR upgrade

2014-01-03 Thread lakshmi guruparan
Hi All,

We recently ungraded from solr 3.5 to solr4.4 . After the upgrade we are
facing the following issue. Could anyone please suggest on what could be
done.

INFO [2013-12-13 07:54:50,149] [threadPoolTaskExecutor-2] [HttpClientUtil]
[HttpClientUtil.java:103] - Creating new http client,
config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
org.apache.http.conn.ssl.SSLInitializationException: Failure initializing
default system SSL context
at
org.apache.http.conn.ssl.SSLSocketFactory.createSystemSSLContext(SSLSocketFactory.java:368)
at
org.apache.http.conn.ssl.SSLSocketFactory.getSystemSocketFactory(SSLSocketFactory.java:204)
at
org.apache.http.impl.conn.SchemeRegistryFactory.createSystemDefault(SchemeRegistryFactory.java:82)
at
org.apache.http.impl.client.SystemDefaultHttpClient.createClientConnectionManager(SystemDefaultHttpClient.java:118)
at
org.apache.http.impl.client.AbstractHttpClient.getConnectionManager(AbstractHttpClient.java:466)
at
org.apache.solr.client.solrj.impl.HttpClientUtil.setMaxConnections(HttpClientUtil.java:179)
at
org.apache.solr.client.solrj.impl.HttpClientConfigurer.configure(HttpClientConfigurer.java:33)
at
org.apache.solr.client.solrj.impl.HttpClientUtil.configureClient(HttpClientUtil.java:115)
at
org.apache.solr.client.solrj.impl.HttpClientUtil.createClient(HttpClientUtil.java:105)
at
org.apache.solr.client.solrj.impl.HttpSolrServer.(HttpSolrServer.java:154)
at
org.apache.solr.client.solrj.impl.HttpSolrServer.(HttpSolrServer.java:127)
at
com.cisco.npa.solr.SolrService.getSolrService(SolrService.java:47)
at
com.cisco.npa.excel.ExcelTask.generateExcelDoc(ExcelTask.java:109)
at com.cisco.npa.excel.ExcelTask.run(ExcelTask.java:75)
at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)

 Caused by: java.io.IOException: Keystore was tampered with, or password
was incorrect
at
sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:771)
at
sun.security.provider.JavaKeyStore$JKS.engineLoad(JavaKeyStore.java:38)
at java.security.KeyStore.load(KeyStore.java:1185)
at
org.apache.http.conn.ssl.SSLSocketFactory.createSystemSSLContext(SSLSocketFactory.java:281)
at
org.apache.http.conn.ssl.SSLSocketFactory.createSystemSSLContext(SSLSocketFactory.java:366)
... 16 more
Caused by: java.security.UnrecoverableKeyException: Password verification
failed
at
sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:769)
... 20 more

Thanks,
Lakshmi


Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
I don't remember your dih-config.xml (could you post it again?)

- remove the trailing ; from the query. It is a valid delimiter only when
you run queries in mysql worlbench;

- I assume there's a parent entity named (name=) article.

- are you sure the column of the article entity is "id" (NB this time case
sensitive)?

Andrea
On 3 Jan 2014 19:33, "PeterKerk"  wrote:

> Hi Andrea,
>
> I think you helped me to get closer, but not quite there yet.
>
> When I replace wtr.object_id='${article.id}'; with wtr.object_id=18
> the cat_name field holds a value, which I checked via the schema browser of
> Solr dashboard!
>
> I then checked my main query SELECT p.*, ( SELECT guid FROM wp_posts WHERE
> id = m.meta_value ) AS imgurl
> FROM wp_posts p
> LEFT JOIN  wp_postmeta m ON(p.id = m.post_id AND m.meta_key =
> '_thumbnail_id' )
> WHERE p.post_type =  'post'
> AND p.post_status =  'publish';
>
> which returns 4 results. For each of these results I checked whether the
> direct query on the database returns a cat_name and it does. So, no null
> values there.
>
>
> When I remove the quotes around the ID like so
>
> 
>
> I get the errors:
>
> Caused by: org.apache.solr.handler.dataimport.DataImportHandlerException:
> Unable
>  to execute query: select wt.name as
> cat_name,wt.slug,wtr.term_taxonomy_id,wtt.t
> erm_id,wtt.taxonomy from  wp_term_relationships wtr INNER JOIN
> wp_term_taxonomy
> wtt ON wtt.term_taxonomy_id=wtr.term_taxonomy_id AND
> wtt.taxonomy='category'
> INN
> ER JOIN wp_terms wt ON wt.term_id=wtt.term_taxonomy_id where
> wtr.object_id=;
> Pro
> cessing Document # 1
> at
> org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAnd
> Throw(DataImportHandlerException.java:71)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.<
> init>(JdbcDataSource.java:253)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSou
> rce.java:210)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSou
> rce.java:38)
> at
> org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEn
> tityProcessor.java:59)
> at
> org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEnti
> tyProcessor.java:73)
> at
> org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(Ent
> ityProcessorWrapper.java:243)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
> r.java:465)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
> r.java:491)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
> r.java:404)
> ... 5 more
> Caused by: com.mysql.jdbc.exceptions.jdbc4.MySQLSyntaxErrorException: You
> have a
> n error in your SQL syntax; check the manual that corresponds to your MySQL
> serv
> er version for the right syntax to use near '' at line 1
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>
> at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
> Source)
>
> at
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
> Sou
> rce)
> at java.lang.reflect.Constructor.newInstance(Unknown Source)
> at com.mysql.jdbc.Util.handleNewInstance(Util.java:411)
> at com.mysql.jdbc.Util.getInstance(Util.java:386)
> at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:1054)
> at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4237)
> at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
> at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
> at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
> at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2819)
> at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2768)
> at com.mysql.jdbc.StatementImpl.execute(StatementImpl.java:894)
> at com.mysql.jdbc.StatementImpl.execute(StatementImpl.java:732)
> at
> org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.<
> init>(JdbcDataSource.java:246)
> ... 13 more
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109398.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>


RE: Indexing URLs for Binaries

2014-01-03 Thread Teague James
Thanks, Mark. I checked there, but pdf files are not listed. There are some
file types in there that I might need in the future, so I appreciate the
info. Any other ideas?

-Original Message-
From: Reyes, Mark 
Sent: Friday, January 03, 2014 1:39 PM
To: solr-user@lucene.apache.org
Subject: Re: Indexing URLs for Binaries

Check suffix-urlfilter.txt in your conf directory for Nutch. You might be
prohibiting those filetypes from the crawl.

- Mark






On 1/3/14, 10:29 AM, "Teague James"  wrote:

>I am using Nutch 1.7 with Solr 4.6.0 to index websites that have links 
>to binary files, such as Word, PDF, etc. The crawler crawls the site 
>but I am not getting the URLs of the links for the binary files no 
>matter how deep I set the settings for the site. I see the labels for 
>the links in the content, but not the URLs. Any ideas on how I could 
>get those URLs back in my crawl?
>


IMPORTANT NOTICE: This e-mail message is intended to be received only by
persons entitled to receive the confidential information it may contain.
E-mail messages sent from Bridgepoint Education may contain information that
is confidential and may be legally privileged. Please do not read, copy,
forward or store this message unless you are an intended recipient of it. If
you received this transmission in error, please notify the sender by reply
e-mail and delete the message and any attachments.=



Re: Tracking down the input that hits an analysis chain bug

2014-01-03 Thread Robert Muir
This exception comes from OffsetAttributeImpl (e.g. you dont need to
index anything to reproduce it).

Maybe you have a missing clearAttributes() call (your tokenizer
'returns true' without calling that first)? This could explain it, if
something like a StopFilter is also present in the chain: basically
the offsets overflow.

the test stuff in BaseTokenStreamTestCase should be able to detect
this as well...

On Fri, Jan 3, 2014 at 1:56 PM, Benson Margulies  wrote:
> Using Solr Cloud with 4.3.1.
>
> We've got a problem with a tokenizer that manifests as calling
> OffsetAtt.setOffsets() with invalid inputs. OK, so, we want to figure out
> what input provokes our code into getting into this pickle.
>
> The problem happens on SolrCloud nodes.
>
> The problem manifests as this sort of thing:
>
> Jan 3, 2014 6:05:33 PM org.apache.solr.common.SolrException log
> SEVERE: java.lang.IllegalArgumentException: startOffset must be
> non-negative, and endOffset must be >= startOffset,
> startOffset=-1811581632,endOffset=-1811581632
>
> How could we get a document ID so that we can tell which document was being
> processed?


Tracking down the input that hits an analysis chain bug

2014-01-03 Thread Benson Margulies
Using Solr Cloud with 4.3.1.

We've got a problem with a tokenizer that manifests as calling
OffsetAtt.setOffsets() with invalid inputs. OK, so, we want to figure out
what input provokes our code into getting into this pickle.

The problem happens on SolrCloud nodes.

The problem manifests as this sort of thing:

Jan 3, 2014 6:05:33 PM org.apache.solr.common.SolrException log
SEVERE: java.lang.IllegalArgumentException: startOffset must be
non-negative, and endOffset must be >= startOffset,
startOffset=-1811581632,endOffset=-1811581632

How could we get a document ID so that we can tell which document was being
processed?


Re: Replication Error

2014-01-03 Thread Shawn Heisey

On 1/3/2014 10:34 AM, Daniel Collins wrote:

We see this a lot as well, my understanding is that recovery asks the
leader for a list of the files that it should download, then it downloads
them.  But if the leader has been merging segments whilst this is going on
(recovery is taking a reasonable period of time and you have an NRT system
where commits/merges are reasonably frequent), then the segments might
disappear during this recovery period, hence the replica can't download
them.

So its an error, but something the system can recover from, since it will
re-recover, which should pick up a (larger) segment next time, which is
less likely to be removed whilst recovery is going on.


This can probably be fixed by increasing the commitReserveDuration 
setting in the master replication configuration.  In the context of 
replication, this controls the amount of time that Solr (Lucene really) 
will hold on to segments that are slated for deletion.  It defaults to 
10 seconds, but when you have a lot of data to replicate, it can take 
considerably longer than 10 seconds for a replication to finish.


http://wiki.apache.org/solr/SolrReplication#Master

Thanks,
Shawn



Re: Indexing URLs for Binaries

2014-01-03 Thread Reyes, Mark
Check suffix-urlfilter.txt in your conf directory for Nutch. You might be
prohibiting those filetypes from the crawl.

- Mark






On 1/3/14, 10:29 AM, "Teague James"  wrote:

>I am using Nutch 1.7 with Solr 4.6.0 to index websites that have links to
>binary files, such as Word, PDF, etc. The crawler crawls the site but I am
>not getting the URLs of the links for the binary files no matter how deep
>I
>set the settings for the site. I see the labels for the links in the
>content, but not the URLs. Any ideas on how I could get those URLs back in
>my crawl?
>


IMPORTANT NOTICE: This e-mail message is intended to be received only by 
persons entitled to receive the confidential information it may contain. E-mail 
messages sent from Bridgepoint Education may contain information that is 
confidential and may be legally privileged. Please do not read, copy, forward 
or store this message unless you are an intended recipient of it. If you 
received this transmission in error, please notify the sender by reply e-mail 
and delete the message and any attachments.

Re: Empty facets on Solr with MySQL

2014-01-03 Thread PeterKerk
Hi Andrea,

I think you helped me to get closer, but not quite there yet.

When I replace wtr.object_id='${article.id}'; with wtr.object_id=18 
the cat_name field holds a value, which I checked via the schema browser of
Solr dashboard!

I then checked my main query SELECT p.*, ( SELECT guid FROM wp_posts WHERE
id = m.meta_value ) AS imgurl 
FROM wp_posts p
LEFT JOIN  wp_postmeta m ON(p.id = m.post_id AND m.meta_key = 
'_thumbnail_id' )
WHERE p.post_type =  'post'
AND p.post_status =  'publish';

which returns 4 results. For each of these results I checked whether the
direct query on the database returns a cat_name and it does. So, no null
values there.


When I remove the quotes around the ID like so



I get the errors:

Caused by: org.apache.solr.handler.dataimport.DataImportHandlerException:
Unable
 to execute query: select wt.name as
cat_name,wt.slug,wtr.term_taxonomy_id,wtt.t
erm_id,wtt.taxonomy from  wp_term_relationships wtr INNER JOIN
wp_term_taxonomy
wtt ON wtt.term_taxonomy_id=wtr.term_taxonomy_id AND wtt.taxonomy='category'
INN
ER JOIN wp_terms wt ON wt.term_id=wtt.term_taxonomy_id where wtr.object_id=;
Pro
cessing Document # 1
at
org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAnd
Throw(DataImportHandlerException.java:71)
at
org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.<
init>(JdbcDataSource.java:253)
at
org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSou
rce.java:210)
at
org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSou
rce.java:38)
at
org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEn
tityProcessor.java:59)
at
org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEnti
tyProcessor.java:73)
at
org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(Ent
ityProcessorWrapper.java:243)
at
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
r.java:465)
at
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
r.java:491)
at
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
r.java:404)
... 5 more
Caused by: com.mysql.jdbc.exceptions.jdbc4.MySQLSyntaxErrorException: You
have a
n error in your SQL syntax; check the manual that corresponds to your MySQL
serv
er version for the right syntax to use near '' at line 1
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method)

at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
Source)

at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
Sou
rce)
at java.lang.reflect.Constructor.newInstance(Unknown Source)
at com.mysql.jdbc.Util.handleNewInstance(Util.java:411)
at com.mysql.jdbc.Util.getInstance(Util.java:386)
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:1054)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4237)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4169)
at com.mysql.jdbc.MysqlIO.sendCommand(MysqlIO.java:2617)
at com.mysql.jdbc.MysqlIO.sqlQueryDirect(MysqlIO.java:2778)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2819)
at com.mysql.jdbc.ConnectionImpl.execSQL(ConnectionImpl.java:2768)
at com.mysql.jdbc.StatementImpl.execute(StatementImpl.java:894)
at com.mysql.jdbc.StatementImpl.execute(StatementImpl.java:732)
at
org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.<
init>(JdbcDataSource.java:246)
... 13 more



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109398.html
Sent from the Solr - User mailing list archive at Nabble.com.


Indexing URLs for Binaries

2014-01-03 Thread Teague James
I am using Nutch 1.7 with Solr 4.6.0 to index websites that have links to
binary files, such as Word, PDF, etc. The crawler crawls the site but I am
not getting the URLs of the links for the binary files no matter how deep I
set the settings for the site. I see the labels for the links in the
content, but not the URLs. Any ideas on how I could get those URLs back in
my crawl?



Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Or, in other words, assuming that the example you reported (article.id=18)
returns a record with cat_name, could you hardcode that id in your
dih-config (replacing ${article.id} with 18) and try to run a full-import?

Another important thing: in your example you are using the object_id as an
int (object_id=18) but not in dih-config.xml where you are using quotes
(object_id=' ${article.id}')although jdbc drivers often handle this
conversion magically this is not very good. Is object_id a string or a
number?
On 3 Jan 2014 18:53, "Andrea Gazzarini"  wrote:

> Yes, that definitely matters...I was talking about a simple java class.
>
> Peter, sincerely? I don't know.
>
> Are you sure you are doing then same queries with the same article.id as
> solr does? and sorry for the stupid question but are you sure that for
> those article ids the cat_name is not null?
>  On 3 Jan 2014 18:37, "PeterKerk"  wrote:
>
>> But when I execute the query directly on MySQL I do get a cat_name column
>> in
>> there:
>>
>> select wt.name as
>> cat_name,wt.slug,wtr.term_taxonomy_id,wtt.term_id,wtt.taxonomy from
>> wp_term_relationships wtr
>> INNER JOIN wp_term_taxonomy wtt ON
>> wtt.term_taxonomy_id=wtr.term_taxonomy_id
>> AND wtt.taxonomy='category'
>> INNER JOIN wp_terms wt ON wt.term_id=wtt.term_taxonomy_id
>> where wtr.object_id=18
>>
>> I see no reason why my configuration in my data-config.xml would not
>> execute
>> successfully:
>>
>> 
>>
>>
>> 
>>
>>
>> I have no transformers on my resultset (I checked my querystring,
>> schema.xml
>> and data-config.xml, since I'm not even sure where it would have to be
>> defined).
>>
>>
>>
>> Andrea Gazzarini-4 wrote
>> > You can debug the resultset in a main class by doing rs.getString
>> > ("cat_name")
>>
>> What do you mean by 'in a main class'? Where can I define that? (ps. I'm
>> working with ASP.NET if that matters)
>>
>> Thanks again! :)
>>
>>
>>
>> --
>> View this message in context:
>> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109388.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>


Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Yes, that definitely matters...I was talking about a simple java class.

Peter, sincerely? I don't know.

Are you sure you are doing then same queries with the same article.id as
solr does? and sorry for the stupid question but are you sure that for
those article ids the cat_name is not null?
 On 3 Jan 2014 18:37, "PeterKerk"  wrote:

> But when I execute the query directly on MySQL I do get a cat_name column
> in
> there:
>
> select wt.name as
> cat_name,wt.slug,wtr.term_taxonomy_id,wtt.term_id,wtt.taxonomy from
> wp_term_relationships wtr
> INNER JOIN wp_term_taxonomy wtt ON
> wtt.term_taxonomy_id=wtr.term_taxonomy_id
> AND wtt.taxonomy='category'
> INNER JOIN wp_terms wt ON wt.term_id=wtt.term_taxonomy_id
> where wtr.object_id=18
>
> I see no reason why my configuration in my data-config.xml would not
> execute
> successfully:
>
> 
>
>
> 
>
>
> I have no transformers on my resultset (I checked my querystring,
> schema.xml
> and data-config.xml, since I'm not even sure where it would have to be
> defined).
>
>
>
> Andrea Gazzarini-4 wrote
> > You can debug the resultset in a main class by doing rs.getString
> > ("cat_name")
>
> What do you mean by 'in a main class'? Where can I define that? (ps. I'm
> working with ASP.NET if that matters)
>
> Thanks again! :)
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109388.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>


Re: Empty facets on Solr with MySQL

2014-01-03 Thread PeterKerk
But when I execute the query directly on MySQL I do get a cat_name column in
there:

select wt.name as
cat_name,wt.slug,wtr.term_taxonomy_id,wtt.term_id,wtt.taxonomy from 
wp_term_relationships wtr
INNER JOIN wp_term_taxonomy wtt ON wtt.term_taxonomy_id=wtr.term_taxonomy_id
AND wtt.taxonomy='category'
INNER JOIN wp_terms wt ON wt.term_id=wtt.term_taxonomy_id
where wtr.object_id=18

I see no reason why my configuration in my data-config.xml would not execute
successfully:




   


I have no transformers on my resultset (I checked my querystring, schema.xml
and data-config.xml, since I'm not even sure where it would have to be
defined).



Andrea Gazzarini-4 wrote
> You can debug the resultset in a main class by doing rs.getString 
> ("cat_name") 

What do you mean by 'in a main class'? Where can I define that? (ps. I'm
working with ASP.NET if that matters)

Thanks again! :)



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109388.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Replication Error

2014-01-03 Thread Daniel Collins
We see this a lot as well, my understanding is that recovery asks the
leader for a list of the files that it should download, then it downloads
them.  But if the leader has been merging segments whilst this is going on
(recovery is taking a reasonable period of time and you have an NRT system
where commits/merges are reasonably frequent), then the segments might
disappear during this recovery period, hence the replica can't download
them.

So its an error, but something the system can recover from, since it will
re-recover, which should pick up a (larger) segment next time, which is
less likely to be removed whilst recovery is going on.


On 3 January 2014 05:32, anand chandak  wrote:

> Hi,
>
> I am hitting this error on replication, can somebody please tell me what's
> wrong here and what can be done to correct this error :
>
>
> [explicit-fetchindex-cmd] ERROR org.apache.solr.handler.ReplicationHandler-
> SnapPull failed :org.apache.solr.common.SolrException: Unable to download
> _av3.fdt completely. Downloaded 0!=497037
>   2 at org.apache.solr.handler.SnapPuller$
> DirectoryFileFetcher.cleanup(SnapPuller.java:1268)
>   3 at org.apache.solr.handler.SnapPuller$DirectoryFileFetcher.
> fetchFile(SnapPuller.java:1148)
>   4 at org.apache.solr.handler.SnapPuller.downloadIndexFiles(
> SnapPuller.java:743)
>   5 at org.apache.solr.handler.SnapPuller.fetchLatestIndex(
> SnapPuller.java:407)
>   6 at org.apache.solr.handler.ReplicationHandler.doFetch(
> ReplicationHandler.java:319)
>   7 at org.apache.solr.handler.ReplicationHandler$1.run(
> ReplicationHandler.java:220)
>
> Thanks,
>
> Anand
>
>


Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Hi Peter,
I can only guess that the result set doesn't contain a "cat_name" (case
insensitive) column. Other option / question: do you have a transformer
(like scriptTransformer) that manipulates the resultset?

You can debug the resultset in a main class by doing rs.getString
("cat_name")

Cheers,
Andrea
On 3 Jan 2014 15:01, "PeterKerk"  wrote:

> Hi Andrea,
>
> Here you go:
>
> **data-config.xml**
> 
>  url="jdbc:mysql://localhost:3306/wordp" user="***" password="***" />
> 
> 
>
>  query="select
> wt.name as cat_name,wt.slug,wtr.term_taxonomy_id,wtt.term_id,wtt.taxonomy
> from
> wp_term_relationships wtr
> INNER JOIN wp_term_taxonomy wtt ON
> wtt.term_taxonomy_id=wtr.term_taxonomy_id AND wtt.taxonomy='category'
> INNER JOIN wp_terms wt ON wt.term_id=wtt.term_taxonomy_id
> where wtr.object_id='${article.id}';">
> 
> 
> 
> 
>
>
> **schema.xml**
>
>  multiValued="true"  />
>  stored="true" multiValued="true" />
> 
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109353.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>


Re: Very long running replication.

2014-01-03 Thread anand chandak
Folks, would really appreciate if somebody can help/throw some light on 
below issue . This issue is blocking our upgrade, we are doing a 3.x to 
4.x upgrade and indexing around 100g of data.


Any help would be highly appreciated.

Thanks,

Anand


On 1/3/2014 11:46 AM, anand chandak wrote:

Thanks Shalin.


I am facing one issue while replicating, as my replication (very large 
index 100g)is happening, I am also doing the indexing and I believe 
the segment_N file is changing because of new commits. So would the 
replication fail if the the filename is different from what it found 
when fetching the filename list.



Basically, I am seeing this exception :


[explicit-fetchindex-cmd] ERROR 
org.apache.solr.handler.ReplicationHandler- SnapPull failed 
:org.apache.solr.common.SolrException: Unable to download _av3.fdt 
completely. Downloaded 0!=497037
  2 at 
org.apache.solr.handler.SnapPuller$DirectoryFileFetcher.cleanup(SnapPuller.java:1268)
  3 at 
org.apache.solr.handler.SnapPuller$DirectoryFileFetcher.fetchFile(SnapPuller.java:1148)
  4 at 
org.apache.solr.handler.SnapPuller.downloadIndexFiles(SnapPuller.java:743)
  5 at 
org.apache.solr.handler.SnapPuller.fetchLatestIndex(SnapPuller.java:407)
  6 at 
org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:319)
  7 at 
org.apache.solr.handler.ReplicationHandler$1.run(ReplicationHandler.java:220)



And I am trying to find the root cause of this issue. Any help ?

Thanks,

Anand


On 1/2/2014 5:32 PM, Shalin Shekhar Mangar wrote:

Replications won't run concurrently. They are scheduled at a fixed
rate and if a particular pull takes longer than the time period then
subsequent executions are delayed until the running one finishes.

On Tue, Dec 31, 2013 at 4:46 PM, anand chandak 
 wrote:

Quick question about solr replication : What happens if there's a
replication running for very large index that runs more than the 
interval
for 2 replication ? would the automatic runs of replication 
interfere with

the current running one or it would not even spawn next iteration of
replication ? Can somebody throw some light ?












Re: Invalid version (expected 2, but 60) or the data in not in 'javabin' format exception while deleting 30k records

2014-01-03 Thread Mark Miller
Just an FYI, newer version of Solr will deploy the proper error message rather 
than that cryptic one.

- Mark

On Jan 3, 2014, at 12:54 AM, Shawn Heisey  wrote:

> On 1/2/2014 10:22 PM, gpssolr2020 wrote:
>> Caused by: java.lang.RuntimeException: Invalid version (expected 2, but 60)
>> or the data in not in 'javabin' format
> 
> 
> 
>> (Account:123+AND+DATE:["2013-11-29T00:00:00Z"+TO+"2013-11-29T23:59:59Z"])+OR+
>> (Account:345+AND+DATE:["2013-11-29T00:00:00Z"+TO+"2013-11-29T23:59:59Z"])+OR+
>> (Account:569+AND+DATE:["2013-11-29T00:00:00Z"+TO+"2013-11-29T23:59:59Z"])+OR+
>> (Account:789+AND+DATE:["2013-11-29T00:00:00Z"+TO+"2013-11-29T23:59:59Z"])+OR+..+OR+30k
>> th record)
>> 
>> But we are able to delete small number of records without any issues.
>> 
>> Can anyone please help us on this.
> 
> Walter is right about the error message.  This indicates that Solr is
> returning an error response that's in HTML or XML format rather than the
> expected javabin.  The server log should actually contain more
> information about what actually went wrong.
> 
> This is probably happening because the request is too big.  In my
> indexing code, I send deleteByQuery requests in batches of 1000 for this
> very reason.
> 
> If that date range is the same for all of the deletes in a batch, you
> could make the request considerably smaller by using the following format:
> 
> DATE:["2013-11-29T00:00:00Z"+TO+"2013-11-29T23:59:59Z"] AND Account:(123
> OR 345 OR 569 OR 789 OR  3th-value)
> 
> The other option is to increase the max form size of your servlet
> container.  In 4.1 or later, this is actually controlled by Solr
> directly, not the container config, with the formdataUploadLimitInKB
> attribute on the requestParsers tag in solrconfig.xml.  The default is
> 2048, or 2MB.  With 3 clauses like you have described, the request
> will be a little bit more than 2MB.
> 
> There is a bug in the early Jetty 8 versions (that included with Solr
> 4.0) that prevented the form size from being set in the jetty config:
> 
> https://bugs.eclipse.org/bugs/show_bug.cgi?id=397130
> 
> The Jetty 6 that is included with Solr 1.x and 3.x can be properly
> configured.
> 
> Thanks,
> Shawn
> 



Re: Empty facets on Solr with MySQL

2014-01-03 Thread PeterKerk
Hi Andrea,

Here you go:

**data-config.xml** 












**schema.xml** 



 



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109353.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Hi Peter,
Umfprtunately I deleted your first email where you wrote a piece of your
schema...the problem seems to be "cat_name" and not cat_name_raw...could
you please post again your schema?
 On 3 Jan 2014 13:40, "PeterKerk"  wrote:

> Hi Andrea,
>
> You were right, I do see errors when setting the required=true
> attribute...what can it be?
>
>
> Logging console homepage:
>
> 13:31:54
> WARN
> SolrWriter
> Error creating document : SolrInputDocument[comment_status=open,
> post_content=algemeen kpn artikeltje ook over vodafone,
> guid=http://www.telefonievergelijken.nl/wordpress/?p=20, post_excerpt=,
> id=20, post_author=1, post_modified=2014-01-02 14:24:28.0,
> post_name=kpn-en-vodafone, post_title=KPN en Vodafone,
> imgurl=http://www.talkman.nl/wordpress/wp-content/uploads/2013/11/taj.png,
> 
> post_date=2013-12-13 14:12:17.0]
>
> Console errors:
>
> 36588 [Thread-15] WARN  org.apache.solr.handler.dataimport.SolrWriter  û
> Error c
> reating document : SolrInputDocument[comment_status=open,
> post_content=algemeen
> kpn artikeltje ook over vodafone,
> guid=http://www.telefonievergelijken.nl/wordpr
> ess/?p=20, post_excerpt=, id=20, post_author=1, post_modified=2014-01-02
> 14:24:2
> 8.0, post_name=kpn-en-vodafone, post_title=KPN en Vodafone,
> imgurl=http://www.ta
> lkman.nl/wordpress/wp-content/uploads/2013/11/taj.png,
> post_date=2013-12-13
> 14:1
> 2:17.0]
> org.apache.solr.common.SolrException: [doc=20] missing required field:
> cat_name
> at
> org.apache.solr.update.DocumentBuilder.toDocument(DocumentBuilder.jav
> a:328)
> at
> org.apache.solr.update.AddUpdateCommand.getLuceneDocument(AddUpdateCo
> mmand.java:73)
> at
> org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandle
> r2.java:208)
> at
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpd
> ateProcessorFactory.java:69)
> at
> org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(Up
> dateRequestProcessor.java:51)
> at
> org.apache.solr.update.processor.DistributedUpdateProcessor.versionAd
> d(DistributedUpdateProcessor.java:545)
> at
> org.apache.solr.update.processor.DistributedUpdateProcessor.processAd
> d(DistributedUpdateProcessor.java:398)
> at
> org.apache.solr.update.processor.LogUpdateProcessor.processAdd(LogUpd
> ateProcessorFactory.java:100)
> at
> org.apache.solr.handler.dataimport.SolrWriter.upload(SolrWriter.java:
> 70)
> at
> org.apache.solr.handler.dataimport.DataImportHandler$1.upload(DataImp
> ortHandler.java:235)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
> r.java:500)
> at
> org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
> r.java:404)
> at
> org.apache.solr.handler.dataimport.DocBuilder.doFullDump(DocBuilder.j
> ava:319)
> at
> org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java
> :227)
> at
> org.apache.solr.handler.dataimport.DataImporter.doFullImport(DataImpo
> rter.java:422)
> at
> org.apache.solr.handler.dataimport.DataImporter.runCmd(DataImporter.j
> ava:487)
> at
> org.apache.solr.handler.dataimport.DataImporter$1.run(DataImporter.ja
> va:468)
> 36596 [Thread-15] INFO  org.apache.solr.handler.dataimport.DocBuilder  û
> Import
> completed successfully
> 36596 [Thread-15] INFO  org.apache.solr.update.UpdateHandler  û start
> commit{,op
>
> timize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit
> =false,prepareCommit=false}
> 36601 [Thread-15] INFO  org.apache.solr.core.SolrCore  û
> SolrDeletionPolicy.onCo
> mmit: commits:num=2
>
> commit{dir=NRTCachingDirectory(org.apache.lucene.store.SimpleFSDirectory
>
> @C:\Dropbox\Databases\solr-4.3.1\example\example-DIH\solr\tv-wordpress\data\inde
> x lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1836cd1;
> maxCach
> eMB=48.0
> maxMergeSizeMB=4.0),segFN=segments_4o,generation=168,filenames=[segment
> s_4o]
>
> commit{dir=NRTCachingDirectory(org.apache.lucene.store.SimpleFSDirectory
>
> @C:\Dropbox\Databases\solr-4.3.1\example\example-DIH\solr\tv-wordpress\data\inde
> x lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1836cd1;
> maxCach
> eMB=48.0
> maxMergeSizeMB=4.0),segFN=segments_4p,generation=169,filenames=[segment
> s_4p]
> 36602 [Thread-15] INFO  org.apache.solr.core.SolrCore  û newest commit =
> 169[seg
> ments_4p]
> 36603 [Thread-15] INFO  org.apache.solr.search.SolrIndexSearcher  û Opening
> Sear
> cher@54dfcf main
> 36604 [Thread-15] INFO  org.apache.solr.update.UpdateHandler  û
> end_commit_flush
>
> 36604 [searcherExecutor-79-thread-1] INFO  org.apache.solr.core.SolrCore  û
> Quer
> ySenderListener sending requests to Searcher@54dfcf
> main{StandardDirectoryReader
> (segments_4p:1330871938621:nrt)}
> 36608 [searcherExecutor-79-thread-1] INFO  org.apache.solr.core.SolrCore  û
> [tv-
> wordpress] webapp=null path=null
> params={start=0&event=newSearcher&q=solr&di

Re: Empty facets on Solr with MySQL

2014-01-03 Thread PeterKerk
Hi Andrea,

You were right, I do see errors when setting the required=true
attribute...what can it be?


Logging console homepage:

13:31:54
WARN
SolrWriter
Error creating document : SolrInputDocument[comment_status=open,​
post_content=algemeen kpn artikeltje ook over vodafone,​
guid=http://www.telefonievergelijken.nl/wordpress/?p=20,​ post_excerpt=,​
id=20,​ post_author=1,​ post_modified=2014-01-02 14:24:28.0,​
post_name=kpn-en-vodafone,​ post_title=KPN en Vodafone,​
imgurl=http://www.talkman.nl/wordpress/wp-content/uploads/2013/11/taj.png,​
post_date=2013-12-13 14:12:17.0]

Console errors:

36588 [Thread-15] WARN  org.apache.solr.handler.dataimport.SolrWriter  û
Error c
reating document : SolrInputDocument[comment_status=open,
post_content=algemeen
kpn artikeltje ook over vodafone,
guid=http://www.telefonievergelijken.nl/wordpr
ess/?p=20, post_excerpt=, id=20, post_author=1, post_modified=2014-01-02
14:24:2
8.0, post_name=kpn-en-vodafone, post_title=KPN en Vodafone,
imgurl=http://www.ta
lkman.nl/wordpress/wp-content/uploads/2013/11/taj.png, post_date=2013-12-13
14:1
2:17.0]
org.apache.solr.common.SolrException: [doc=20] missing required field:
cat_name
at
org.apache.solr.update.DocumentBuilder.toDocument(DocumentBuilder.jav
a:328)
at
org.apache.solr.update.AddUpdateCommand.getLuceneDocument(AddUpdateCo
mmand.java:73)
at
org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandle
r2.java:208)
at
org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpd
ateProcessorFactory.java:69)
at
org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(Up
dateRequestProcessor.java:51)
at
org.apache.solr.update.processor.DistributedUpdateProcessor.versionAd
d(DistributedUpdateProcessor.java:545)
at
org.apache.solr.update.processor.DistributedUpdateProcessor.processAd
d(DistributedUpdateProcessor.java:398)
at
org.apache.solr.update.processor.LogUpdateProcessor.processAdd(LogUpd
ateProcessorFactory.java:100)
at
org.apache.solr.handler.dataimport.SolrWriter.upload(SolrWriter.java:
70)
at
org.apache.solr.handler.dataimport.DataImportHandler$1.upload(DataImp
ortHandler.java:235)
at
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
r.java:500)
at
org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilde
r.java:404)
at
org.apache.solr.handler.dataimport.DocBuilder.doFullDump(DocBuilder.j
ava:319)
at
org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java
:227)
at
org.apache.solr.handler.dataimport.DataImporter.doFullImport(DataImpo
rter.java:422)
at
org.apache.solr.handler.dataimport.DataImporter.runCmd(DataImporter.j
ava:487)
at
org.apache.solr.handler.dataimport.DataImporter$1.run(DataImporter.ja
va:468)
36596 [Thread-15] INFO  org.apache.solr.handler.dataimport.DocBuilder  û
Import
completed successfully
36596 [Thread-15] INFO  org.apache.solr.update.UpdateHandler  û start
commit{,op
timize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit
=false,prepareCommit=false}
36601 [Thread-15] INFO  org.apache.solr.core.SolrCore  û
SolrDeletionPolicy.onCo
mmit: commits:num=2
   
commit{dir=NRTCachingDirectory(org.apache.lucene.store.SimpleFSDirectory
@C:\Dropbox\Databases\solr-4.3.1\example\example-DIH\solr\tv-wordpress\data\inde
x lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1836cd1;
maxCach
eMB=48.0
maxMergeSizeMB=4.0),segFN=segments_4o,generation=168,filenames=[segment
s_4o]
   
commit{dir=NRTCachingDirectory(org.apache.lucene.store.SimpleFSDirectory
@C:\Dropbox\Databases\solr-4.3.1\example\example-DIH\solr\tv-wordpress\data\inde
x lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1836cd1;
maxCach
eMB=48.0
maxMergeSizeMB=4.0),segFN=segments_4p,generation=169,filenames=[segment
s_4p]
36602 [Thread-15] INFO  org.apache.solr.core.SolrCore  û newest commit =
169[seg
ments_4p]
36603 [Thread-15] INFO  org.apache.solr.search.SolrIndexSearcher  û Opening
Sear
cher@54dfcf main
36604 [Thread-15] INFO  org.apache.solr.update.UpdateHandler  û
end_commit_flush

36604 [searcherExecutor-79-thread-1] INFO  org.apache.solr.core.SolrCore  û
Quer
ySenderListener sending requests to Searcher@54dfcf
main{StandardDirectoryReader
(segments_4p:1330871938621:nrt)}
36608 [searcherExecutor-79-thread-1] INFO  org.apache.solr.core.SolrCore  û
[tv-
wordpress] webapp=null path=null
params={start=0&event=newSearcher&q=solr&distri
b=false&rows=10} hits=0 status=0 QTime=4
36609 [searcherExecutor-79-thread-1] INFO  org.apache.solr.core.SolrCore  û
[tv-
wordpress] webapp=null path=null
params={start=0&event=newSearcher&q=rocks&distr
ib=false&rows=10} hits=0 status=0 QTime=0
36610 [searcherExecutor-79-thread-1] INFO  org.apache.solr.core.SolrCore  û
[tv-
wordpress] webapp=null path=null
params={event=newSearcher&q=static+newSearcher+
warming+query+from+solrconfig.xml&distrib=false} hits=0 status

Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Hi Peter
Sorry I didn't see your question about the log level. I'm here with my
mobile so cannot guide you step by step (don't remember exactly) but is
very very easy. Prior to solr 4 there was a "logging" hyperlink on the
homepage of the console. Starting from 4 you should see the same link on
the left menu

Best,
Andrea
On 3 Jan 2014 10:21, "PeterKerk"  wrote:

> Hi Andrea,
>
> I would say the JDBC driver must be working because when I leave out the
> required="true" from the cat_name field, 4 documents are imported. Since my
> entire DB currently holds only 4 records, there's no need for a LIMIT
> clause
> I guess?
>
>
> Andrea Gazzarini-4 wrote
> > In the solr console set to DEBUG / FINEST the level of DIH classes
>
> How do I do that?
>
> Thanks!
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109290.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>


Re: Empty facets on Solr with MySQL

2014-01-03 Thread Andrea Gazzarini
Sorry I didn't get the point...so that means

-  if you put the required clause you must have some kind of errors
(because no documets are imported)

- your query doesn't throw out that field

I'm quite sure that if you put again the  required attribute and set to
finest the log level you must see something useful

4 documents are ok you don't need limit at all

Andrea
On 3 Jan 2014 10:21, "PeterKerk"  wrote:

> Hi Andrea,
>
> I would say the JDBC driver must be working because when I leave out the
> required="true" from the cat_name field, 4 documents are imported. Since my
> entire DB currently holds only 4 records, there's no need for a LIMIT
> clause
> I guess?
>
>
> Andrea Gazzarini-4 wrote
> > In the solr console set to DEBUG / FINEST the level of DIH classes
>
> How do I do that?
>
> Thanks!
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109290.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>


Re: Grouping results with group.limit return wrong numFound ?

2014-01-03 Thread tasmaniski
Hi and thanks all, 

There is no easy solution implemented in Solr for my problem.
I will try "namedistinct" you said, might improve workaround solution.

Perhaps, the main problem are the names  :)  
because *group.limit*  do same as basic  rows
   and 
group.offset do same as basic  start
  . 
Should be called group.rows and group.start.

So, group.limit isn't real limit, it's only NumRows to return. 


offtopics:
Maybe, we have a bug.
I noticed that *group.offset* doesn't work if we use *group.main=true*



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Grouping-results-with-group-limit-return-wrong-numFound-tp4108174p4109316.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Empty facets on Solr with MySQL

2014-01-03 Thread PeterKerk
Hi Andrea, 

I would say the JDBC driver must be working because when I leave out the
required="true" from the cat_name field, 4 documents are imported. Since my
entire DB currently holds only 4 records, there's no need for a LIMIT clause
I guess?


Andrea Gazzarini-4 wrote
> In the solr console set to DEBUG / FINEST the level of DIH classes

How do I do that?

Thanks!



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109290.html
Sent from the Solr - User mailing list archive at Nabble.com.