Re: Update jar file in Solr 4.4.0

2016-06-14 Thread thakkar.aayush
Actually my changes in updateProcessor.0.1.jar were not reflecting
(functionality wise). I was getting no errors.

Well I dropped the jar file in shared folder only updateProcessor.0.1.jar .
The entry added in solrconfig file was 
**

In /updateProcessor.0.1.jar/ I had class file with the path
/org.apache.solr.update.processor.MyUpdateProcessorFactory/

However I have made some changes and it is working as expected.
*Solution that worked for me:* I changed entry in solrconfig to 
**

Then created new jar file /updateProcessor.0.2.jar/ with following class:
/org.apache.solr.update.processor.MyUpdateProcessorFactory2/

Thanks for your help. I will check with team about zookeepers though :)

Regards,
Aayush



First, having 5 Zookeeper nodes to manage 4 Solr nodes
is serious overkill. Three should be more than sufficient.

what did you put in your configuration? Does your 
directive in solrconfig.xml mention updateProcessor.0.1?

And what error are you seeing exactly?

When Solr starts up, part of the voluminous messages
are where exactly it looks for jar files. So you should
be able to see exactly what Solr is aware of.

If you didn't specify a  directive, one assumes you
dropped the jar somewhere in the Tomcat hive. Is
it in the right place? Did you restart Tomcat? (not sure
this last is necessary, but just in case...)

Best,
Erick

On Mon, Jun 13, 2016 at 7:22 PM, thakkar.aayush
thakkar.aayush@ wrote:
> I have Solr cloud configuration which we run on 4 servers. We use tomcat
> as
> web server for solr. I have 5 zookeepers to maintain the data-replication.
> I
> have added a jar file with custom update processor. This is in shared
> folder
> which is mention in solr.xml**While creating the first version of this jar
> file I gave the name updateProcessor.0.1.jar as the file name. Even though
> it was shared, jar files were added in all the 4 servers.But now I have to
> update the updateProcessor. For this I created updateProcessor0.2.jar. I
> deleted the updateProcessor.0.1.jar from each sever and added a new one.
> But
> changes were not seen ?Any ideas what I am doing wrong? Should this is be
> checked using zkcli ?
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/Update-jar-file-in-Solr-4-4-0-tp4282164.html
> Sent from the Solr - User mailing list archive at Nabble.com.




--
View this message in context: 
http://lucene.472066.n3.nabble.com/Update-jar-file-in-Solr-4-4-0-tp4282164p4282328.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Define search query parameters in Solr or let clients applicationscraft them?

2016-06-14 Thread Derek Poh

Hi Scott, thank you for sharing your solution, appreciate it.

To me interms of maintainability I think it will bebetter to define all 
the parameters either at the client end or solr end.


On 6/15/2016 9:47 AM, scott.chu wrote:

In my case, I write a HTTP gateway between application and Solr engine. This is 
long existed before I use Solr as SE. Back that time, I figure out one day I 
might replace our old SE and it would cause two dilemma:
1> If our applications directly call the API of THE search engines, when we 
replace it with another SE, all the calling statements have to be rewritten. It 
would be a very hard job for us, especially when the number and size of 
applications get bigger.
2> We have applications written in different languages and we from time to time 
need to maunally test status of SE by our system engineers.
Furthermore, we want to fix some default parameters in the gateway for 
simplicity and security issues (e.g. Shortening the size of HTTP call, Prevent 
db names, field names, etc. shown in the HTTP call, etc.)
And these considerations ended up with a gateway design.

For your question, IMHO, I wouldn't define query parameters in Solr unless you 
think they WOULD BE GLOBALIZED. You can consider our solution.


scott.chu,scott@udngroup.com
2016/6/15 (週三)
- Original Message -
From: Derek Poh
To: solr-user
CC:
Date: 2016/6/13 (週一) 11:21
Subject: Define search query parameters in Solr or let clients 
applicationscraft them?


Hi

Would like to get some advice on should the queries parameters be define
in Solr or let the clients applications define and pass the queries
parameters to Solr?

Regards,
Derek



--
CONFIDENTIALITY NOTICE

This e-mail (including any attachments) may contain confidential and/or 
privileged information. If you are not the intended recipient or have received 
this e-mail in error, please inform the sender immediately and delete this 
e-mail (including any attachments) from your computer, and you must not use, 
disclose to anyone else or copy this e-mail (including any attachments), 
whether in whole or in part.

This e-mail and any reply to it may be monitored for security, legal, 
regulatory compliance and/or other appropriate reasons.


-
未在此訊息中找到病毒。
已透過 AVG 檢查 - www.avg.com
版本: 2015.0.6201 / 病毒庫: 4598/12409 - 發佈日期: 06/12/16



--
CONFIDENTIALITY NOTICE 

This e-mail (including any attachments) may contain confidential and/or privileged information. If you are not the intended recipient or have received this e-mail in error, please inform the sender immediately and delete this e-mail (including any attachments) from your computer, and you must not use, disclose to anyone else or copy this e-mail (including any attachments), whether in whole or in part. 


This e-mail and any reply to it may be monitored for security, legal, 
regulatory compliance and/or other appropriate reasons.

Re: Regarding CDCR SOLR 6

2016-06-14 Thread Bharath Kumar
Hi Renaud,

Thank you so much for your response. It is very helpful and it helped me
understand the need for turning on buffering.

Is it recommended to keep the buffering enabled all the time on the source
cluster? If the target cluster is up and running and the cdcr is started,
can i turn off the buffering on the source site?

As you have mentioned, the transaction logs are kept on the source cluster,
until the data is replicated on the target cluster, once the cdcr is
started. Is there a possibility that target cluster is out of sync with the
source cluster and we need to do a hard recovery from the source cluster to
sync up the target cluster?

Also i have the below configuration on the source cluster to synchronize
the update logs.
   
1000
  

Regarding the monitoring of the replication, i am planning to add a script
to check the queue size, to make sure the disk is not full in case the
target site is down and the transaction log size keeps growing on the
source site.
Is there any other recommended approach?

Thanks again, your inputs were very helpful.

On Tue, Jun 14, 2016 at 7:10 PM, Bharath Kumar 
wrote:

> Hi Renaud,
>
> Thank you so much for your response. It is very helpful and it helped me
> understand the need for turning on buffering.
>
> Is it recommended to keep the buffering enabled all the time on the source
> cluster? If the target cluster is up and running and the cdcr is started,
> can i turn off the buffering on the source site?
>
> As you have mentioned, the transaction logs are kept on the source
> cluster, until the data is replicated on the target cluster, once the cdcr
> is started, is there a possibility that if on the target cluster
>
>
>
> On Tue, Jun 14, 2016 at 6:50 AM, Davis, Daniel (NIH/NLM) [C] <
> daniel.da...@nih.gov> wrote:
>
>> I must chime in to clarify something - in case 2, would the source
>> cluster eventually start a log reader on its own?   That is, would the CDCR
>> heal over time, or would manual action be required?
>>
>> -Original Message-
>> From: Renaud Delbru [mailto:renaud@siren.solutions]
>> Sent: Tuesday, June 14, 2016 4:51 AM
>> To: solr-user@lucene.apache.org
>> Subject: Re: Regarding CDCR SOLR 6
>>
>> Hi Bharath,
>>
>> The buffer is useful when you need to buffer updates on the source
>> cluster before starting cdcr, if the source cluster might receive updates
>> in the meanwhile and you want to be sure to not miss them.
>>
>> To understand this better, you need to understand how cdcr clean
>> transaction logs. Cdcr when started (with the START action) will
>> instantiate a log reader for each target cluster. The position of the log
>> reader will indicate cdcr which transaction logs it can clean. If all the
>> log readers are beyond a certain point, then cdcr can clean all the
>> transaction logs up to this point.
>>
>> However, there might be cases when the source cluster will be up without
>> any log readers instantiated:
>> 1) The source cluster is started, but cdcr is not started yet
>> 2) the source cluster is started, cdcr is started, but the target cluster
>> was not accessible when cdcr was started. In this case, cdcr will not be
>> able to instantiate a log reader for this cluster.
>>
>> In these two scenarios, if updates are received by the source cluster,
>> then they might be cleaned out from the transaction log as per the normal
>> update log cleaning procedure.
>> That is where the buffer becomes useful. When you know that while
>> starting up your clusters and cdcr, you will be in one of these two
>> scenarios, then you can activate the buffer to be sure to not miss updates.
>> Then when the source and target clusters are properly up and cdcr
>> replication is properly started, you can turn off this buffer.
>>
>> --
>> Renaud Delbru
>>
>> On 14/06/16 06:41, Bharath Kumar wrote:
>> > Hi,
>> >
>> > I have setup cross data center replication using solr 6, i want to
>> > know why the buffer needs to be enabled on the source cluster? Even if
>> > the buffer is not enabled, i am able to replicate the data between
>> > source and target sites. What is the advantages of enabling the buffer
>> > on the source site? If i enable the buffer, the transaction logs are
>> > never deleted and over a period of time we are running out of disk.
>> > Can you please let me know why the buffer enabling is required?
>> >
>>
>>
>
>
> --
> Thanks & Regards,
> Bharath MV Kumar
>
> "Life is short, enjoy every moment of it"
>



-- 
Thanks & Regards,
Bharath MV Kumar

"Life is short, enjoy every moment of it"


Re: Regarding CDCR SOLR 6

2016-06-14 Thread Bharath Kumar
Hi Renaud,

Thank you so much for your response. It is very helpful and it helped me
understand the need for turning on buffering.

Is it recommended to keep the buffering enabled all the time on the source
cluster? If the target cluster is up and running and the cdcr is started,
can i turn off the buffering on the source site?

As you have mentioned, the transaction logs are kept on the source cluster,
until the data is replicated on the target cluster, once the cdcr is
started, is there a possibility that if on the target cluster



On Tue, Jun 14, 2016 at 6:50 AM, Davis, Daniel (NIH/NLM) [C] <
daniel.da...@nih.gov> wrote:

> I must chime in to clarify something - in case 2, would the source cluster
> eventually start a log reader on its own?   That is, would the CDCR heal
> over time, or would manual action be required?
>
> -Original Message-
> From: Renaud Delbru [mailto:renaud@siren.solutions]
> Sent: Tuesday, June 14, 2016 4:51 AM
> To: solr-user@lucene.apache.org
> Subject: Re: Regarding CDCR SOLR 6
>
> Hi Bharath,
>
> The buffer is useful when you need to buffer updates on the source cluster
> before starting cdcr, if the source cluster might receive updates in the
> meanwhile and you want to be sure to not miss them.
>
> To understand this better, you need to understand how cdcr clean
> transaction logs. Cdcr when started (with the START action) will
> instantiate a log reader for each target cluster. The position of the log
> reader will indicate cdcr which transaction logs it can clean. If all the
> log readers are beyond a certain point, then cdcr can clean all the
> transaction logs up to this point.
>
> However, there might be cases when the source cluster will be up without
> any log readers instantiated:
> 1) The source cluster is started, but cdcr is not started yet
> 2) the source cluster is started, cdcr is started, but the target cluster
> was not accessible when cdcr was started. In this case, cdcr will not be
> able to instantiate a log reader for this cluster.
>
> In these two scenarios, if updates are received by the source cluster,
> then they might be cleaned out from the transaction log as per the normal
> update log cleaning procedure.
> That is where the buffer becomes useful. When you know that while starting
> up your clusters and cdcr, you will be in one of these two scenarios, then
> you can activate the buffer to be sure to not miss updates. Then when the
> source and target clusters are properly up and cdcr replication is properly
> started, you can turn off this buffer.
>
> --
> Renaud Delbru
>
> On 14/06/16 06:41, Bharath Kumar wrote:
> > Hi,
> >
> > I have setup cross data center replication using solr 6, i want to
> > know why the buffer needs to be enabled on the source cluster? Even if
> > the buffer is not enabled, i am able to replicate the data between
> > source and target sites. What is the advantages of enabling the buffer
> > on the source site? If i enable the buffer, the transaction logs are
> > never deleted and over a period of time we are running out of disk.
> > Can you please let me know why the buffer enabling is required?
> >
>
>


-- 
Thanks & Regards,
Bharath MV Kumar

"Life is short, enjoy every moment of it"


Re: Define search query parameters in Solr or let clients applicationscraft them?

2016-06-14 Thread scott.chu
In my case, I write a HTTP gateway between application and Solr engine. This is 
long existed before I use Solr as SE. Back that time, I figure out one day I 
might replace our old SE and it would cause two dilemma:
1> If our applications directly call the API of THE search engines, when we 
replace it with another SE, all the calling statements have to be rewritten. It 
would be a very hard job for us, especially when the number and size of 
applications get bigger.
2> We have applications written in different languages and we from time to time 
need to maunally test status of SE by our system engineers.
Furthermore, we want to fix some default parameters in the gateway for 
simplicity and security issues (e.g. Shortening the size of HTTP call, Prevent 
db names, field names, etc. shown in the HTTP call, etc.)
And these considerations ended up with a gateway design.

For your question, IMHO, I wouldn't define query parameters in Solr unless you 
think they WOULD BE GLOBALIZED. You can consider our solution.


scott.chu,scott@udngroup.com
2016/6/15 (週三)
- Original Message - 
From: Derek Poh 
To: solr-user 
CC: 
Date: 2016/6/13 (週一) 11:21
Subject: Define search query parameters in Solr or let clients 
applicationscraft them?


Hi 

Would like to get some advice on should the queries parameters be define 
in Solr or let the clients applications define and pass the queries 
parameters to Solr? 

Regards, 
Derek 



-- 
CONFIDENTIALITY NOTICE 

This e-mail (including any attachments) may contain confidential and/or 
privileged information. If you are not the intended recipient or have received 
this e-mail in error, please inform the sender immediately and delete this 
e-mail (including any attachments) from your computer, and you must not use, 
disclose to anyone else or copy this e-mail (including any attachments), 
whether in whole or in part. 

This e-mail and any reply to it may be monitored for security, legal, 
regulatory compliance and/or other appropriate reasons. 


- 
未在此訊息中找到病毒。 
已透過 AVG 檢查 - www.avg.com 
版本: 2015.0.6201 / 病毒庫: 4598/12409 - 發佈日期: 06/12/16


Re: Define search query parameters in Solr or let clients applications craft them?

2016-06-14 Thread Derek Poh

Hi Emir

Yaguess one way is to implement a policy where new queries from client 
application have to be reviewcouple with periodic search log grooming as 
you have suggested.


On 6/14/2016 4:12 PM, Emir Arnautovic wrote:

Hi Derek,
Unless you lock all your parameters, there will always be a chance of 
inefficient queries. Only way to fight that is to have full control of 
Solr interface and provide some search API, or to do regular search 
log grooming.


Emir

On 14.06.2016 03:05, Derek Poh wrote:

Hi Emir

Thank you for pointing out the cons of defining them in Solr config.

One of the thing I am worry about in letting clientapplication 
defined the parametersis the developers will use or include 
unnecessary, wrong and resource intensive parameters.



On 6/13/2016 5:50 PM, Emir Arnautovic wrote:

Hi Derek,
Maybe I am looking this from perspective who is working with other 
peoples' setups, but I prefer when it is defined in Solr configs: I 
can get sense of queries from looking at configs, you have mechanism 
to lock some parameters, updates are centralized... However, it does 
come with some cons: it is less expressive than what you can do in 
client code, you have to reload cores when you want to change, 
people tend to override it from client so you get configs in two 
places.


HTH,
Emir

On 13.06.2016 05:21, Derek Poh wrote:

Hi

Would like to get some advice on should the queries parameters be 
define in Solr or let the clients applications define and pass the 
queries parameters to Solr?


Regards,
Derek



--
CONFIDENTIALITY NOTICE
This e-mail (including any attachments) may contain confidential 
and/or privileged information. If you are not the intended 
recipient or have received this e-mail in error, please inform the 
sender immediately and delete this e-mail (including any 
attachments) from your computer, and you must not use, disclose to 
anyone else or copy this e-mail (including any attachments), 
whether in whole or in part.
This e-mail and any reply to it may be monitored for security, 
legal, regulatory compliance and/or other appropriate reasons.





--
CONFIDENTIALITY NOTICE
This e-mail (including any attachments) may contain confidential 
and/or privileged information. If you are not the intended recipient 
or have received this e-mail in error, please inform the sender 
immediately and delete this e-mail (including any attachments) from 
your computer, and you must not use, disclose to anyone else or copy 
this e-mail (including any attachments), whether in whole or in part.
This e-mail and any reply to it may be monitored for security, legal, 
regulatory compliance and/or other appropriate reasons.





--
CONFIDENTIALITY NOTICE 

This e-mail (including any attachments) may contain confidential and/or privileged information. If you are not the intended recipient or have received this e-mail in error, please inform the sender immediately and delete this e-mail (including any attachments) from your computer, and you must not use, disclose to anyone else or copy this e-mail (including any attachments), whether in whole or in part. 


This e-mail and any reply to it may be monitored for security, legal, 
regulatory compliance and/or other appropriate reasons.

Recommendations for analyzing Korean?

2016-06-14 Thread Solr List
Hi -

What's the current recommendation for searching/analyzing Korean?

The reference guide only lists CJK:
https://cwiki.apache.org/confluence/display/solr/Language+Analysis

I see a bunch of work was done on
https://issues.apache.org/jira/browse/LUCENE-4956, but it doesn't look like
that was ever committed - and the last comment was years ago.

There seem to be a few version of this in the wild, both more recent:
https://github.com/juncon/arirang.lucene-analyzer-5.0.0, and the original:
https://sourceforge.net/projects/lucenekorean/ but I'm not sure what's the
canonical source at this point.

I also see this: https://bitbucket.org/eunjeon/mecab-ko-lucene-analyzer

Suggestions?

Thanks,

Tom


Re: result grouping in sharded index

2016-06-14 Thread Jay Potharaju
Any suggestions on how to handle result grouping in sharded index?


On Mon, Jun 13, 2016 at 1:15 PM, Jay Potharaju 
wrote:

> Hi,
> I am working on a functionality that would require me to group documents
> by a id field. I read that the ngroups feature would not work in a sharded
> index.
> Can someone recommend how to handle this in a sharded index?
>
>
> Solr Version: 5.5
>
>
> https://cwiki.apache.org/confluence/display/solr/Result+Grouping#ResultGrouping-DistributedResultGroupingCaveats
>
> --
> Thanks
> Jay
>
>



-- 
Thanks
Jay Potharaju


Re: Update jar file in Solr 4.4.0

2016-06-14 Thread Erick Erickson
First, having 5 Zookeeper nodes to manage 4 Solr nodes
is serious overkill. Three should be more than sufficient.

what did you put in your configuration? Does your 
directive in solrconfig.xml mention updateProcessor.0.1?

And what error are you seeing exactly?

When Solr starts up, part of the voluminous messages
are where exactly it looks for jar files. So you should
be able to see exactly what Solr is aware of.

If you didn't specify a  directive, one assumes you
dropped the jar somewhere in the Tomcat hive. Is
it in the right place? Did you restart Tomcat? (not sure
this last is necessary, but just in case...)

Best,
Erick

On Mon, Jun 13, 2016 at 7:22 PM, thakkar.aayush
 wrote:
> I have Solr cloud configuration which we run on 4 servers. We use tomcat as
> web server for solr. I have 5 zookeepers to maintain the data-replication. I
> have added a jar file with custom update processor. This is in shared folder
> which is mention in solr.xml**While creating the first version of this jar
> file I gave the name updateProcessor.0.1.jar as the file name. Even though
> it was shared, jar files were added in all the 4 servers.But now I have to
> update the updateProcessor. For this I created updateProcessor0.2.jar. I
> deleted the updateProcessor.0.1.jar from each sever and added a new one. But
> changes were not seen ?Any ideas what I am doing wrong? Should this is be
> checked using zkcli ?
>
>
>
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/Update-jar-file-in-Solr-4-4-0-tp4282164.html
> Sent from the Solr - User mailing list archive at Nabble.com.


Re: Boosting exact match fields.

2016-06-14 Thread Erick Erickson
If these are the complete field, i.e. your document
contains exactly "ear phones" and not "ear phones
are great" use a copyField to put it into an "exact_match"
field that uses a much simpler analysis chain based
on KeywordTokenizer (plus, perhaps things like
lowercaseFilter, maybe strip punctuation and the like".
Then you add a clause on exact_match boosted
really high.

Best,
Erick

On Tue, Jun 14, 2016 at 1:01 AM, Naveen Pajjuri
 wrote:
> Hi,
>
> I have documents with a field (data type definition for that field is
> below) values as ear phones, sony ear phones, philips ear phones. when i
> query for earphones sony ear phones is the top result where as i want ear
> phones as top result. please suggest how to boost exact matches. PS: I have
> earphones => ear phones in my synonyms.txt and the datatype definition for
> that field keywords is  positionIncrementGap="100">   "solr.WhitespaceTokenizerFactory"/>  ignoreCase="true" words="stopwords.txt"/>  "solr.LowerCaseFilterFactory"/>  synonyms="synonyms.txt" ignoreCase="true" expand="true"/>  "solr.RemoveDuplicatesTokenFilterFactory"/>   "query">   "solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>  class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true"
> expand="true"/>   "solr.RemoveDuplicatesTokenFilterFactory"/>   REGARDS,
> Naveen


Re: Solr - Error when trying to index the date field.

2016-06-14 Thread Erick Erickson
BTW, the easiest way to check what schema
file you are _actually_ using is through the
admin UI. Select a core, go to files>>and click
on the schema file in question

Best,
Erick

On Tue, Jun 14, 2016 at 5:53 AM, Shawn Heisey  wrote:
> On 6/14/2016 6:09 AM, Syedabbasmehdi Rizvi wrote:
>> Schema file has a field called timestamp of type date. Still it shows
>> that error.
>
> Did you restart Solr or reload the core/collection after modifying the
> schema?
>
> Are you absolutely sure that you are looking at the active schema file?
> Users are often confused about which file is the correct one.
>
> The active schema file will be named either managed-schema or
> schema.xml, depending on how solrconfig.xml is configured.  The examples
> in 5.5 and later use managed-schema exclusively.  Before 5.5, it might
> be either one, depending on which configset you used.
>
> If you're in SolrCloud mode, the active schema will live in zookeeper,
> in the configuration linked to the collection.  If you're not in cloud
> mode, look for a "conf" directory inside the Instance directory that you
> can see on the core overview page in the admin UI.  That is where the
> schema will live.
>
> Thanks,
> Shawn
>


Re: Solr indexing sequentially or randomly?

2016-06-14 Thread Zheng Lin Edwin Yeo
Thank you.

On 14 June 2016 at 20:03, Mikhail Khludnev 
wrote:

> Sequentially.
>
> On Tue, Jun 14, 2016 at 12:32 PM, Zheng Lin Edwin Yeo <
> edwinye...@gmail.com>
> wrote:
>
> > Hi,
> >
> > i would like to find out, does Solr writes to the disk sequentially or
> > randomly during indexing?
> > I'm using Solr 6.0.1.
> >
> > Regards,
> > Edwin
> >
>
>
>
> --
> Sincerely yours
> Mikhail Khludnev
> Principal Engineer,
> Grid Dynamics
>
> 
> 
>


RE: Regarding CDCR SOLR 6

2016-06-14 Thread Davis, Daniel (NIH/NLM) [C]
I must chime in to clarify something - in case 2, would the source cluster 
eventually start a log reader on its own?   That is, would the CDCR heal over 
time, or would manual action be required?

-Original Message-
From: Renaud Delbru [mailto:renaud@siren.solutions] 
Sent: Tuesday, June 14, 2016 4:51 AM
To: solr-user@lucene.apache.org
Subject: Re: Regarding CDCR SOLR 6

Hi Bharath,

The buffer is useful when you need to buffer updates on the source cluster 
before starting cdcr, if the source cluster might receive updates in the 
meanwhile and you want to be sure to not miss them.

To understand this better, you need to understand how cdcr clean transaction 
logs. Cdcr when started (with the START action) will instantiate a log reader 
for each target cluster. The position of the log reader will indicate cdcr 
which transaction logs it can clean. If all the log readers are beyond a 
certain point, then cdcr can clean all the transaction logs up to this point.

However, there might be cases when the source cluster will be up without any 
log readers instantiated:
1) The source cluster is started, but cdcr is not started yet
2) the source cluster is started, cdcr is started, but the target cluster was 
not accessible when cdcr was started. In this case, cdcr will not be able to 
instantiate a log reader for this cluster.

In these two scenarios, if updates are received by the source cluster, then 
they might be cleaned out from the transaction log as per the normal update log 
cleaning procedure.
That is where the buffer becomes useful. When you know that while starting up 
your clusters and cdcr, you will be in one of these two scenarios, then you can 
activate the buffer to be sure to not miss updates. Then when the source and 
target clusters are properly up and cdcr replication is properly started, you 
can turn off this buffer.

--
Renaud Delbru

On 14/06/16 06:41, Bharath Kumar wrote:
> Hi,
>
> I have setup cross data center replication using solr 6, i want to 
> know why the buffer needs to be enabled on the source cluster? Even if 
> the buffer is not enabled, i am able to replicate the data between 
> source and target sites. What is the advantages of enabling the buffer 
> on the source site? If i enable the buffer, the transaction logs are 
> never deleted and over a period of time we are running out of disk. 
> Can you please let me know why the buffer enabling is required?
>



Re: Solr - Error when trying to index the date field.

2016-06-14 Thread Shawn Heisey
On 6/14/2016 6:09 AM, Syedabbasmehdi Rizvi wrote:
> Schema file has a field called timestamp of type date. Still it shows
> that error. 

Did you restart Solr or reload the core/collection after modifying the
schema?

Are you absolutely sure that you are looking at the active schema file? 
Users are often confused about which file is the correct one.

The active schema file will be named either managed-schema or
schema.xml, depending on how solrconfig.xml is configured.  The examples
in 5.5 and later use managed-schema exclusively.  Before 5.5, it might
be either one, depending on which configset you used.

If you're in SolrCloud mode, the active schema will live in zookeeper,
in the configuration linked to the collection.  If you're not in cloud
mode, look for a "conf" directory inside the Instance directory that you
can see on the core overview page in the admin UI.  That is where the
schema will live.

Thanks,
Shawn



RE: Solr - Error when trying to index the date field.

2016-06-14 Thread Syedabbasmehdi Rizvi
Schema file has a field called timestamp of type date. Still it shows that 
error.



-Original Message-
From: Erik Hatcher [mailto:erik.hatc...@gmail.com] 
Sent: Tuesday, June 14, 2016 5:31 PM
To: solr-user@lucene.apache.org
Subject: Re: Solr - Error when trying to index the date field.

You apparently don’t have a `timestamp` field defined in your schema.  The 
error message is:

 unknown field ‘timestamp’





> On Jun 14, 2016, at 5:18 AM, Syedabbasmehdi Rizvi  
> wrote:
> 
> Hi,
> 
> I am trying to index a CSV file that contains a date field. I have the 
> date field configured in schema and config.xml But somehow, it shows an error 
> when I try to index this file which says:
> 
> SimplePostTool version 5.0.0
> Posting files to [base] url http://localhost:8983/solr/polycom/update using 
> content-type text/csv...
> POSTing file book1.csv to [base]
> SimplePostTool: WARNING: Solr returned an error #400 (Bad Request) for 
> url: http ://localhost:8983/solr/polycom/update
> SimplePostTool: WARNING: Response:  encoding="UTF-8"?>   name="status">4002  name="error">org.apache.
> solr.common.SolrException name="root-error-class">org.apache.solr.comm
> on.SolrExceptionERROR: [doc=552998] 
> unknown field 't imestamp'400 
> 
> SimplePostTool: WARNING: IOException while reading response: 
> java.io.IOException
> : Server returned HTTP response code: 400 for URL: 
> http://localhost:8983/solr/po lycom/update
> 
> A help would really be very useful. Kindly share the steps to do so if 
> possible.
> 
> Regards
> Abbas
> 
> 
> ::DISCLAIMER::
> --
> --
> 
> 
> The contents of this e-mail and any attachment(s) are confidential and 
> intended for the named recipient(s) only.
> E-mail transmission is not guaranteed to be secure or error-free as 
> information could be intercepted, corrupted, lost, destroyed, arrive 
> late or incomplete, or may contain viruses in transmission. The e mail and 
> its contents (with or without referred errors) shall therefore not attach any 
> liability on the originator or HCL or its affiliates.
> Views or opinions, if any, presented in this email are solely those of 
> the author and may not necessarily reflect the views or opinions of 
> HCL or its affiliates. Any form of reproduction, dissemination, 
> copying, disclosure, modification, distribution and / or publication of this 
> message without the prior written consent of authorized representative of HCL 
> is strictly prohibited. If you have received this email in error please 
> delete it and notify the sender immediately.
> Before opening any email and/or attachments, please check them for viruses 
> and other defects.
> 
> --
> --
> 



Re: Solr indexing sequentially or randomly?

2016-06-14 Thread Mikhail Khludnev
Sequentially.

On Tue, Jun 14, 2016 at 12:32 PM, Zheng Lin Edwin Yeo 
wrote:

> Hi,
>
> i would like to find out, does Solr writes to the disk sequentially or
> randomly during indexing?
> I'm using Solr 6.0.1.
>
> Regards,
> Edwin
>



-- 
Sincerely yours
Mikhail Khludnev
Principal Engineer,
Grid Dynamics





Re: Solr - Error when trying to index the date field.

2016-06-14 Thread Erik Hatcher
You apparently don’t have a `timestamp` field defined in your schema.  The 
error message is:

 unknown field ‘timestamp’





> On Jun 14, 2016, at 5:18 AM, Syedabbasmehdi Rizvi  
> wrote:
> 
> Hi,
> 
> I am trying to index a CSV file that contains a date field. I have the date 
> field configured in schema and config.xml
> But somehow, it shows an error when I try to index this file which says:
> 
> SimplePostTool version 5.0.0
> Posting files to [base] url http://localhost:8983/solr/polycom/update using 
> content-type text/csv...
> POSTing file book1.csv to [base]
> SimplePostTool: WARNING: Solr returned an error #400 (Bad Request) for url: 
> http
> ://localhost:8983/solr/polycom/update
> SimplePostTool: WARNING: Response: 
> 
> 400 name="QTime">2
>  name="error-class">org.apache.
> solr.common.SolrException name="root-error-class">org.apache.solr.comm
> on.SolrExceptionERROR: [doc=552998] unknown field 
> 't
> imestamp'400
> 
> SimplePostTool: WARNING: IOException while reading response: 
> java.io.IOException
> : Server returned HTTP response code: 400 for URL: 
> http://localhost:8983/solr/po
> lycom/update
> 
> A help would really be very useful. Kindly share the steps to do so if 
> possible.
> 
> Regards
> Abbas
> 
> 
> ::DISCLAIMER::
> 
> 
> The contents of this e-mail and any attachment(s) are confidential and 
> intended for the named recipient(s) only.
> E-mail transmission is not guaranteed to be secure or error-free as 
> information could be intercepted, corrupted,
> lost, destroyed, arrive late or incomplete, or may contain viruses in 
> transmission. The e mail and its contents
> (with or without referred errors) shall therefore not attach any liability on 
> the originator or HCL or its affiliates.
> Views or opinions, if any, presented in this email are solely those of the 
> author and may not necessarily reflect the
> views or opinions of HCL or its affiliates. Any form of reproduction, 
> dissemination, copying, disclosure, modification,
> distribution and / or publication of this message without the prior written 
> consent of authorized representative of
> HCL is strictly prohibited. If you have received this email in error please 
> delete it and notify the sender immediately.
> Before opening any email and/or attachments, please check them for viruses 
> and other defects.
> 
> 



Re: Issues with coordinates in Solr during updating of fields

2016-06-14 Thread Zheng Lin Edwin Yeo
Hi David,

Thanks for your explanation.
I don't see there's a need to to use the gps_0_coordinate and
gps_1_coordinate field for the time being, as the query which I'm using are
querying directly to the gps field.

Regards,
Edwin


On 14 June 2016 at 12:39, David Smiley  wrote:

> Zheng,
> There are a few Solr FieldTypes that are basically composite fields -- a
> virtual field of other fields.  AFAIK they are all spatial related.  You
> don't necessarily need to pay attention to the fact that gps_1_coordinate
> exists under the hood unless you wish to customize the options on that
> field type in the schema.  e.g. if you don't need it for filtering (perhaps
> using RPT for that) then you can set indexed=false.
> ~ David
>
> On Fri, Jun 10, 2016 at 8:43 PM Zheng Lin Edwin Yeo 
> wrote:
>
> > Would like to check, what is the use of the gps_0_coordinate and
> > gps_1_coordinate
> > field then? Is it just to store the data points, or does it have any
> other
> > use?
> >
> > When I do the query, I found that we are only querying the gps_field,
> which
> > is something like this:
> > http://localhost:8983/solr/collection1/highlight?q=*:*={!geofilt
> > pt=1.5,100.0
> > <
> http://localhost:8983/solr/collection1/highlight?q=*:*=%7B!geofiltpt=1.5,100.0
> >
> > sfield=gps d=5}
> >
> >
> > Regards,
> > Edwin
> >
> > On 27 May 2016 at 08:48, Erick Erickson  wrote:
> >
> > > Should be fine. When the location field is
> > > re-indexed (as it is with Atomic Updates)
> > > the two fields will be filled back in.
> > >
> > > Best,
> > > Erick
> > >
> > > On Thu, May 26, 2016 at 4:45 PM, Zheng Lin Edwin Yeo
> > >  wrote:
> > > > Thanks Erick for your reply.
> > > >
> > > > It works when I remove the 'stored="true" ' from the gps_0_coordinate
> > and
> > > > gps_1_coordinate.
> > > >
> > > > But will this affect the search functions of the gps coordinates in
> the
> > > > future?
> > > >
> > > > Yes, I am referring to Atomic Updates.
> > > >
> > > > Regards,
> > > > Edwin
> > > >
> > > >
> > > > On 27 May 2016 at 02:02, Erick Erickson 
> > wrote:
> > > >
> > > >> Try removing the 'stored="true" ' from the gps_0_coordinate and
> > > >> gps_1_coordinate.
> > > >>
> > > >> When you say "...tried to do an update on any other fileds" I'm
> > assuming
> > > >> you're
> > > >> talking about Atomic Updates, which require that the destinations of
> > > >> copyFields are single valued. Under the covers the location type is
> > > >> split and copied to the other two fields so I suspect that's what's
> > > going
> > > >> on.
> > > >>
> > > >> And you could also try one of the other types, see:
> > > >> https://cwiki.apache.org/confluence/display/solr/Spatial+Search
> > > >>
> > > >> Best,
> > > >> Erick
> > > >>
> > > >> On Thu, May 26, 2016 at 1:46 AM, Zheng Lin Edwin Yeo
> > > >>  wrote:
> > > >> > Anyone has any solutions to this problem?
> > > >> >
> > > >> > I tried to remove the gps_0_coordinate and gps_1_coordinate, but I
> > > will
> > > >> get
> > > >> > the following error during indexing.
> > > >> > ERROR: [doc=id1] unknown field 'gps_0_coordinate'
> > > >> >
> > > >> > Regards,
> > > >> > Edwin
> > > >> >
> > > >> >
> > > >> > On 25 May 2016 at 11:37, Zheng Lin Edwin Yeo <
> edwinye...@gmail.com>
> > > >> wrote:
> > > >> >
> > > >> >> Hi,
> > > >> >>
> > > >> >> I have an implementation of storing the coordinates in Solr
> during
> > > >> >> indexing.
> > > >> >> During indexing, I will only store the value in the field name
> > > ="gps".
> > > >> For
> > > >> >> the field name = "gps_0_coordinate" and "gps_1_coordinate", the
> > value
> > > >> will
> > > >> >> be auto filled and indexed from the "gps" field.
> > > >> >>
> > > >> >> > > >> required="false"/>
> > > >> >> > > >> stored="true" required="false"/>
> > > >> >> > > >> stored="true" required="false"/>
> > > >> >>
> > > >> >> But when I tried to do an update on any other fields in the
> index,
> > > Solr
> > > >> >> will try to add another value in the "gps_0_coordinate" and
> > > >> >> "gps_1_coordinate". However, as these 2 fields are not
> > multi-Valued,
> > > it
> > > >> >> will lead to an error:
> > > >> >> multiple values encountered for non multiValued field
> > > gps_0_coordinate:
> > > >> >> [1.0,1.0]
> > > >> >>
> > > >> >> Does anyone knows how we can solve this issue?
> > > >> >>
> > > >> >> I am using Solr 5.4.0
> > > >> >>
> > > >> >> Regards,
> > > >> >> Edwin
> > > >> >>
> > > >>
> > >
> >
> --
> Lucene/Solr Search Committer, Consultant, Developer, Author, Speaker
> LinkedIn: http://linkedin.com/in/davidwsmiley | Book:
> http://www.solrenterprisesearchserver.com
>


Solr indexing sequentially or randomly?

2016-06-14 Thread Zheng Lin Edwin Yeo
Hi,

i would like to find out, does Solr writes to the disk sequentially or
randomly during indexing?
I'm using Solr 6.0.1.

Regards,
Edwin


Filter query wrt main query on block join

2016-06-14 Thread Pranaya Behera

Hi,
 I have indexed nested documents into solr.
How do I filter on the main query using block join query?
Here is what I have in the sense of documents:
Document A -> id, name, title, is_parent=true
Document B -> id, x,y,z
Document C -> id, a , b
Document B & C are child to A. I want to get all the parent which 
children has x and y. So the main query becomes:

q={!parent which="is_parent:true"}x:"Some string" y:"Some other string"

Now I want to filter on the result set of the previous query on how many 
parent has children a.
Is fq={!parent which="is_parent:true"}a:"Specific String" along with the 
q i.e. specified above is correct ?


The main query i.e. "q" is it correct in terms of syntax. If not how can 
I improve that?
What would be a correct "fq" for filtering the resultset based on the 
children the resultset's each document has ?




Re: Block Join Facet not giving results.

2016-06-14 Thread Pranaya Behera

Here it is:
https://gist.github.com/shadow-fox/150c1e5d11cccd4a5bafd307c717ff85

On Tuesday 14 June 2016 01:03 PM, Mikhail Khludnev wrote:


OK. And how does response looks like on meaningful child.facet.field 
request with debugQuery?


14 июня 2016 г. 8:12 пользователь "Pranaya Behera" 
> написал:


Hi Mikhail,
Here is the response for

 q=*:*=true:

https://gist.github.com/shadow-fox/495c50cda339e2a18550e41a524f03f0


On Tuesday 14 June 2016 01:59 AM, Mikhail Khludnev wrote:

Can you post response on q=*:*=true?

On Mon, Jun 13, 2016 at 5:01 PM, Pranaya Behera
>
wrote:

Hi,

 I have followed what the documentation says in this page:
https://cwiki.apache.org/confluence/display/solr/BlockJoin+Faceting

This is my current select requestHandler in solrconfig.xml


 
   explicit
   10
 
 
   bjqFacetComponent
 


And the bjqFacetComponent is:




   
 /bjqfacet
   
   
 bjqFacetComponent
   



   
 /bjqdocsetfacet
   
   
 bjqDocsetFacetComponent
   


As the documentation says.
I am using solr 6.0.1, I have copied the schema to
solr/server/configsets/
and uploaded to zookeeper via command line and then
reloaded the collection
and re-indexed the collection as well. But the select
handler never
responds to child.facet.field for a field in child
documents. It always
gives me zero result with nothing inside the array. I have
looked at the
document that I am indexing and found that indeed there is
data in my child
document to match the facet field, but alas no results.
It neither gives results with select handler nor with
bjqfacet handler.
With select handler all I am getting is the keys but not
the values i.e.
count , counts are always zero. With bjqfacet handler I am
getting an empty
array, no keys no values.

--
Thanks & Regards
Pranaya Behera




-- 
Thanks & Regards

Pranaya Behera



--
Thanks & Regards
Pranaya Behera



Solr - Error when trying to index the date field.

2016-06-14 Thread Syedabbasmehdi Rizvi
Hi,

I am trying to index a CSV file that contains a date field. I have the date 
field configured in schema and config.xml
But somehow, it shows an error when I try to index this file which says:

SimplePostTool version 5.0.0
Posting files to [base] url http://localhost:8983/solr/polycom/update using 
content-type text/csv...
POSTing file book1.csv to [base]
SimplePostTool: WARNING: Solr returned an error #400 (Bad Request) for url: http
://localhost:8983/solr/polycom/update
SimplePostTool: WARNING: Response: 

4002
org.apache.
solr.common.SolrExceptionorg.apache.solr.comm
on.SolrExceptionERROR: [doc=552998] unknown field 't
imestamp'400

SimplePostTool: WARNING: IOException while reading response: java.io.IOException
: Server returned HTTP response code: 400 for URL: http://localhost:8983/solr/po
lycom/update

A help would really be very useful. Kindly share the steps to do so if possible.

Regards
Abbas


::DISCLAIMER::


The contents of this e-mail and any attachment(s) are confidential and intended 
for the named recipient(s) only.
E-mail transmission is not guaranteed to be secure or error-free as information 
could be intercepted, corrupted,
lost, destroyed, arrive late or incomplete, or may contain viruses in 
transmission. The e mail and its contents
(with or without referred errors) shall therefore not attach any liability on 
the originator or HCL or its affiliates.
Views or opinions, if any, presented in this email are solely those of the 
author and may not necessarily reflect the
views or opinions of HCL or its affiliates. Any form of reproduction, 
dissemination, copying, disclosure, modification,
distribution and / or publication of this message without the prior written 
consent of authorized representative of
HCL is strictly prohibited. If you have received this email in error please 
delete it and notify the sender immediately.
Before opening any email and/or attachments, please check them for viruses and 
other defects.




Re: Solr6 CDCR issue with a 3 cloud design

2016-06-14 Thread Renaud Delbru

Hi dmitry,

Was a commit operation sent to the 2 target clusters after the 
replication ? Replicated documents will not appeared until a commit 
operation is sent.


What is the output of the monitoring actions QUEUES and ERRORS ? Are you 
seeing any errors reported ? Are you seeing the queue size not equal to 0 ?


--
Renaud Delbru

On 09/06/16 08:55, dmitry.medve...@barclays.com wrote:

I've set up a 3 cloud CDCR: Source => Target1-Source2 => Target2 CDCR
environment, and the replication process works perfectly, but:

when I shutdown Target1-Source2 cloud (the mediator, for testing for
resilience), index/push some docs to Source1 cloud, get back
Target1-Source2 cloud online after several min, then I only part of the
docs are replicated to the 2 Target clouds (7 of 10 docs tested).

Anyone has an idea what is the reason for such a behavior?

Configurations attached.

Thanks in advance,

Dmitry Medvedev.

___

This message is for information purposes only, it is not a
recommendation, advice, offer or solicitation to buy or sell a product
or service nor an official confirmation of any transaction. It is
directed at persons who are professionals and is not intended for retail
customer use. Intended for recipient only. This message is subject to
the terms at: www.barclays.com/emaildisclaimer
.

For important disclosures, please see:
www.barclays.com/salesandtradingdisclaimer
 regarding market
commentary from Barclays Sales and/or Trading, who are active market
participants; and in respect of Barclays Research, including disclosures
relating to specific issuers, please see http://publicresearch.barclays.com.

___





Re: Regarding CDCR SOLR 6

2016-06-14 Thread Renaud Delbru

Hi Bharath,

The buffer is useful when you need to buffer updates on the source 
cluster before starting cdcr, if the source cluster might receive 
updates in the meanwhile and you want to be sure to not miss them.


To understand this better, you need to understand how cdcr clean 
transaction logs. Cdcr when started (with the START action) will 
instantiate a log reader for each target cluster. The position of the 
log reader will indicate cdcr which transaction logs it can clean. If 
all the log readers are beyond a certain point, then cdcr can clean all 
the transaction logs up to this point.


However, there might be cases when the source cluster will be up without 
any log readers instantiated:

1) The source cluster is started, but cdcr is not started yet
2) the source cluster is started, cdcr is started, but the target 
cluster was not accessible when cdcr was started. In this case, cdcr 
will not be able to instantiate a log reader for this cluster.


In these two scenarios, if updates are received by the source cluster, 
then they might be cleaned out from the transaction log as per the 
normal update log cleaning procedure.
That is where the buffer becomes useful. When you know that while 
starting up your clusters and cdcr, you will be in one of these two 
scenarios, then you can activate the buffer to be sure to not miss 
updates. Then when the source and target clusters are properly up and 
cdcr replication is properly started, you can turn off this buffer.


--
Renaud Delbru

On 14/06/16 06:41, Bharath Kumar wrote:

Hi,

I have setup cross data center replication using solr 6, i want to know why
the buffer needs to be enabled on the source cluster? Even if the buffer is
not enabled, i am able to replicate the data between source and target
sites. What is the advantages of enabling the buffer on the source site? If
i enable the buffer, the transaction logs are never deleted and over a
period of time we are running out of disk. Can you please let me know why
the buffer enabling is required?





Re: Define search query parameters in Solr or let clients applications craft them?

2016-06-14 Thread Emir Arnautovic

Hi Derek,
Unless you lock all your parameters, there will always be a chance of 
inefficient queries. Only way to fight that is to have full control of 
Solr interface and provide some search API, or to do regular search log 
grooming.


Emir

On 14.06.2016 03:05, Derek Poh wrote:

Hi Emir

Thank you for pointing out the cons of defining them in Solr config.

One of the thing I am worry about in letting clientapplication defined 
the parametersis the developers will use or include unnecessary, wrong 
and resource intensive parameters.



On 6/13/2016 5:50 PM, Emir Arnautovic wrote:

Hi Derek,
Maybe I am looking this from perspective who is working with other 
peoples' setups, but I prefer when it is defined in Solr configs: I 
can get sense of queries from looking at configs, you have mechanism 
to lock some parameters, updates are centralized... However, it does 
come with some cons: it is less expressive than what you can do in 
client code, you have to reload cores when you want to change, people 
tend to override it from client so you get configs in two places.


HTH,
Emir

On 13.06.2016 05:21, Derek Poh wrote:

Hi

Would like to get some advice on should the queries parameters be 
define in Solr or let the clients applications define and pass the 
queries parameters to Solr?


Regards,
Derek



--
CONFIDENTIALITY NOTICE
This e-mail (including any attachments) may contain confidential 
and/or privileged information. If you are not the intended recipient 
or have received this e-mail in error, please inform the sender 
immediately and delete this e-mail (including any attachments) from 
your computer, and you must not use, disclose to anyone else or copy 
this e-mail (including any attachments), whether in whole or in part.
This e-mail and any reply to it may be monitored for security, 
legal, regulatory compliance and/or other appropriate reasons.





--
CONFIDENTIALITY NOTICE
This e-mail (including any attachments) may contain confidential 
and/or privileged information. If you are not the intended recipient 
or have received this e-mail in error, please inform the sender 
immediately and delete this e-mail (including any attachments) from 
your computer, and you must not use, disclose to anyone else or copy 
this e-mail (including any attachments), whether in whole or in part.
This e-mail and any reply to it may be monitored for security, legal, 
regulatory compliance and/or other appropriate reasons.


--
Monitoring * Alerting * Anomaly Detection * Centralized Log Management
Solr & Elasticsearch Support * http://sematext.com/



Re: Re-create shard with compositeId router and known hash range

2016-06-14 Thread Henrik Brautaset Aronsen
On Mon, Jun 13, 2016 at 4:59 PM, Erick Erickson 
wrote:

> Yes, Solr will pick that up. You won't have any replicas
> though so you'll have to ADDREPLICA afterwards.
> You could use the EMPTY option on the creteNodeSet
> of the Collections API to create a dummy collection
> to see what a no-replica shard should look like as
> a model
>

Thanks, that's good to know.  I'll definitely try that if the problem
reoccurs.

Henrik


Boosting exact match fields.

2016-06-14 Thread Naveen Pajjuri
Hi,

I have documents with a field (data type definition for that field is
below) values as ear phones, sony ear phones, philips ear phones. when i
query for earphones sony ear phones is the top result where as i want ear
phones as top result. please suggest how to boost exact matches. PS: I have
earphones => ear phones in my synonyms.txt and the datatype definition for
that field keywords is REGARDS,
Naveen


Re: Block Join Facet not giving results.

2016-06-14 Thread Mikhail Khludnev
OK. And how does response looks like on meaningful child.facet.field
request with debugQuery?
14 июня 2016 г. 8:12 пользователь "Pranaya Behera" 
написал:

> Hi Mikhail,
> Here is the response for
>
>  q=*:*=true:
>
> https://gist.github.com/shadow-fox/495c50cda339e2a18550e41a524f03f0
>
>
> On Tuesday 14 June 2016 01:59 AM, Mikhail Khludnev wrote:
>
>> Can you post response on q=*:*=true?
>>
>> On Mon, Jun 13, 2016 at 5:01 PM, Pranaya Behera 
>> wrote:
>>
>> Hi,
>>>
>>>  I have followed what the documentation says in this page:
>>> https://cwiki.apache.org/confluence/display/solr/BlockJoin+Faceting
>>>
>>> This is my current select requestHandler in solrconfig.xml
>>>
>>> 
>>>  
>>>explicit
>>>10
>>>  
>>>  
>>>bjqFacetComponent
>>>  
>>> 
>>>
>>> And the bjqFacetComponent is:
>>> >> class="org.apache.solr.search.join.BlockJoinFacetComponent"/>
>>> >> class="org.apache.solr.search.join.BlockJoinDocSetFacetComponent"/>
>>>
>>> >> class="org.apache.solr.handler.component.SearchHandler">
>>>
>>>  /bjqfacet
>>>
>>>
>>>  bjqFacetComponent
>>>
>>> 
>>>
>>> >> class="org.apache.solr.handler.component.SearchHandler">
>>>
>>>  /bjqdocsetfacet
>>>
>>>
>>>  bjqDocsetFacetComponent
>>>
>>> 
>>>
>>> As the documentation says.
>>> I am using solr 6.0.1, I have copied the schema to
>>> solr/server/configsets/
>>> and uploaded to zookeeper via command line and then reloaded the
>>> collection
>>> and re-indexed the collection as well. But the select handler never
>>> responds to child.facet.field for a field in child documents. It always
>>> gives me zero result with nothing inside the array. I have looked at the
>>> document that I am indexing and found that indeed there is data in my
>>> child
>>> document to match the facet field, but alas no results.
>>> It neither gives results with select handler nor with bjqfacet handler.
>>> With select handler all I am getting is the keys but not the values i.e.
>>> count , counts are always zero. With bjqfacet handler I am getting an
>>> empty
>>> array, no keys no values.
>>>
>>> --
>>> Thanks & Regards
>>> Pranaya Behera
>>>
>>>
>>>
>>
> --
> Thanks & Regards
> Pranaya Behera
>
>


RE: [E] Re: Question about Data Import Handler

2016-06-14 Thread Mikhail Khludnev
Sas,
I have no idea why it might not work, perhaps debugging DIH in Solr Admin
or just via request param might answer this question.
09 июня 2016 г. 19:43 пользователь "Jamal, Sarfaraz"
 написал:

> I am on SOLR6 =)
>
> Thanks,
>
> Sas
>
> -Original Message-
> From: Mikhail Khludnev [mailto:mkhlud...@griddynamics.com]
> Sent: Thursday, June 9, 2016 12:42 PM
> To: solr-user 
> Subject: [E] Re: Question about Data Import Handler
>
> which version of Solr do you run?
>
> On Thu, Jun 9, 2016 at 6:23 PM, Jamal, Sarfaraz <
> sarfaraz.ja...@verizonwireless.com.invalid> wrote:
>
> > Hi Guys,
> >
> > I have a question about the data import handler and its configuration
> > file
> >
> > This is what a part of my data-config looks like:
> >
> >
> > 
> > 
> >
> > 
> > 
> >  
> > ===
> >
> > I would like it so that when its indexed, it returns in xml the
> > following when on that doc.
> >
> > -
> > This Is my name
> > This is my description 
> >
> > The best I have gotten it to do so far is to add to the values in name
> > and description, which are fields on the doc.
> >
> > Thanks for any help -
> >
> > P.S. I shall be replying to the other threads as well, I Just took a
> > break from it to come work on another part of SOLR.
> >
> > Sas
> >
>
>
>
> --
> Sincerely yours
> Mikhail Khludnev
> Principal Engineer,
> Grid Dynamics
>
> 
> 
>