doing before, so there may be
> something I am missing myself.
>
>
> On Thu, 17 Sep 2020 at 12:46, Pratik Patel wrote:
> >
> > Thanks for your reply Alexandre.
> >
> > I have "_root_" and "_nest_path_" fields in my schema but not
&
uld also include _nest_path and
> _nest_parent_. Your particular exception seems to be triggering
> something (maybe a bug) related to - possibly - missing _nest_path_
> field.
>
> See:
> https://lucene.apache.org/solr/guide/8_5/indexing-nested-documents.html#indexing-nested-documents
etFieldFromHierarchy(SolrInputDocument
> completeHierarchy, String fieldPath) {
> final List docPaths =
> StrUtils.splitSmart(fieldPath.substring(1), '/');
> ..
>}
Any idea what's wrong here?
Thanks
On Wed, Se
Hello Everyone,
I am trying to update a field of a child document using atomic updates
feature. I am using solr and solrJ version 8.5.0
I have ensured that my schema satisfies the conditions for atomic updates
and I am able to do atomic updates on normal documents but with nested
child documents,
filter and you can implement any logic you want: you consume
> three tokens and emit tokens based on adjacent tokens.
>
> HTH,
> Emir
> --
> Monitoring - Log Management - Alerting - Anomaly Detection
> Solr & Elasticsearch Consulting Support Training - http://sematext.com/
Hello Everyone,
Let's say I have an analyzer which has following token stream as an output.
*token stream : [], a, ab, [], c, [], d, de, def .*
Now let's say I want to add another filter which will drop a certain tokens
based on whether adjacent token on the right side is [] or some string.
l Bernstein
> > http://joelsolr.blogspot.com/
> >
> >
> > On Tue, Jan 28, 2020 at 11:52 AM pratik@semandex
> > wrote:
> >
> >> Joel Bernstein wrote
> >> > Ok, that sounds like a bug. I can create a ticket for this.
> >> >
> >
Thanks Shawn! This is what I needed.
On Wed, Nov 20, 2019 at 3:59 PM Shawn Heisey wrote:
> On 11/20/2019 1:34 PM, Pratik Patel wrote:
> > Let's say I have a collection called "collection1" which uses config set
> > "config_set_1".
> > Now, using &q
Hello Everyone,
Let's say I have a collection called "collection1" which uses config set
"config_set_1".
Now, using "upconfig" command, I upload a new configuration called
"config_set_2". How can I make "collection1" use "config_set_2" instead of
"config_set_1"?
I know that if I upload new config
Hi Joel,
Looks like this is going to be very helpful, thank you! I am wondering
whether the visualizations are generated through third party library or is
it something which would be part of solr distribution?
https://github.com/apache/lucene-solr/blob/visual-guide/solr/solr-ref-guide/src/visualiz
Hello Everyone,
We use SolrJ with POJOs to index documents into solr. If a POJO has a field
annotated with @child then SolrJ automatically adds those objects as
children of the POJO. This works fine and indexing is done properly.
However, when I retrieve the same document through same POJO using
uest = new QueryRequest(paramsLoc, SolrRequest.METHOD.POST);
Is this also a bug?
On Tue, Jul 2, 2019 at 10:17 AM Pratik Patel wrote:
> Great, thanks!
>
> On Tue, Jul 2, 2019 at 6:37 AM Joel Bernstein wrote:
>
>> Ok, that sounds like a bug. I can create a ticket for this.
>>
Great, thanks!
On Tue, Jul 2, 2019 at 6:37 AM Joel Bernstein wrote:
> Ok, that sounds like a bug. I can create a ticket for this.
>
> On Mon, Jul 1, 2019 at 5:57 PM Pratik Patel wrote:
>
> > I think the problem was that my streaming expression was always returning
> &g
I think the problem was that my streaming expression was always returning
just one node. When I added more data so that I can have more than one
node, I started seeing the result.
On Mon, Jul 1, 2019 at 11:21 AM Pratik Patel wrote:
> Hello Everyone,
>
> I am trying to execute
Hello Everyone,
I am trying to execute following streaming expression with "scoreNodes"
function in it. This is taken from the documentation.
scoreNodes(top(n="50",
sort="count(*) desc",
nodes(baskets,
random(baskets, q="productID:ABC
a little bit larger than your largest
> file).
> If possible you can try to avoid storing the NLP / ML models in Solr but
> provide them on a share or similar where all Solr nodes have access to.
>
> > Am 11.06.2019 um 00:32 schrieb Pratik Patel :
> >
> > Hi,
> >
Hi,
I am trying to upload a configuration to "MiniSolrCloudCluster" in my unit
test. This configuration has some binary files for NLP related
functionality. Some of these binary files are bigger than 5 MB. If I try to
upload configuration with these files then it doesn't work. I can set up
the clu
ata in new collection. I see that new collection is created but it seems
to be without any data.
Am I missing something here? Any idea what could be the cause of this?
Thanks!
Pratik
On Thu, Jun 6, 2019 at 11:18 AM Pratik Patel wrote:
> Thanks for the reply Alexandre, only special thi
If your children documents have a link to parent documents (like parent id
or something) then you can use graph traversal to do this.
On Mon, Jun 10, 2019 at 8:01 AM Jai Jamba
wrote:
> Can anyone help me in this ?
>
>
>
> --
> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
tion” without the configfiles, but it will
> obviously fail when try to use it.
>
>
> On the surface, this znode already exists:
>
> /solr/configs/collection2
>
> So it looks like somehow you're
>
> > On Jun 4, 2019, at 12:29 PM, Pratik Patel pra...@semandex.net&
t is done differently.
>
> Regards,
> Alex
>
> On Wed, Jun 5, 2019, 6:29 PM Pratik Patel, wrote:
>
> > Hello Everyone,
> >
> > I am trying to write some unit tests for solr queries which requires some
> > data in specific state. There is a way to load th
Hello Everyone,
I am trying to write some unit tests for solr queries which requires some
data in specific state. There is a way to load this data through json files
but the problem is that the required data needs to have parent-child blocks
to be present.
Because of this, I would prefer if there
Hello Everyone,
I am trying to run a simple unit test using solr test framework. At this
point, all I am trying to achieve is to be able to upload some
configuration and create a collection using solr test framework.
Following is the simple code which I am trying to run.
private static final Str
/url?u=https-3A__github.com_apache_lucene-2Dsolr_blob_master_solr_solrj_src_test_org_apache_solr_client_solrj_io_stream_StreamExpressionTest.java&d=DwIBaQ&c=RoP1YumCXCgaWHvlZYR8PZh8Bv7qIrMUB65eapI_JnE&r=lUsTzFRk0CX38HvagQ0wd52D67dA0fx_D6M6F3LHzAU&m=9tFliF4KA1tiG2lGmDJWO34hyq9-Sz1inAxRP
Hello Everyone,
I want to write unit tests for some solr queries which are being triggered
through java code. These queries includes complex streaming expressions and
faceting queries which requires large number of documents to be present in
solr index. I can not create and push so many documents
Hello Everyone,
Is there a way to paginate the results of Streaming Expression?
Let's say I have a simple gatherNodes function which has count operation at
the end of it. I can sort by the count fine but now I would like to be able
to select specific sub set of result based on pagination paramete
Hello Everyone,
I have a Bean object which can have child documents of classes Child_type1
and Child_type2. When I try to index this document, I get an error message
"Doc cannot have more than one Field with child=true".
I looked at the mailing list but couldn't find any solution for this.
Any su
Problem #1 can probably be solved by using "fetch" function. (
https://lucene.apache.org/solr/guide/6_6/stream-decorators.html#fetch)
Problem #2 and #3 can be solved by normalizing the graph connections and by
applying cartesianProduct on multi valued field, as described here.
http://lucene.472066
Joel Bernstein
> http://joelsolr.blogspot.com/
>
>
> On Mon, Jan 21, 2019 at 3:02 PM Pratik Patel wrote:
>
> > Aman,
> >
> > Thanks for the reply!
> >
> > I have tried with corrected query but it doesn't solve the problem. also,
> > my tags filter matches multipl
I see two rows params, looks like which will be overwritten by rows=2, and
> then your tags filter is resulting only one document. Please remove extra
> rows and try.
>
> On Mon, Jan 21, 2019, 08:44 Pratik Patel
> > Hi Everyone!
> >
> > I am trying to use MLT request
Hi Everyone!
I am trying to use MLT request handler. My query matches more than one
documents but the response always seems to pick up the first document and
interestingTerms also seems to be corresponding to that single document
only.
What I am expecting is that if my query matches multiple docu
alyzers-common/org/apache/lucene/analysis/shingle/ShingleFilterFactory.html
> . I'll try to fix that later (SOLR-12996).
>
> Regards,
>Alex.
> On Fri, 16 Nov 2018 at 10:44, Pratik Patel wrote:
> >
> > @Markus @Walter, @Alexandre is right. The culprit was not Sto
ou using? significantTerms?
>
> On Thu, Nov 15, 2018 at 7:09 PM Alexandre Rafalovitch
> wrote:
>
> > I think the underscore actually comes from the Shingles (parameter
> > fillerToken). Have you tried setting it to empty string?
> >
> > Regards,
> > Al
Hi Markus,
Thanks for the reply. I tried using ShingleFilter and it seems to
be working. However, I am hitting an issue when it is used with
StopWordFilter. StopWordFilter leaves an underscore "_" for removed words
and it kind of screws up the data in index.
I tried setting enablePositionIncremen
Hello Everyone,
Standard way of tokenizing in solr would divide the text by white space in
solr.
Is there a way by which we can index multi-term phrases like "Machine
Learning" instead of "Machine", "Learning"?
Is it possible to create a specific field type for such phrases which has
its own inde
u invisage. As in, is
> there training corpus, are you looking at NGram techniques, etc.
>
> Regards,
> Alex.
> On Wed, 17 Oct 2018 at 13:40, Pratik Patel wrote:
> >
> > Hi Everyone,
> >
> > I have been using Semantic Knowledge Graph for document summarizat
Hi Everyone,
I have been using Semantic Knowledge Graph for document summarization, term
correlation and document similarity. It has produced very good results
after appropriate tuning.
I was wondering if there is any way the Semantic Knowledge Graph can be
used to for Named Entity Extraction lik
Hello All,
I am trying to understand how exactly query time boosting works in solr.
Primarily, I want to understand if absolute boost values matter or is it
just the relative difference between various boost values which decides
scoring. Let's take following two queries for example.
// case1: q p
I am trying to use Semantic Knowledge Graph in my java based application.
I have a Semantic Knowledge Graph query which works fine if I trigger it
through browser using restlet client. Following is the query.
{
"queries": [
"foo:\"5a6127a7234e76473a816f1c\""
],
"compare": [
{
; ConcatOperationTest.java <https://github.com/apache/
> lucene-solr/blob/branch_6_4/solr/solrj/src/test/org/
> apache/solr/client/solrj/io/stream/ops/ConcatOperationTest.java>
>
>
> > On Jun 27, 2018, at 1:27 PM, Aroop Ganguly
> wrote:
> >
> > It should, but
rc/java/org/apache/solr/client/solrj/io/eval/AppendEvaluator.java <
> https://github.com/apache/lucene-solr/blob/master/solr/
> solrj/src/java/org/apache/solr/client/solrj/io/eval/AppendEvaluator.java>
> >>
> >>
> >>> On Jun 27, 2018, at 12:58 PM, Pra
Hello,
Is there a function which can be used in Streaming Expressions to
concatenate two strings? I want to use it just like add(1,2) in a Streaming
Expression. Essentially, I want to achieve something as follows.
select(
search(..),
conceptid as foo,
storeid as bar
concat(foo,bar
Hello Everyone,
Is there any java library for building Streaming Expressions? Currently, I
am using solr's java client and building Streaming Expressions as follows.
StreamFactory factory = new StreamFactory().withCollectionZkHost( collName,
zkHost )
.withFunctionName("gatherN
he potential improvement in performance, I was curious. Please
let me know if you feel there is a better approach.
Thanks
- Pratik
On Thu, Jun 21, 2018 at 7:05 PM, Joel Bernstein wrote:
> Currently the gatherNodes expression can only be filtered by a traditional
> filter query. I'
We can limit the scope of graph traversal by applying some filter along the
way as follows.
gatherNodes(emails,
walk="john...@apache.org->from",
fq="body:(solr rocks)",
gather="to")
Is it possible to replace "body:(solr rocks)" by some streaming expression
lik
Using cursor marker might help as explained in this documentation
https://lucene.apache.org/solr/guide/6_6/pagination-of-results.html
On Fri, May 18, 2018 at 4:13 PM, Deepak Goel wrote:
> I wonder if in-memory-filesystem would help...
>
> On Sat, 19 May 2018, 01:03 Erick Erickson,
> wrote:
>
>
@Alessandro I will see if I can reproduce the same issue just by turning
off omitNorms on field type. I'll open another mail thread if required.
Thanks.
On Thu, Feb 15, 2018 at 6:12 AM, Howe, David
wrote:
>
> Hi Alessandro,
>
> Some interesting testing today that seems to have gotten me closer t
You are right, in my case this field type was applied to many text fields.
These includes many copy fields and dynamic fields as well. In my case,
only specifying omitNorms=true for field type "text_general" fixed the
issue. I didn't do anything else or had any other bug.
On Wed, Feb 14, 2018 at 1
I had a similar issue with index size after upgrading to version 6.4.1 from
5.x. The issue for me was that the field which caused index size to be
increased disproportionately had a field type("text_general") for which
default value of omitNorms was not true. Turning it on explicitly on field
fixed
tially always test this out on a local, stand-alone instance
> until I can index a few documents and query them, it's faster than
> always having to remember to move them to ZooKeeper
>
> Best,
> Erick
>
> On Wed, Feb 7, 2018 at 7:14 PM, Pratik Patel wrote:
> > He
hough, to just start all over.
>
> BTW, do not delete any field that begins and ends with an underscore,
> e.g. _version_ unless you know exactly what the consequences are
>
> Best,
> Erick
>
> On Wed, Feb 7, 2018 at 2:59 PM, Pratik Patel wrote:
> > Hello all,
&g
Hello all,
I have added some fields to default managed-schema file. I was wondering if
it is safe to take default managed-schema file as is and add your own
fields to it in production. What is the best practice for this? As I
understand, it should be safe to use default schema as base if documents
Roll up needs documents to be sorted by the "over" field.
Check this for more details
http://lucene.472066.n3.nabble.com/Streaming-Expressions-rollup-function-returning-results-with-duplicate-tuples-td4342398.html
On Wed, Nov 1, 2017 at 3:41 PM, Kojo wrote:
> Wrap cartesianProduct function with
ection? Or just use this tuples in query time?
>
> 2017-10-30 11:00 GMT-02:00 Pratik Patel :
>
> > By including Cartesian function in Streaming Expression pipeline, you can
> > convert a tuple having one multivalued field into multiple tuples where
> > each tuple holds
This solution looks like normalizing data like a m2m table in sql database,
> is it?
>
>
>
> 2017-10-29 21:51 GMT-02:00 Pratik Patel :
>
> > For now, you can probably use Cartesian function of Streaming Expressions
> > which Joel implemented to solve the same problem.
&
For now, you can probably use Cartesian function of Streaming Expressions
which Joel implemented to solve the same problem.
https://issues.apache.org/jira/browse/SOLR-10292
http://joelsolr.blogspot.com/2017/03/streaming-nlp-is-coming-in-solr-66.html
Regards,
Pratik
On Sat, Oct 28, 2017 at 7:38 P
Thanks for the clarification. I'll change my code to accommodate this
behavior.
On Thu, Oct 5, 2017 at 6:24 PM, Chris Hostetter
wrote:
> : > "startTime":"2013-02-10T18:36:07.000Z"
> ...
> : handler. It gets added successfully but when I retrieve this document
> back
> : using "id
Hello Everyone,
Say I have a document like one below.
> {
> "id":"test",
> "startTime":"2013-02-10T18:36:07.000Z"
> }
I add this document to solr index using the admin UI and "update" request
handler. It gets added successfully but when I retrieve this document back
using "id"
Hi Everyone,
We are running solr 6.4.1 in cloud mode on CentOS production server.
Currently, we are using the embedded zookeeper. It is a simple set up with
one collection and one shard.
By default, Jetty server binds to all interfaces which is not safe so we
have changed the bin/solr script. We
Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Thu, Jun 29, 2017 at 10:06 AM, Pratik Patel
> wrote:
>
> > Hey Everyone,
> >
> > This is about the facet function of Streaming Expression. Is there any
> way
> > to set limit for number of facets to infinite?
Hey Everyone,
This is about the facet function of Streaming Expression. Is there any way
to set limit for number of facets to infinite? The *bucketSizeLimit
parameter *seems to accept only those numbers which are greater than 0.
Thanks,
Pratik
n
> wrote:
>
> > You'll need to use the sort expression to sort the nodes by schemaType
> > first. The rollup expression is doing a MapReduce rollup that requires
> the
> > the records to be sorted by the "over" fields.
> >
> > Joel Bernstein
>
Hi,
I have a streaming expression which uses rollup function. My understanding
is that rollup takes an incoming stream and aggregates over given buckets.
However, with following query the result contains duplicate tuples.
Following is the streaming expression.
rollup(
fetch(
collecti
Hi,
When we do a Carrot Clustering query on a set of solr documents we get back
following type of response.
DDR
3.9599865057283354
TWINX2048-3200PRO
VS1GB400C3
VDBDB1A16
iPod
11.959228467
solr.blogspot.com/
>
> On Tue, Mar 14, 2017 at 7:53 PM, Pratik Patel wrote:
>
> > Wow, this is interesting! Is it going to be a new addition to solr or is
> it
> > already available cause I can not find it in documentation? I am using
> solr
> > version 6.4.1.
>
I am trying to write a streaming expression in solrj. Following is the
query that I want to implement in Java.
having(
> gatherNodes(collection1,
> search(collection1,q="*:*",fl="conceptid",sort="conceptid
> asc",fq=storeid:"524efcfd505637004b1f6f24",fq=tags:"Company",fq=tags:"Prospects2",
> qt="/
> scatter="leaves",
> count(*)),
> gt(count(*),1))),
> fl="concept_name",
> on="ancestors=conceptid")
>
> Joel Bernstein
> http:
t names for ancestors?
>
> Can you explain a little more about the use case?
>
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Tue, Mar 14, 2017 at 11:08 AM, Pratik Patel
> wrote:
>
> > I have two types of documents in my index. eventLink and concept
I have two types of documents in my index. eventLink and concepttData.
eventLink { ancestors:[,] }
conceptData-{ id:id1, conceptid, concept_name . }
Both are in same collection.
In my query, I am doing a gatherNodes query wrapped in some other function
and ultimately I am getting a b
has string
> comparisons.
>
> In the expression you're working with it would be much more performant
> though to filter the query on the storeid.
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
> On Mon, Mar 13, 2017 at 1:06 PM, Pratik Patel wrote:
>
>
it's not a stable version*
On Mon, Mar 13, 2017 at 1:34 PM, Pratik Patel wrote:
> Thanks Joel! This is just a simplified sample query that I created to
> better demonstrate the issue. I am not sure whether I want to upgrade to
> solr 6.5 as only developer version is available
Hi,
I am trying to write a streaming expression with 'having' function in it.
Following is my simple query.
having(
>search(collection1,q="*:*",fl="storeid",sort="storeid
> asc",fq=tags:"Company"),
>eq(storeid,524efcfd505637004b1f6f24)
> )
Here, storeid is a field of type "string" in s
I am trying to do a graph traversal query using gatherNode function. I am
seeding a streaming expression to get some documents and then I am trying
to map their ids(conceptid) to a multi valued field "participantIds" and
gather nodes.
Here is the query I am doing.
gatherNodes(collection1,
> sear
;
> http://www.solr-start.com/ - Resources for Solr users, new and experienced
>
>
> On 22 February 2017 at 12:02, Pratik Patel wrote:
> > Asterisks were not for formatting, I was trying to use a wildcard
> operator.
> > Here is another example query and "parse
Erick Erickson
wrote:
> That's not what I'm looking for. Way down near the end there should be
> an entry like
> "parsed_query toString"
>
> This line is pretty suspicious: 82, "params":{ "q":"Description_note:*
> and *"
>
> Attach &debug=query to your query and look at the parsed query that's
> returned.
> That'll tell you what was searched at least.
>
> You can also use the TermsComponent to examine terms in a field directly.
>
> Best,
> Erick
>
> On Tue, Feb 21, 2017 at 2
I have a field type in schema which has been applied stopwords list.
I have verified that path of stopwords file is correct and it is being
loaded fine in solr admin UI. When I analyse these fields using "Analysis" tab
of the solr admin UI, I can see that stopwords are being filtered out.
However,
9/02/scaling-lucene-and-solr/#d0e71
Thanks,
Pratik
On Tue, Feb 21, 2017 at 12:03 PM, Pratik Patel wrote:
> I am using the schema from solr 5 which does not have any field with
> docValues enabled.In fact to ensure that everything is same as solr 5
> (except the breaking changes) I am usin
use of doc values should actually blow
> > up the size of your index considerably if they are in fields that get
> sent
> > a lot of data.
> >
> > On Tue, Feb 21, 2017 at 10:50 AM, Pratik Patel
> wrote:
> >
> >> Thanks for the reply. I can see that in
ck what index file extensions
> contribute most to the difference? That could give a hint.
>
> Regards,
> Alex
>
> On 21 Feb 2017 9:47 AM, "Pratik Patel" wrote:
>
> > Here is the same question in stackOverflow for better format.
> >
> > http:
Here is the same question in stackOverflow for better format.
http://stackoverflow.com/questions/42370231/solr-
dynamic-field-blowing-up-the-index-size
Recently, I upgraded from solr 5.0 to solr 6.4.1. I can run my app fine but
the problem is that index size with solr 6 is way too large. In solr
81 matches
Mail list logo