Is there any way to use geo_bounds as a sub aggregator?
I have a geohash_grid aggregation, and it will be very useful to get bounds
for each bucket of the data.
Right now my ES 1.3.0 says that he Could not find aggregator type [bounds]
in [aggregations]].
--
You received this message because
I have created a gist with an analyzer that uses filter shingle in attempt to
match sub phrases. For instance I have entries in the table with discrete
phrases likeEGFRLungCancerLung Cancerand I want to match these when
searching the phrase 'EGFR related lung cancerMy expectation is that the
Using ES 1.2.2
The below aggregation will give me my average requests per second and the
average response times for each second.
Is this the only way to do it or is there better way? Since scripting is a
bit slow?
aggs: {
tps: {
date_histogram: {
field:
You should be able to achieve this with the below query. If you still
cannot get it working could you provide a cURL example which reproduces
your issue?
curl -XGET http://localhost:9200/geo/_search; -d'
{
size: 0,
aggs: {
geohash : {
geohash_grid: {
field: location,
The best option could be to compute that at index time.
On a client level or by using the new transform feature:
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/mapping-transform.html#mapping-transform
HTH
--
David ;-)
Twitter : @dadoonet / @elasticsearchfr / @scrutmydocs
Hello everyone, I'm to faraway to consider myself an experienced user .
But I would like to make a multimaster elasticsearch cluster
infrastructure.
As it's described bellow:
master1
http://pastebin.com/PZ1saiL5
master2
http://pastebin.com/KV2Jmqby
master3
http://pastebin.com/296Lh5Dy
When you define the shards and replica's it's at a cluster level, it
doesn't create them on each local server, it spreads them out.
Ultimately what you will find is that the number of shards and replicas
will be based on which ever nodes is the master at the time.
Regards,
Mark Walkom
Indeed the short-circuit rule applies. I tried to reproduce the issue
locally on master but didn't manage to.
How do you check that the native script is executed against data that fails
the first clause? Can you provide us with a curl recreation of the issue?
On Wed, Jul 23, 2014 at 6:31 PM,
Here is an example: https://gist.github.com/svartalf/c3cb76c40dd4fa1374fa
I'm using two inner aggregation functions, and with top_hits removed,
geo_bounds works sometimes, and sometimes fails with a `response_2.json`
(in the example) or `response_3.json`.
With an enabled top_hits aggregator it
I think the problem is that you have an extra aggregations object
surrounding the geo_bounds aggregation. Try the following:
curl -XGET http://localhost:9200/postings/_search; -d'
{
query: {
filtered: {
filter: {
geo_bounding_box: {
Dear Anand,
I really appreciate your question because I face a similar problem, still
unsolved (you may refer to this post
https://groups.google.com/forum/?fromgroups#!topicsearchin/elasticsearch/suggest$20OR$20completion$20OR$20typeahead/elasticsearch/yiebM3dm4EE).
I hope someone comes up
Hi Mark,
thanks! Worked like a charm!
Best regards,
Pv
On 24 Jul 2014, at 00:58, Mark Walkom ma...@campaignmonitor.com wrote:
If this is production you really want an odd number of nodes to reduce
potential split brain issues.
However in your case, just add the new node to the cluster,
My answer is very late, but here :
?php
require '../vendor/autoload.php';
$client = new Elasticsearch\Client();
$params = array();
$params['body'] = array('testField' = 'abc');
$params['index'] = 'my_index';
$params['type'] = 'my_type';
hi folks,
I am frequently getting this error :-
*2014-07-24 01:45:37 Commons Daemon procrun stdout initialized*
*java.lang.OutOfMemoryError: Java heap space*
*Dumping heap to java_pid84012.hprof ...*
*Heap dump file created [1081959347 bytes in 4.653 secs]*
to avoid this I have increase the es
One of the coolest and most celebrated features of Elasticsearch is its
ability to index JSON in what we know to be a “quasi schemaless” fashion.
Elasticsearch does this by automatically flattening fields, whenever
possible, while resorting to nested field (“Blockjoin”) when objects are
Thanks the reply Adrien.
I've taken another look and confirm that it was an error on my part (sorry)!
It is behaving as expected - my error was the term based filter was working
correctly and when changing it to a non-matching term it filtered out
everything as is expected.
Thanks again,
Hi Ivan,
(P.S.: I am one of the developer of the SIREn plugin)
it would be possible for SIREn to support such functionality (but it is not
yet implemented), as each element / node in the tree as a unique identifier
that is retrieved at search time. Therefore, one could use this identifier
to
Hi Brian,
Our apologies for the issues with the web site, we had some problems on our
web server yesterday.
What you have described is very close to the indexing model in SIREn. SIREn
provides an optimised Lucene's Codec for such data structure, and provide
query operators on top of this
Hi David,
does that mean in conclusion that there will be a rss input for logstash
soon as replacement of the river? Same goes for wikipedia?
Greets,
Andrej
Am Mittwoch, 23. Juli 2014 14:13:51 UTC+2 schrieb David Pilato:
Rivers will be deprecated in favor of logstash inputs.
Deprecated does
Hi Andrej
That's the plan indeed.
About soon, I don't know :)
--
David Pilato | Technical Advocate | Elasticsearch.com
@dadoonet | @elasticsearchfr
Le 24 juillet 2014 à 13:42:25, Andrej Rosenheinrich
(andrej.rosenheinr...@unister.de) a écrit:
Hi David,
does that mean in conclusion that
Hi,
Let's say I have 2 indices: *one whose name is index_name and an other
index with alias **index_name *(called *second_index*)*.*
In this way there is a conflict.
*How ElasticSearch handles such situation?*
Does it give priority to the alias (this means when you are using
*index_name* it
We're about to stand up an elasticsearch cluster and we're facing the task
of determining the correct number of shards to allocate for our single
index. We have 30 servers (16cpus, 48gb memory), each of which hosting one
node. The concern is only with query performance. Indexing performance
Hello,
in EC@,i installed elasticsearch with
wget
http://download.elasticsearch.org/elasticsearch/elasticsearch/elasticsearch-1.3.0.zip
sudo unzip elasticsearch-1.3.0.zip -d /usr/local/elasticsearch
cd /usr/local/elasticsearch/elasticsearch-1.3.0
sudo bin/plugin -install
Hi David,
Any update on upgrading FSRiver with elastic-1.2.1.
Thanks,
chandra shekar.k
On Tuesday, 1 July 2014 23:54:46 UTC+5:30, Alain Désilets wrote:
I am trying to use the FSRiver with elastic-1.2.1.
I installed the plugin as follows:
bin\plugin -install
Hi,
*PREDICTIVE SEARCH:*
How can we achieve the feature predictive search with ElasticSearch API.
can any one help me if any body worked on predictive search feature in
elasticsearch API.
We are using Elasticsearch 1.2.1 version.
*Differences Between ElasticSearch 1.0.1 and
Cool! So for the new feature...
All I need is an empty field called avgRespTime and in my mapping define a
transform for this field which compute the time difference between thw two
stamp times?
--
You received this message because you are subscribed to the Google Groups
elasticsearch group.
How can I use ES score in the script?
The following script works and I am able to sort the results based on this
script:
String script = doc['studyscorenormal'].value;
ScriptSortBuilder sortScript = SortBuilders.scriptSort(script,
number).order(SortOrder.DESC);
But when I am
Hi!
My cluster, currently default named, Damon Dran, appear to be working fine.
The head plugin shows yellow and queries work. I have indices and documents
added.
I added Marvel plugin, Jetty and set some protection in realm.properties,
but when I restarted node, I see the error
[ERROR]
Gateways have been removed. You can't use that anymore.
--
David Pilato | Technical Advocate | Elasticsearch.com
@dadoonet | @elasticsearchfr
Le 24 juillet 2014 à 14:47:48, Anthony Oleary (anthony.ole...@kweekweek.com) a
écrit:
Hello,
in EC@,i installed elasticsearch with
wget
Ok..thanks david ..one quick confirmation required ..any major changes in
1.2.1 over 1.0.1..?
On Thursday, 24 July 2014 19:16:12 UTC+5:30, David Pilato wrote:
Not yet here. Still work to do on official plugins first!
--
*David Pilato* | *Technical Advocate* | *Elasticsearch.com*
Hi,
I have indexed my logs into Elasticsearch. The logs contain requests and
responses to web services used by customers. I want to use Kibana to
graphically display usage of web services (requests) by customers over a
period of time. I also want to provide the ability to zoom in and out. I
Not yet here. Still work to do on official plugins first!
--
David Pilato | Technical Advocate | Elasticsearch.com
@dadoonet | @elasticsearchfr
Le 24 juillet 2014 à 15:05:36, shekarflex (shekarf...@gmail.com) a écrit:
Hi David,
Any update on upgrading FSRiver with elastic-1.2.1.
Dear,
I've upgraded from 1.2.2 to 1.3.0 today and I've found one issue.
I've a 3 node cluster (one without data) running on CentOS 6.5. I've done a
rolling upgrade: first upgraded the no data node (server_0), then shutdown
server_1, upgraded (with the RPM), and restarted it. All shards came
What do you mean? Are you talking about elasticsearch?
If so, a lot of changes between 1.0 and 1.2 and 1.3
If you are talking about FSRiver, I'm not sure yet about what we will have in
FSRiver 1.x versions.
--
David Pilato | Technical Advocate | Elasticsearch.com
@dadoonet | @elasticsearchfr
thanks David
I got the link about gateways from
http://www.elasticsearch.org/tutorials/elasticsearch-on-ec2/
what would you recommend i use for clustering on EC2 instead of gateways or
how to configure the config.yml ?
Anthony
On Thursday, July 24, 2014 2:47:32 PM UTC+1, David Pilato wrote:
david i am talking about elasticsearch not fsriver.
as we started working with version 1.2.1.
- we got the requirement to index binary document initially we tried with
elastic-search attachment plugin but it will index only base64 content so
if we index 50,100mb files system hangs and we quit
You should read this:
http://www.elasticsearch.org/blog/elasticsearch-1-1-0-released/
http://www.elasticsearch.org/blog/elasticsearch-1-2-0-released/
http://www.elasticsearch.org/blog/elasticsearch-1-3-0-released/
It will give you an overview of major changes between major releases.
--
David
Just use local disk or EBS with provisioned IOs.
You don't need to store your indices on S3. If you want to do that for backup
purpose, have a look at snapshot and restore API.
Basically, in elasticsearch.yml file, remove:
gateway.type: s3
gateway.s3.bucket: codetest
Gateway has been removed
Hi
I'm trying to disable DELETE API option from sense or via HTTP and was not
able to do it . My requirement is make DELETE API not available so that
user will not accidentally delete any index . I tried the following in .yml
file and restarted the server ,but still i'm able to DELETE any
Hello all,
I have indexed some documents using the ElasticSearch Java API. How can I
use Kibana to access these documents?
Thanks
Vivek K
--
You received this message because you are subscribed to the Google Groups
elasticsearch group.
To unsubscribe from this group and stop receiving
Hello, I am trying to follow the *field collapse example* on this page
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/search-aggregations-metrics-top-hits-aggregation.html#_field_collapse_example,
which uses the new 1.3.0 *top_hits* aggregator to return the top scoring
Basically there will be no conflict, because ElasticSearch does not allow
you to set an alias that is the same of an index name.
Solved!
Thanks :-)
On Thursday, July 24, 2014 2:27:12 PM UTC+2, Giovanni Gaglione wrote:
Hi,
Let's say I have 2 indices: *one whose name is index_name and an
I can search in documents for a term and get the count of that term in the
document with
{
query: {
query_string: {
query: myterm
}
},
script_fields: {
term_count: {
script: _index['text']['myterm'].tf()
}
}
}
Now suppose that the query is something like
Thanks again David,
One last question, how would i tell elasticsearch to use the EBS? or is it
when i create a EC2 to link it to EBS and it works away without telling
eleasticsearch.
would you recommend EBS in order for the data not to be lost.
On Thursday, July 24, 2014 3:48:53 PM UTC+1,
Hi,
the JDBC river/feeder plugin has been updated to Elasticsearch 1.3.0
https://github.com/jprante/elasticsearch-river-jdbc
As always, feedback/contribution is welcome!
Best,
Jörg
--
You received this message because you are subscribed to the Google Groups
elasticsearch group.
To
Title: Re: Top Hits aggregator from example not working for me
Hello!
The error is about scripting and dynamic scripting being disabled for MVEL. The simplest way to make it work is turning on dynamic scripting by adding script.disable_dynamic: false to your elasticsearch.yml file. However for
Hi!
I'm really enjoying all the possibilities brought about by the move from
facets to aggregations. However, I still can't figure out the relationship
between facets or buckets and analyzers. Is it not possible at all to get
the buckets out of an analyzed field?
Specifically, I need to
Well, don't know what happened but suddenly the shard replicated.
I was trying to copy the index to a new one with stream2es and when the
copy started the cluster state turned green...
The problem, for me, is solved.
On Thursday, July 24, 2014 11:13:41 AM UTC-3, Antonio Augusto Santos wrote:
Thanks, Rafal. The *top_hits *aggregator is working now, but could you
explain why this is a security risk? We will need to use this feature in
production and now I am feeling uneasy about it.
On Thursday, July 24, 2014 10:28:32 AM UTC-5, Rafał Kuć wrote:
Hello!
The error is about
Title: Re: Top Hits aggregator from example not working for me
Hello Daniel!
MVEL scripts can't be sandboxed and they provide a lot of features, however when running Elasticsearch with too much user rights one can exploit that situation, i.e. http://bouk.co/blog/elasticsearch-rce/. This
Hello Team,
We are using Elastic search for last 2years and we are currently in 0.90.2
release which was upgraded on Jan 2014.
For our requirement aggregation (top_hits) fits and we planned to upgrade to
1.3.0
We are getting following exception when elastic search tries to recover
shards.
It's a matter of EC2 configuration. Elasticsearch does not really care to know.
I'd probably start with local disks if possible. Replication is done by
elasticsearch. So if you have more than one node, your data could be replicated
on another machine.
--
David Pilato | Technical Advocate |
@Jeremy: I've placed the newrelic installer in /usr/share/elasticsearch and
ran the installer but it would not work. How did you get it working?
On Friday, July 11, 2014 8:21:04 AM UTC+2, Jérémy Lecour wrote:
Le jeudi 14 février 2013 01:59:13 UTC+1, Shairon Toledo a écrit :
I did that,
I am not sure if there is a cleaner way to bypassing score, but if you
explicitly sort against another value that is not the score, then by
default scoring will not occur.
Perhaps if you trace the code for sort, you can find a setting that
disables scoring in general.
--
Ivan
On Wed, Jul 23,
Thanks for chiming in Renaud. Hopefully I will have a chance to test out
the plugin soon. My use case for nested documents is fairly simple.
--
Ivan
On Thu, Jul 24, 2014 at 4:00 AM, ren...@sindicetech.com wrote:
Hi Brian,
Our apologies for the issues with the web site, we had some problems
You can switch to Lucene expression language to access score.
The example in the docs is a bit unfortunate, as it uses MVEL, which has to
be enabled in ES 1.3 explicitly.
Here is a modified example I made using the expression language:
https://gist.github.com/jprante/7a1763c7073f2b8084a6
Jörg
Hi,
I'm trying to insert data into elasticsearch from pig. I can insert into ES
on the same node as Pig just fine, but when I try to insert it into ES on
another node, I'm running into errors. I'm using the following command -
STORE result INTO 'medi/hps_with_prev' USING
One help, i am new user to elastic search ..like to know the recommended
java version for the elastic search 1.1.0
On Tuesday, July 1, 2014 1:36:44 PM UTC-5, Clinton Gormley wrote:
Hi all
Recently we've had a few spam emails that have made it through Google's
filters, and there have been a
Hi,
Is there a way to index rest web service (lets say that web service gives
some list of XML objects)..?
I am having a requirement to index the list of product details which will
be returned from a web service.
Any suggestions will be helpful to me.
Thanks,
Srinivas
--
You received
Hi,
The problem: Scroll scan does not return all data. The problem is not
permanent, elasticsearch might return 100% of data or only 50% and it never
throws an exception. A good indicator of the problem is _shard section in
the response, e.g.:
_shards: {successful: 40, failed: 0, total: 50}
That gives me the following error:
{
error: SearchPhaseExecutionException[Failed to execute phase [query],
all shards failed; shardFailures
{[mzcfN2MzQjSKlWPHndRROA][personsearch][0]:
SearchParseException[[personsearch][0]: query[_all:icu],from[-1],size[-1]:
Parse Failure [Failed to parse
Also, I was hoping that this example will give me the entire contents of
the top scoring document, not just its score.
Here is the result I get when running this query:
https://gist.github.com/384d88ee0015009033d7.git
{
took: 6,
timed_out: false,
_shards: {
total: 5,
Hi Nikolas,
We tried the following:
{
persistent: {
cluster: {
routing: {
allocation: {
balance: {
index: 0.05,
shard: 0.05,
primary: 0.9,
threshold: 1.0
}
Hello,
We've got a big index with millions of documents describing a relationship
between entities. One entity and it's many friends. We score them in a
few different ways and later on query and show paged results by those
scores.
The question is how were warmers supposed to be used most
I manage my ES cluster using Ansible. Specifically I wrote this
role https://github.com/LaneCommunityCollege/aspects_elasticsearch to do
most of the work for me.
The issue right now is that if I make a change to my elasticsearch.yml
configuration, Ansible automatically issues a 'service
I don't imagine there is anything you are missing. I don't know where to go
from here. Sorry!
On Jul 24, 2014 3:20 PM, Daniel Schonfeld downwindab...@gmail.com wrote:
Hi Nikolas,
We tried the following:
{
persistent: {
cluster: {
routing: {
allocation: {
You're better off starting a new thread around this :)
(But take a look here
http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/setup.html
)
Regards,
Mark Walkom
Infrastructure Engineer
Campaign Monitor
email: ma...@campaignmonitor.com
web: www.campaignmonitor.com
On 25
Hello,
I am running a pretty simply filter query in elasticsearch; the took field
in elasticsearch's response indicates that the query is quite fast, and
takes only a couple miliseconds. If I run the request via a CURL call from
the command line and time that, it takes slightly longer around
Is there are a better solution for now, maybe something with the
aggregations?
On Friday, August 17, 2012 3:50:37 PM UTC+1, Martijn v Groningen wrote:
You mean like:
query for items: match_all
Facet A: facet_filter B:b, C:c
Facet B: facet_filter A:a, C:c
Facet C:
Sorry, one more quick note: the time goes down several hundred milliseconds
if I only fetch the first 3 results, instead of 100, which leads me to
believe the issue has something to do with serializing or the client's
handling of the ES response; however, I am not sure what could be the
I am passing in a full phrase and using a shingle filter to match sub
phrases in that string. What I would like is to be able to 'mark' the
initial request with what is matched and what is not matched. Is there a
way to highlight or get the offset position of where the match was made
within
The only way to load the config is with a restart of the service.
You could use the cat API to check on the status before reloading -
http://www.elasticsearch.org/guide/en/elasticsearch/reference/master/cat-health.html
Regards,
Mark Walkom
Infrastructure Engineer
Campaign Monitor
email:
Instead of stopping nodes, setting the enabled:false flag
(http://www.elasticsearch.org/guide/en/marvel/current/index.html#_the_monitoring_cluster)
is there a way to do this on the fly and have marvel stop
collecting/sending data?
seeing if there is a way to do this without having to restart
since we do not have it all setup yet, just wondering if you can do this
marvel.agent.enabled
on the fly like other configs, or if it is only a restart node option
On Thursday, July 24, 2014 4:23:36 PM UTC-7, Scott Decker wrote:
Instead of stopping nodes, setting the enabled:false flag (
Hi ,
I have one index having data. After that i created one new field called
TYPE. I want to update all records for newly added field(TYPE=TS) .
Can you please provide query for the above requirement.
Thanks in advance.
Thanks,
Srikanth.
--
You received this message because you are
It is really odd, but it failed for me on the precision levels 6 and 7 with
a ArrayIndexOutOfBoundsException, and working for 5 and 8 levels.
I had put a big query examples here:
https://gist.github.com/svartalf/5db600e7ea2907fbda3a
четверг, 24 июля 2014 г., 18:05:30 UTC+9 пользователь Colin
Bump nothing?
Everything was working fine for a few months and now I cant get much to
work. I went down to one node to reduce any conflicts between nodes.Deleted
and rebuilt indexes. Even reloaded data from the past, Issues I'm still
facing:
This shows up all the time: [2014-07-24
Just wanted to bring attention to the new and *very* useful transform
scripts that were introduced in elasticsearch 1.3 [1]. This feature allows
you to manipulate the source BEFORE it is indexed so you can do things like
add/remove fields, change field values, etc. Groovy scripts will be the
Hi all
I was tinkering a bit with ElasticSearch. I think my target runtime may
end up being embedded, just based on our deployment model. Probably not a
big deal to get multiple nodes talking to each other and no additional
hardware to deploy.
I was trying to start the application in
I wanted to do conditional copy_to and Andrian suggested implementing
scripted transforms instead. Much more flexible. They mesh well with the
shift to groovy too because groovy is much more stable. Stable enough to
run on every insert.
I'm glad you are excited by it. It was fun to build and I
Hello Srikanth ,
Update by query is not a feature in Elasticsearch , but there is a plugin
which can do it.
PLUGIN - https://github.com/yakaz/elasticsearch-action-updatebyquery
Thanks
Vineeth
On Fri, Jul 25, 2014 at 5:01 AM, srikanth ramineni ri.srika...@gmail.com
wrote:
Hi ,
I
Yea, I really like it. I have been thinking about the exact same thing for
a while but never had the time to put it together. I do have some things I
would like to add such as the ability to stop a document from being indexed
when the doc has/does not have a specific value. At any rate, great
Hi,
Is this link
http://www.elasticsearch.org/tutorials/elasticsearch-on-ec2/still
applicable on EC2, cause it's since August 2011 and the release of ES is
0.19..? But I used it as a guidelines.
(http://www.elasticsearch.org/tutorials/elasticsearch-on-ec2/)
--
You received this message
The only solution I could come out is also to modify the source code. I
just think it would be nice if there is a existing flag or configuration to
completely skip the score evaluation process.
Thank you, Ivan!
Ivan Brusic於 2014年7月25日星期五UTC+8上午1時03分53秒寫道:
I am not sure if there is a
84 matches
Mail list logo