Some shards on my ES-Instance get an IndexCorruptedException. I assume this
was caused by java 1.7u51, which may cause such issues. But how to recover
from this corrupted shard? The currently affected shards seems endless to
be recovered and the ES-logs grow up to several megabytes of size. I
Hi all,
I'm writing a web crawler in Node.js and indexing with ElasticSearch.
However, I've ran into a problem where the code hangs at the indexing
function.
Here's how the client is initialised:
var es_client = new elasticsearch.Client({
host: localhost:9200,
log:
Same problem here.
Log says:
Parse Failure [Failed to parse source
[{size:1,sort:{@timestamp:{order:desc}}}]]
Verified that the index(es) have the timestamp field.
Running Marvel 1.2.1 and Elasticsearch 1.3.2.
Any help would be appreciated as we cant get our Marvel up and running.
Den
Thanks a lot! Looks like this is exactly what we need.
Vagif
On Tuesday, November 25, 2014 4:15:42 AM UTC+1, chenlin rao wrote:
use templated or scripted dashboards, check:
http://www.elasticsearch.org/guide/en/kibana/current/templated-and-scripted-dashboards.html
2014-11-24 20:48
Hello,
We just had something very weird on our ES cluster, and I wanted to know if
this behavior was either documented or expected or...
ES version: 1.0.1
1 node:
master: yes
http: yes
data: no
3 nodes:
master: no
http: yes
data: yes
The virtual machine with the master node was
Hi I am new to ES but my company is starting to use it
When I set up an river I have scheduled it to check for data changes at an
30 min interval, my largest index on dev includes 230k documents but in
production is expected to grow to 300million docs
this 230k index is a heavy load on the
Can you post more information about the IndexCorruptedException, stacktrace,
or reproduction?
I have to correct me: CorruptIndexException
[2014-11-21 20:15:44,225][WARN ][cluster.action.shard ] [isetta]
[X][2] received shard failed for [ips-mac-mips-mac-k01-2014.08][2],
Hi Erlendur,
In your case, you should use the column strategy instead of simple one. The
column strategy requires two columns in the SQL DB.
- cerated_at
- update_at
Cheers, Ramy
Am Dienstag, 25. November 2014 11:04:17 UTC+1 schrieb Erlendur Hákonarson:
Hi I am new to ES but my company is
Sorry...
- created_at
- updated_at
Am Dienstag, 25. November 2014 11:55:18 UTC+1 schrieb Ramy:
Hi Erlendur,
In your case, you should use the column strategy instead of simple one.
The column strategy requires two columns in the SQL DB.
- cerated_at
- update_at
Cheers, Ramy
Am Dienstag,
Hello,
I am investigating the reason for ES to raise some exceptions while
processing data shipped by Logstash, and something doesn't make sense. The
software is running on Windows machines, and the log file is a JSON
document with dateTime fields formatted using invariant culture. Here's an
Setup:
Rsyslog provides log from mutiple servers to a central server. Directory
structure is like:
/var/log/rsyslog/SERVER-1/2014-10-21/*.log
/var/log/rsyslog/SERVER-2/2014-10-21/*.log
So multiple server directories each having multiple dates within and
multiple log files.
Now,
Newbie here. This is probably straightforward, but I wanted to get
best-practices recommendations if applicable.
My main ES use and query case is to search for a kind of Event. All Events
are associated with (i.e. contain) a Country object. Besides querying for
other kinds of Event data,
Thanks Ramy
but how does that strategy work
is there any doc on strategies I can view?
the only one I found was on the jprante github wiki and that only describes
the simple strategy
and if I am using tables from a system that I have no control over and
those columns created_at and updated_at
Hi,
the connection is lazy so it will only be opened once you make a
request - just instantiating a client like this will not create any
connections.
Once a connection is created the python client will try to hold on to
it as long as possible (it uses urllib3 to do the connection pooling
itself)
maybe this link will helps you:
https://github.com/jprante/elasticsearch-river-jdbc/pull/137
and this code snippet:
{
*strategy: column,*
type: jdbc,
jdbc: {
url: db server connect string,
user: username,
schedule: 0 20/30 * * * ?,
password: password,
index:
Hi,
In attempt to increase JVM heap size for Elasticsearch on a Windows box, I
defined the following environment variable:
ES_HEAP_SIZE=6g
(The machine has 12G RAM in total).
I have restarted ES and queried its memory usage using _nodes/stat HTTP GET
command. However, it seems that it's
No suggestions?
/A
On Thursday, November 13, 2014 4:44:35 PM UTC+1, Alexander Jiteg wrote:
Follow up question; What If I change the terms in the looked up filter
after the percolator has been created? Seems that the terms are cached but
even though I have experimented with cache
After I had uninstalled and reinstalled Elasticsearch Windows service, it
correctly increased the JVM heap size. Just restarting it without
reinstalling the service didn't work.
Vagif
On Tuesday, November 25, 2014 2:45:04 PM UTC+1, Vagif Abilov wrote:
Hi,
In attempt to increase JVM heap
Hi all --
Can anyone point out how one can enable debug-level logging for discovery
on a standalone client using the Node client? I have the following code,
trying to join a single-node remote cluster. The discovery times out after
30 seconds (yet transport works fine, and starting a
We have a mapping where one of the fields is an integer, but we want to
change this to a double. We want to avoid re-indexing, since there will be
a lot of documents at migration time. Hence, we were considering using a
multi_field (now apparently deprecated, but I guess the same applies for
Hi,
I'm using Spring Data Elasticsearch, and so some of the queries are
generated. I'd like to be able to log all queries that hit Elasticsearch,
simply to be able to better understand what is being generated and run.
I found some old (2010) discussions saying that this isn't possible - but
Hi Guys,
We are running a 3 node cluster, and each node returns a different number
of documents when issued a direct HTTP _count call.
The cluster holds about 150K documents and the differences range from 30~50
documents, but are still troubling.
This shouldn't be a split brain problem, since
minimum_master_nodes still doesn't protect you from all possible failure
scenarios, see http://aphyr.com/posts/317-call-me-maybe-elasticsearch
What version are you running?
--
Itamar Syn-Hershko
http://code972.com | @synhershko https://twitter.com/synhershko
Freelance Developer Consultant
Hi Eugen,
You can use the slowlog.log file to view queries that were sent: it
prints out the JSON of the query received.
This log is off by default and only prints logs for operations that took
longer than durations specific in the elasticsearch.yml config file.
You can turn it on and set the
Refactoring my statement from script to a straight update { doc,
upsert_as_doc } seems to have done the trick. So rather than diagnose
what's odd about the script, this has resolved my issue. Yeah, lazy
solution, but a more optimal one ;)
--
You received this message because you are
Pretty simple requirement:
Find all documents where the field vendorid has no value and give it a
value of 6
Would I do this with an update call and use script? I'm a little unclear
how. Or is there a more canonical way to do something like this?
--
You received this message because you are
1.0.0
On Tuesday, November 25, 2014 6:41:36 PM UTC+2, Itamar Syn-Hershko wrote:
minimum_master_nodes still doesn't protect you from all possible failure
scenarios, see http://aphyr.com/posts/317-call-me-maybe-elasticsearch
What version are you running?
--
Itamar Syn-Hershko
I did read about that option, thanks for letting me know that it's the way
to go. I though that there would be another way, but since there's not,
I'll go ahead and do that.
Thank you,
Eugen.
On Tue, Nov 25, 2014 at 6:46 PM, Yosi Haran y...@my6sense.com wrote:
Hi Eugen,
You can use the
Hi Dror,
On Tue, Nov 25, 2014 at 2:29 PM, Dror Atariah dror...@gmail.com wrote:
Hi Adrien,
I have two comments/questions:
1) For me, the documentation is still somehow confusing, and the
difference between the *cardinality* and *value_count* aggregations is
not 100% clear.
I have to
Thanks for your quick reply!
On Tue, Nov 25, 2014 at 6:41 PM, Adrien Grand
adrien.gr...@elasticsearch.com wrote:
On Tue, Nov 25, 2014 at 2:29 PM, Dror Atariah dror...@gmail.com wrote:
1) For me, the documentation is still somehow confusing, and the
difference between the *cardinality* and
I'm trying to use ElasticsearchTestCase in my test framework to start
getting ES deployed in production.
However, I'm in maven dependency hell and wanted some advice.
I'm trying to use this:
Hi Kevin,
You should look at some plugins code, such as:
https://github.com/elasticsearch/elasticsearch-analysis-icu
I think this should help.
Best.
David
Le 25 nov. 2014 à 20:45, Kevin Burton burtona...@gmail.com a écrit :
I'm trying to use ElasticsearchTestCase in my test framework to
yeah.. doing that I am still getting the following no class def error:
... this really sucks. Should just work out of the box. Testing is kind
of the first thing I would expect most people to want to do...
java.lang.NoClassDefFoundError: org/apache/lucene/index/LeafReader
at
That seems like a lot of work for little data reduction, especially as ES
compresses data as well.
Maybe someone else has an idea though.
On 25 November 2014 at 23:30, joshua harr joshua.h...@gmail.com wrote:
Newbie here. This is probably straightforward, but I wanted to get
best-practices
Thank you Mark!
On Monday, November 24, 2014 4:47:32 PM UTC-8, Mark Walkom wrote:
You should be able to search for that string in kibana to find more info
about it.
On 25 November 2014 at 11:04, shriyansh jain shriyan...@gmail.com
javascript: wrote:
Thank you mark for your reply. What
Hi! Sorry you've had a snag with this. This is a good catch! Thanks for
finding this.
The problem is in filter_by_timestamp:
dateregex = get_date_regex(timestring)
if object_type == 'index':
regex = ^ + prefix + ( + dateregex + ) + suffix + $
elif object_type ==
Hello,
Are the kibana indexes associated with the elasticsearch indexes in any
way. As I closed few elasticsearch indexes, all the kibana saved dashboards
disappeared.
Is there any way to bring back those dashboards.
Thanks!
Shriyansh
--
You received this message because you are
The kibana dashboards are saved in the kibana-int index.
On 26 November 2014 at 08:11, shriyansh jain shriyanshaj...@gmail.com
wrote:
Hello,
Are the kibana indexes associated with the elasticsearch indexes in any
way. As I closed few elasticsearch indexes, all the kibana saved dashboards
Just figured out they are stored in kibana-int index.!
Thanks!
Shriyansh
On Tuesday, November 25, 2014 1:12:20 PM UTC-8, shriyansh jain wrote:
Hello,
Are the kibana indexes associated with the elasticsearch indexes in any
way. As I closed few elasticsearch indexes, all the kibana saved
Hello,
Are the kibana indexes associated with the elasticsearch indexes in any
way. As I closed few elasticsearch indexes, all the kibana saved dashboards
disappeared.
Is there any way to bring back those indexes.
Thanks!
Shriyansh
--
You received this message because you are subscribed to
Hi,
On Mon, Nov 24, 2014 at 11:51 AM, Lasse Schou lassesc...@gmail.com wrote:
The way I understand the Shard Query Cache
http://www.elasticsearch.org/guide/en/elasticsearch/reference/1.4/index-modules-shard-query-cache.html,
it invalidates *ALL *cached results in a shard when the shard
Hi all,
I need to query an index with tens of millions of short documents.
The result set may contain 100,000 documents, and I need to process a
single field from each document. It those are simple stored fields in *.fdt
file - it will take forever +-.
I thought document values will
Got it, thanks.
Lasse
Den tirsdag den 25. november 2014 skrev Adrien Grand
adrien.gr...@elasticsearch.com:
Hi,
On Mon, Nov 24, 2014 at 11:51 AM, Lasse Schou lassesc...@gmail.com
javascript:_e(%7B%7D,'cvml','lassesc...@gmail.com'); wrote:
The way I understand the Shard Query Cache
Thanks so much.
But the answer is very frustrating.
Getting large result sets will always be slow - even if I need just a
single field.
Only aggregations and facets enjoy document fields - we commoners need to
dig our fields from the *.fdt file.
Bugger – and thanks again
--
You
Ended up resolving this. I had to take the 1.4.0 pom and use the
dependencies there instead of the dependencies from master which was
2.0.0-SNAPSHOT...
... things sort of work now but still running into some small problems.
--
You received this message because you are subscribed to the
I notice that ElasticsearchIntegrationTest doesn't open a port on 9300...
how would I go about connecting my own created transport clients to it?
I want to test my code that creates my own transport clients.
--
You received this message because you are subscribed to the Google Groups
Doing a thorough job of deleting all indexes got my installation back to
status yellow, apparently fully functional when I re-established my test
index with a new name. Sledgehammer, but I didn't need to sow the ground
with salt.
On Friday, November 21, 2014 11:21:26 AM UTC-8, Pitaga wrote:
You can get it to green if you get rid of the replicas - curl -XPUT
localhost:9200/*/_settings -d '{ index : { number_of_replicas : 0 } }'
On 26 November 2014 at 09:36, Pitaga ach...@blarg.net wrote:
Doing a thorough job of deleting all indexes got my installation back to
status yellow,
48 matches
Mail list logo