Thank you very much. Tony On Friday, August 29, 2014 3:27:33 PM UTC-4, Jörg Prante wrote: > > Quick guide: > > - install Java 7 (or Java 8), Apache Maven, and git, also ensure internet > connection to the Maven central repo > > - clone 1.3 branch only (you could also clone the whole repo and switch to > the branch): git clone https://github.com/elasticsearch/elasticsearch.git > --branch 1.3 --single-branch es-1.3 > > - enter folder es-1.3 > > - start build: mvn -DskipTests clean install > > - wait a few minutes while Maven loads all dependent artifacts and > compiles ~3000 source files > > The result will be a complete build of all binaries. In the 'target' > folder, after the "Build complete" message of Maven, you will see a file > "elasticsearch-<VERSION>.jar" > > <VERSION> is something like "1.3.3-SNAPSHOT". You can copy this file into > your existing Elasticsearch 1.3.x installation "lib" folder. Do not forget > to adjust "bin/elasticsearch.in.sh" to point to the new > elasticsearch-<VERSION>.jar file in the classpath configuration (at the top > lines). This must be the first jar on the classpath so it can patch Lucene > jars. > > If you have already data in the existing Elasticsearch I recommend to > backup everything before starting the new snapshot build - no guarantees, > use at your own risk. > > Jörg > > > > > On Fri, Aug 29, 2014 at 7:36 PM, <tony....@iqor.com <javascript:>> wrote: > >> The easiest for me is to install fresh binaries but I'm not shy about >> learning about Maven while I build it from source. >> >> Thanks >> Tony >> >> >> On Friday, August 29, 2014 11:21:34 AM UTC-4, Jörg Prante wrote: >> >>> Do you want to build from source? Or do you want to install a fresh >>> binary? >>> >>> At jenkins.elasticsearch.org I can not find any snapshot builds but it >>> may be just me. >>> >>> It would be a nice add-on to provide snapshot builds for users that >>> eagerly await bug fixes or take a ride on the bleeding edge before the next >>> release arrives, without release notes etc. >>> >>> Jörg >>> >>> >>> On Fri, Aug 29, 2014 at 4:29 PM, <tony....@iqor.com> wrote: >>> >>>> Thanks again and sorry to bother you guys but I'm new to Github and >>>> don't know what do do from here. Can you point me to the right place >>>> where >>>> I can take the next step to put this patch on my server? I only know how >>>> to untar the tarball I downloaded from the main ES page. >>>> >>>> Thanks. >>>> Tony >>>> >>>> >>>> On Wednesday, August 27, 2014 1:35:06 PM UTC-4, tony....@iqor.com >>>> wrote: >>>>> >>>>> Kudos! >>>>> >>>>> Tony >>>>> >>>>> On Wednesday, August 27, 2014 1:16:11 PM UTC-4, Jörg Prante wrote: >>>>>> >>>>>> All praise should go to the fantastic Elasticsearch team who did not >>>>>> hesitate to test the fix immediately and replaced it with a better >>>>>> working >>>>>> solution, since the lzf-compress software is having weaknesses regarding >>>>>> threadsafety. >>>>>> >>>>>> Jörg >>>>>> >>>>>> >>>>>> On Wed, Aug 27, 2014 at 7:01 PM, Ivan Brusic <iv...@brusic.com> >>>>>> wrote: >>>>>> >>>>>>> Amazing job. Great work. >>>>>>> >>>>>>> -- >>>>>>> Ivan >>>>>>> >>>>>>> >>>>>>> On Tue, Aug 26, 2014 at 12:41 PM, joerg...@gmail.com < >>>>>>> joerg...@gmail.com> wrote: >>>>>>> >>>>>>>> I fixed the issue by setting the safe LZF encoder in LZFCompressor >>>>>>>> and opened a pull request >>>>>>>> >>>>>>>> https://github.com/elasticsearch/elasticsearch/pull/7466 >>>>>>>> >>>>>>>> Jörg >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Aug 26, 2014 at 8:17 PM, joerg...@gmail.com < >>>>>>>> joerg...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Still broken with lzf-compress 1.0.3 >>>>>>>>> >>>>>>>>> https://gist.github.com/jprante/d2d829b497db4963aea5 >>>>>>>>> >>>>>>>>> Jörg >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, Aug 26, 2014 at 7:54 PM, joerg...@gmail.com < >>>>>>>>> joerg...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Thanks for the logstash mapping command. I can reproduce it now. >>>>>>>>>> >>>>>>>>>> It's the LZF encoder that bails out at org.elasticsearch.common. >>>>>>>>>> compress.lzf.impl.UnsafeChunkEncoderBE._getInt >>>>>>>>>> >>>>>>>>>> which uses in turn sun.misc.Unsafe.getInt >>>>>>>>>> >>>>>>>>>> I have created a gist of the JVM crash file at >>>>>>>>>> >>>>>>>>>> https://gist.github.com/jprante/79f4b4c0b9fd83eb1c9b >>>>>>>>>> >>>>>>>>>> There has been a fix in LZF lately https://github.com/ning >>>>>>>>>> /compress/commit/db7f51bddc5b7beb47da77eeeab56882c650bff7 >>>>>>>>>> >>>>>>>>>> for version 1.0.3 which has been released recently. >>>>>>>>>> >>>>>>>>>> I will build a snapshot ES version with LZF 1.0.3 and see if this >>>>>>>>>> works... >>>>>>>>>> >>>>>>>>>> Jörg >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Mon, Aug 25, 2014 at 11:30 PM, <tony....@iqor.com> wrote: >>>>>>>>>> >>>>>>>>>>> I captured a WireShark trace of the interaction between ES and >>>>>>>>>>> Logstash 1.4.1. The error occurs even before my data is sent. Can >>>>>>>>>>> you try >>>>>>>>>>> to reproduce it on your testbed with this message I captured? >>>>>>>>>>> >>>>>>>>>>> curl -XPUT http://amssc103-mgmt-app2:9200/_template/logstash -d >>>>>>>>>>> @y >>>>>>>>>>> >>>>>>>>>>> Contests of file 'y": >>>>>>>>>>> { "template" : "logstash-*", "settings" : { >>>>>>>>>>> "index.refresh_interval" : "5s" }, "mappings" : { "_default_" >>>>>>>>>>> : { >>>>>>>>>>> "_all" : {"enabled" : true}, "dynamic_templates" : [ { >>>>>>>>>>> >>>>>>>>>>> "string_fields" : { "match" : "*", >>>>>>>>>>> "match_mapping_type" >>>>>>>>>>> : "string", "mapping" : { "type" : "string", >>>>>>>>>>> "index" >>>>>>>>>>> : "analyzed", "omit_norms" : true, "fields" : { >>>>>>>>>>> >>>>>>>>>>> "raw" : {"type": "string", "index" : "not_analyzed", >>>>>>>>>>> "ignore_above" : >>>>>>>>>>> 256} } } } } ], >>>>>>>>>>> "properties" : >>>>>>>>>>> { "@version": { "type": "string", "index": "not_analyzed" >>>>>>>>>>> }, >>>>>>>>>>> "geoip" : { "type" : "object", "dynamic": >>>>>>>>>>> true, >>>>>>>>>>> "path": "full", "properties" : { >>>>>>>>>>> >>>>>>>>>>> "location" : { "type" : "geo_point" } } } >>>>>>>>>>> } } >>>>>>>>>>> }} >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Monday, August 25, 2014 3:53:18 PM UTC-4, tony....@iqor.com >>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>> I have no plugins installed (yet) and only changed >>>>>>>>>>>> "es.logger.level" to DEBUG in logging.yml. >>>>>>>>>>>> >>>>>>>>>>>> elasticsearch.yml: >>>>>>>>>>>> cluster.name: es-AMS1Cluster >>>>>>>>>>>> node.name: "KYLIE1" >>>>>>>>>>>> node.rack: amssc2client02 >>>>>>>>>>>> path.data: /export/home/apontet/elasticsearch/data >>>>>>>>>>>> path.work: /export/home/apontet/elasticsearch/work >>>>>>>>>>>> path.logs: /export/home/apontet/elasticsearch/logs >>>>>>>>>>>> network.host: ******** <===== sanitized line; file >>>>>>>>>>>> contains actual server IP >>>>>>>>>>>> discovery.zen.ping.multicast.enabled: false >>>>>>>>>>>> discovery.zen.ping.unicast.hosts: ["s1", "s2", "s3", "s5" , >>>>>>>>>>>> "s6", "s7"] <===== Also sanitized >>>>>>>>>>>> >>>>>>>>>>>> Thanks, >>>>>>>>>>>> Tony >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Saturday, August 23, 2014 6:29:40 AM UTC-4, Jörg Prante >>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> I tested a simple "Hello World" document on Elasticsearch >>>>>>>>>>>>> 1.3.2 with Oracle JDK 1.7.0_17 64-bit Server VM, Sparc Solaris >>>>>>>>>>>>> 10, default >>>>>>>>>>>>> settings. >>>>>>>>>>>>> >>>>>>>>>>>>> No issues. >>>>>>>>>>>>> >>>>>>>>>>>>> So I would like to know more about the settings in >>>>>>>>>>>>> elasticsearch.yml, the mappings, and the installed plugins. >>>>>>>>>>>>> >>>>>>>>>>>>> Jörg >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Sat, Aug 23, 2014 at 11:25 AM, joerg...@gmail.com < >>>>>>>>>>>>> joerg...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> I have some Solaris 10 Sparc V440/V445 servers available and >>>>>>>>>>>>>> can try to reproduce over the weekend. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Jörg >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Sat, Aug 23, 2014 at 4:37 AM, Robert Muir < >>>>>>>>>>>>>> rober...@elasticsearch.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> How big is it? Maybe i can have it anyway? I pulled two >>>>>>>>>>>>>>> ancient ultrasparcs out of my closet to try to debug your >>>>>>>>>>>>>>> issue, but >>>>>>>>>>>>>>> unfortunately they are a pita to work with (dead nvram battery >>>>>>>>>>>>>>> on both, >>>>>>>>>>>>>>> zeroed mac address, etc.) Id still love to get to the bottom of >>>>>>>>>>>>>>> this. >>>>>>>>>>>>>>> On Aug 22, 2014 3:59 PM, <tony....@iqor.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Adrien, >>>>>>>>>>>>>>>> It's a bunch of garbled binary data, basically a dump of >>>>>>>>>>>>>>>> the process image. >>>>>>>>>>>>>>>> Tony >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Thursday, August 21, 2014 6:36:12 PM UTC-4, Adrien Grand >>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi Tony, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Do you have more information in the core dump file? (cf. >>>>>>>>>>>>>>>>> the "Core dump written" line that you pasted) >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Aug 21, 2014 at 7:53 PM, <tony....@iqor.com> >>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>>>>>> I installed ES 1.3.2 on a spare Solaris 11/ T4-4 SPARC >>>>>>>>>>>>>>>>>> server to scale out of small x86 machine. I get a similar >>>>>>>>>>>>>>>>>> exception >>>>>>>>>>>>>>>>>> running ES with JAVA_OPTS=-d64. When Logstash 1.4.1 sends >>>>>>>>>>>>>>>>>> the first >>>>>>>>>>>>>>>>>> message I get the error below on the ES process: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>> # A fatal error has been detected by the Java Runtime >>>>>>>>>>>>>>>>>> Environment: >>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>> # SIGBUS (0xa) at pc=0xffffffff7a9a3d8c, pid=14473, >>>>>>>>>>>>>>>>>> tid=209 >>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>> # JRE version: 7.0_25-b15 >>>>>>>>>>>>>>>>>> # Java VM: Java HotSpot(TM) 64-Bit Server VM (23.25-b01 >>>>>>>>>>>>>>>>>> mixed mode solaris-sparc compressed oops) >>>>>>>>>>>>>>>>>> # Problematic frame: >>>>>>>>>>>>>>>>>> # V [libjvm.so+0xba3d8c] Unsafe_GetInt+0x158 >>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>> # Core dump written. Default location: >>>>>>>>>>>>>>>>>> /export/home/elasticsearch/elasticsearch-1.3.2/core or >>>>>>>>>>>>>>>>>> core.14473 >>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>> # If you would like to submit a bug report, please visit: >>>>>>>>>>>>>>>>>> # http://bugreport.sun.com/bugreport/crash.jsp >>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> --------------- T H R E A D --------------- >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Current thread (0x0000000107078000): JavaThread >>>>>>>>>>>>>>>>>> "elasticsearch[KYLIE1][http_server_worker][T#17]{New I/O >>>>>>>>>>>>>>>>>> worker #147}" daemon [_thread_in_vm, id=209, >>>>>>>>>>>>>>>>>> stack(0xffffffff5b800000, >>>>>>>>>>>>>>>>>> 0xffffffff5b840000)] >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> siginfo:si_signo=SIGBUS: si_errno=0, si_code=1 >>>>>>>>>>>>>>>>>> (BUS_ADRALN), si_addr=0x0000000709cc09e7 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I can run ES using 32bit java but have to shrink >>>>>>>>>>>>>>>>>> ES_HEAPS_SIZE more than I want to. Any assistance would be >>>>>>>>>>>>>>>>>> appreciated. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>>>> Tony >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Tuesday, July 22, 2014 5:43:28 AM UTC-4, David Roberts >>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> After upgrading from Elasticsearch 1.0.1 to 1.2.2 I'm >>>>>>>>>>>>>>>>>>> getting JVM core dumps on Solaris 10 on SPARC. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> # A fatal error has been detected by the Java Runtime >>>>>>>>>>>>>>>>>>> Environment: >>>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>>> # SIGBUS (0xa) at pc=0xffffffff7e452d78, pid=15483, >>>>>>>>>>>>>>>>>>> tid=263 >>>>>>>>>>>>>>>>>>> # >>>>>>>>>>>>>>>>>>> # JRE version: Java(TM) SE Runtime Environment >>>>>>>>>>>>>>>>>>> (7.0_55-b13) (build 1.7.0_55-b13) >>>>>>>>>>>>>>>>>>> # Java VM: Java HotSpot(TM) 64-Bit Server VM (24.55-b03 >>>>>>>>>>>>>>>>>>> mixed mode solaris-sparc compressed oops) >>>>>>>>>>>>>>>>>>> # Problematic frame: >>>>>>>>>>>>>>>>>>> # V [libjvm.so+0xc52d78] Unsafe_GetLong+0x158 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> I'm pretty sure the problem here is that Elasticsearch >>>>>>>>>>>>>>>>>>> is making increasing use of "unsafe" functions in Java, >>>>>>>>>>>>>>>>>>> presumably to speed >>>>>>>>>>>>>>>>>>> things up, and some CPUs are more picky than others about >>>>>>>>>>>>>>>>>>> memory >>>>>>>>>>>>>>>>>>> alignment. In particular, x86 will tolerate misaligned >>>>>>>>>>>>>>>>>>> memory access >>>>>>>>>>>>>>>>>>> whereas SPARC won't. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Somebody has tried to report this to Oracle in the past >>>>>>>>>>>>>>>>>>> and (understandably) Oracle has said that if you're going >>>>>>>>>>>>>>>>>>> to use unsafe >>>>>>>>>>>>>>>>>>> functions you need to understand what you're doing: >>>>>>>>>>>>>>>>>>> http://bugs.java.com/bugdatabase/view_bug.do?bug_ >>>>>>>>>>>>>>>>>>> id=8021574 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> A quick grep through the code of the two versions of >>>>>>>>>>>>>>>>>>> Elasticsearch shows that the new use of "unsafe" memory >>>>>>>>>>>>>>>>>>> access functions is >>>>>>>>>>>>>>>>>>> in the BytesReference, MurmurHash3 and HyperLogLogPlusPlus >>>>>>>>>>>>>>>>>>> classes: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> bash-3.2$ git checkout v1.0.1 >>>>>>>>>>>>>>>>>>> Checking out files: 100% (2904/2904), done. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> bash-3.2$ find . -name '*.java' | xargs grep UnsafeUtils >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/util/ >>>>>>>>>>>>>>>>>>> UnsafeUtils.java:public enum UnsafeUtils { >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/search/aggregations/ >>>>>>>>>>>>>>>>>>> bucket/BytesRefHash.java: if (id == -1L || >>>>>>>>>>>>>>>>>>> UnsafeUtils.equals(key, get(id, spare))) { >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/search/aggregations/ >>>>>>>>>>>>>>>>>>> bucket/BytesRefHash.java: } else if >>>>>>>>>>>>>>>>>>> (UnsafeUtils.equals(key, get(curId, spare))) { >>>>>>>>>>>>>>>>>>> ./src/test/java/org/elasticsearch/benchmark/common/util/ >>>>>>>>>>>>>>>>>>> BytesRefComparisonsBenchmark.java:import >>>>>>>>>>>>>>>>>>> org.elasticsearch.common.util.UnsafeUtils; >>>>>>>>>>>>>>>>>>> ./src/test/java/org/elasticsearch/benchmark/common/util/ >>>>>>>>>>>>>>>>>>> BytesRefComparisonsBenchmark.java: >>>>>>>>>>>>>>>>>>> return UnsafeUtils.equals(b1, b2); >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> bash-3.2$ git checkout v1.2.2 >>>>>>>>>>>>>>>>>>> Checking out files: 100% (2220/2220), done. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> bash-3.2$ find . -name '*.java' | xargs grep UnsafeUtils >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/bytes/ >>>>>>>>>>>>>>>>>>> BytesReference.java:import >>>>>>>>>>>>>>>>>>> org.elasticsearch.common.util.UnsafeUtils; >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/bytes/ >>>>>>>>>>>>>>>>>>> BytesReference.java: return >>>>>>>>>>>>>>>>>>> UnsafeUtils.equals(a.array(), a.arrayOffset(), b.array(), >>>>>>>>>>>>>>>>>>> b.arrayOffset(), >>>>>>>>>>>>>>>>>>> a.length()); >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/hash/ >>>>>>>>>>>>>>>>>>> MurmurHash3.java:import org.elasticsearch.common.util. >>>>>>>>>>>>>>>>>>> UnsafeUtils; >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/hash/ >>>>>>>>>>>>>>>>>>> MurmurHash3.java: return >>>>>>>>>>>>>>>>>>> UnsafeUtils.readLongLE(key, blockOffset); >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/hash/ >>>>>>>>>>>>>>>>>>> MurmurHash3.java: long k1 = >>>>>>>>>>>>>>>>>>> UnsafeUtils.readLongLE(key, i); >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/hash/ >>>>>>>>>>>>>>>>>>> MurmurHash3.java: long k2 = >>>>>>>>>>>>>>>>>>> UnsafeUtils.readLongLE(key, i + 8); >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/util/ >>>>>>>>>>>>>>>>>>> BytesRefHash.java: if (id == -1L || >>>>>>>>>>>>>>>>>>> UnsafeUtils.equals(key, get(id, spare))) { >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/util/ >>>>>>>>>>>>>>>>>>> BytesRefHash.java: } else if >>>>>>>>>>>>>>>>>>> (UnsafeUtils.equals(key, get(curId, spare))) { >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/common/util/ >>>>>>>>>>>>>>>>>>> UnsafeUtils.java:public enum UnsafeUtils { >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/search/aggregations/ >>>>>>>>>>>>>>>>>>> metrics/cardinality/HyperLogLogPlusPlus.java:import >>>>>>>>>>>>>>>>>>> org.elasticsearch.common.util.UnsafeUtils; >>>>>>>>>>>>>>>>>>> ./src/main/java/org/elasticsearch/search/aggregations/ >>>>>>>>>>>>>>>>>>> metrics/cardinality/HyperLogLogPlusPlus.java: >>>>>>>>>>>>>>>>>>> return UnsafeUtils.readIntLE(readSpare.bytes, >>>>>>>>>>>>>>>>>>> readSpare.offset); >>>>>>>>>>>>>>>>>>> ./src/test/java/org/elasticsearch/benchmark/common/util/ >>>>>>>>>>>>>>>>>>> BytesRefComparisonsBenchmark.java:import >>>>>>>>>>>>>>>>>>> org.elasticsearch.common.util.UnsafeUtils; >>>>>>>>>>>>>>>>>>> ./src/test/java/org/elasticsearch/benchmark/common/util/ >>>>>>>>>>>>>>>>>>> BytesRefComparisonsBenchmark.java: >>>>>>>>>>>>>>>>>>> return UnsafeUtils.equals(b1, b2); >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Presumably one of these three new uses is what is >>>>>>>>>>>>>>>>>>> causing the JVM SIGBUS error I'm seeing. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> A quick look at the MurmurHash3 class shows that the >>>>>>>>>>>>>>>>>>> hash128 method accepts an arbitrary offset and passes it to >>>>>>>>>>>>>>>>>>> an unsafe >>>>>>>>>>>>>>>>>>> function with no check that it's a multiple of 8: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> public static Hash128 hash128(byte[] key, int >>>>>>>>>>>>>>>>>>> offset, int length, long seed, Hash128 hash) { >>>>>>>>>>>>>>>>>>> long h1 = seed; >>>>>>>>>>>>>>>>>>> long h2 = seed; >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> if (length >= 16) { >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> final int len16 = length & 0xFFFFFFF0; // >>>>>>>>>>>>>>>>>>> higher multiple of 16 that is lower than or equal to length >>>>>>>>>>>>>>>>>>> final int end = offset + len16; >>>>>>>>>>>>>>>>>>> for (int i = offset; i < end; i += 16) { >>>>>>>>>>>>>>>>>>> long k1 = UnsafeUtils.readLongLE(key, i); >>>>>>>>>>>>>>>>>>> long k2 = UnsafeUtils.readLongLE(key, i >>>>>>>>>>>>>>>>>>> + 8); >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> This is a recipe for generating JVM core dumps on >>>>>>>>>>>>>>>>>>> architectures such as SPARC, Itanium and PowerPC that don't >>>>>>>>>>>>>>>>>>> support >>>>>>>>>>>>>>>>>>> unaligned 64 bit memory access. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Does Elasticsearch have any policy for support of >>>>>>>>>>>>>>>>>>> hardware other than x86? If not, I don't think many people >>>>>>>>>>>>>>>>>>> would care but >>>>>>>>>>>>>>>>>>> you really ought to clearly say so on your platform support >>>>>>>>>>>>>>>>>>> page. If you >>>>>>>>>>>>>>>>>>> do intend to support non-x86 architectures then you need to >>>>>>>>>>>>>>>>>>> be much more >>>>>>>>>>>>>>>>>>> careful about the use of unsafe memory accesses. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> David >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> You received this message because you are subscribed to >>>>>>>>>>>>>>>>>> the Google Groups "elasticsearch" group. >>>>>>>>>>>>>>>>>> To unsubscribe from this group and stop receiving emails >>>>>>>>>>>>>>>>>> from it, send an email to elasticsearc...@googlegroups.c >>>>>>>>>>>>>>>>>> om. >>>>>>>>>>>>>>>>>> To view this discussion on the web visit >>>>>>>>>>>>>>>>>> https://groups.google.com/d/msgid/elasticsearch/eb7f4c23- >>>>>>>>>>>>>>>>>> b63e-4c2e-87c3-029fc58449fc%40googlegroups.com >>>>>>>>>>>>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/eb7f4c23-b63e-4c2e-87c3-029fc58449fc%40googlegroups.com?utm_medium=email&utm_source=footer> >>>>>>>>>>>>>>>>>> . >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> For more options, visit https://groups.google.com/d/op >>>>>>>>>>>>>>>>>> tout. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> Adrien Grand >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> You received this message because you are subscribed to the >>>>>>>>>>>>>>>> Google Groups "elasticsearch" group. >>>>>>>>>>>>>>>> To unsubscribe from this group and stop receiving emails >>>>>>>>>>>>>>>> from it, send an email to elasticsearc...@googlegroups.com. >>>>>>>>>>>>>>>> To view this discussion on the web visit >>>>>>>>>>>>>>>> https://groups.google.com/d/msgid/elasticsearch/12aa33de- >>>>>>>>>>>>>>>> ccc7-485a-8c52-562f3e91a535%40googlegroups.com >>>>>>>>>>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/12aa33de-ccc7-485a-8c52-562f3e91a535%40googlegroups.com?utm_medium=email&utm_source=footer> >>>>>>>>>>>>>>>> . >>>>>>>>>>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> You received this message because you are subscribed to the >>>>>>>>>>>>>>> Google Groups "elasticsearch" group. >>>>>>>>>>>>>>> To unsubscribe from this group and stop receiving emails >>>>>>>>>>>>>>> from it, send an email to elasticsearc...@googlegroups.com. >>>>>>>>>>>>>>> To view this discussion on the web visit >>>>>>>>>>>>>>> https://groups.google.com/d/msgid/elasticsearch/CAMUKNZXOKeJ >>>>>>>>>>>>>>> q8Datx2KY7cSfJXDH1YGDNmQjNWDQ2jci%3DfN31Q%40mail.gmail.com >>>>>>>>>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/CAMUKNZXOKeJq8Datx2KY7cSfJXDH1YGDNmQjNWDQ2jci%3DfN31Q%40mail.gmail.com?utm_medium=email&utm_source=footer> >>>>>>>>>>>>>>> . >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>> You received this message because you are subscribed to the >>>>>>>>>>> Google Groups "elasticsearch" group. >>>>>>>>>>> To unsubscribe from this group and stop receiving emails from >>>>>>>>>>> it, send an email to elasticsearc...@googlegroups.com. >>>>>>>>>>> To view this discussion on the web visit >>>>>>>>>>> https://groups.google.com/d/msgid/elasticsearch/c62191ea-543 >>>>>>>>>>> b-462d-95e9-aff125c0a6f0%40googlegroups.com >>>>>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/c62191ea-543b-462d-95e9-aff125c0a6f0%40googlegroups.com?utm_medium=email&utm_source=footer> >>>>>>>>>>> . >>>>>>>>>>> >>>>>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> -- >>>>>>>> You received this message because you are subscribed to the Google >>>>>>>> Groups "elasticsearch" group. >>>>>>>> To unsubscribe from this group and stop receiving emails from it, >>>>>>>> send an email to elasticsearc...@googlegroups.com. >>>>>>>> To view this discussion on the web visit >>>>>>>> https://groups.google.com/d/msgid/elasticsearch/CAKdsXoHrOOq >>>>>>>> hgOSiRhmweSR5wLs%2BJiO70_CSRO%2BFS2zOU9VKzg%40mail.gmail.com >>>>>>>> <https://groups.google.com/d/msgid/elasticsearch/CAKdsXoHrOOqhgOSiRhmweSR5wLs%2BJiO70_CSRO%2BFS2zOU9VKzg%40mail.gmail.com?utm_medium=email&utm_source=footer> >>>>>>>> . >>>>>>>> >>>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> You received this message because you are subscribed to the Google >>>>>>> Groups "elasticsearch" group. >>>>>>> To unsubscribe from this group and stop receiving emails from it, >>>>>>> send an email to elasticsearc...@googlegroups.com. >>>>>>> To view this discussion on the web visit >>>>>>> https://groups.google.com/d/msgid/elasticsearch/CALY%3DcQAvC >>>>>>> oJwN8tSJXa8%3DZHMDYw_mpHc0Q866fcso_1LZCFiyw%40mail.gmail.com >>>>>>> <https://groups.google.com/d/msgid/elasticsearch/CALY%3DcQAvCoJwN8tSJXa8%3DZHMDYw_mpHc0Q866fcso_1LZCFiyw%40mail.gmail.com?utm_medium=email&utm_source=footer> >>>>>>> . >>>>>>> >>>>>>> For more options, visit https://groups.google.com/d/optout. >>>>>>> >>>>>> >>>>>> -- >>>> You received this message because you are subscribed to the Google >>>> Groups "elasticsearch" group. >>>> To unsubscribe from this group and stop receiving emails from it, send >>>> an email to elasticsearc...@googlegroups.com. >>>> To view this discussion on the web visit https://groups.google.com/d/ >>>> msgid/elasticsearch/205b2250-f6ee-4141-a49a-ff22aade0fe9% >>>> 40googlegroups.com >>>> <https://groups.google.com/d/msgid/elasticsearch/205b2250-f6ee-4141-a49a-ff22aade0fe9%40googlegroups.com?utm_medium=email&utm_source=footer> >>>> . >>>> >>>> For more options, visit https://groups.google.com/d/optout. >>>> >>> >>> -- >> You received this message because you are subscribed to the Google Groups >> "elasticsearch" group. >> To unsubscribe from this group and stop receiving emails from it, send an >> email to elasticsearc...@googlegroups.com <javascript:>. >> To view this discussion on the web visit >> https://groups.google.com/d/msgid/elasticsearch/45c53e81-2c8a-40cd-aadd-15891a09e30d%40googlegroups.com >> >> <https://groups.google.com/d/msgid/elasticsearch/45c53e81-2c8a-40cd-aadd-15891a09e30d%40googlegroups.com?utm_medium=email&utm_source=footer> >> . >> >> For more options, visit https://groups.google.com/d/optout. >> > >
-- You received this message because you are subscribed to the Google Groups "elasticsearch" group. To unsubscribe from this group and stop receiving emails from it, send an email to elasticsearch+unsubscr...@googlegroups.com. To view this discussion on the web visit https://groups.google.com/d/msgid/elasticsearch/cd24c022-1b65-4857-98da-a90415164663%40googlegroups.com. For more options, visit https://groups.google.com/d/optout.