Ok thanks. I'll wait for the patch.
Steve On 5 March 2015 at 12:28, Zeeshan Lakhani <zlakh...@basho.com> wrote: > Hey Steve, > > Your data will remain in Riak, but Search will then use AAE to rebuild the > data on the Solr side. This is only to fix Solr’s core data issues. If > you’re worried about an interruption on the SOLR side, as there would be > one to reindex the data, you could then wait until things are sured-up in > our next release/patch-release to help solve these issues. > > As I mentioned, we’re working on a fix for this with an internal patch > that we’re testing at this moment. > > Thanks. > > Zeeshan Lakhani > programmer | > software engineer at @basho | > org. member/founder of @papers_we_love | paperswelove.org > twitter => @zeeshanlakhani > > On Mar 5, 2015, at 12:13 PM, Steve Garon <steve.ga...@gmail.com> wrote: > > Does deleting core.properties or moving search-root/index deletes any of > my index data? Will that cause any interruption to SOLR? Cause I'm having > problems on a production cluster and I really don't want to have to backup > 300 millions keys again or cause any interruptions to the system ... > > > Steve > > On 5 March 2015 at 10:11, Zeeshan Lakhani <zlakh...@basho.com> wrote: > >> Hey Steve, >> >> We’re currently tracking this issue, >> https://github.com/basho/yokozuna/issues/442, and are working on testing >> out a patch internally. I will update you as soon as we get clarity there. >> >> As a possible workaround, I would attempt to delete the `core.properties` >> file located in the search root_directory/index of each node >> (./data/yz/<<index_name>>). This file should then be recreated on the next >> attempt to creating the core on the Solr side, >> https://github.com/basho/yokozuna/blob/92ca14cc35b46c8e7ac86cad6d92547e68e8d917/src/yz_index.erl#L171 >> . >> >> If that doesn’t work, you can then delete or `mv` that search-root/index >> directory out of the way, for which it will then get recreated and AAE will >> sync the data. >> >> Thanks. >> >> Zeeshan Lakhani >> programmer | >> software engineer at @basho | >> org. member/founder of @papers_we_love | paperswelove.org >> twitter => @zeeshanlakhani >> >> On Mar 5, 2015, at 9:02 AM, Steve Garon <steve.ga...@gmail.com> wrote: >> >> In solr.log, the kind of exceptions that I'm getting right now is the >> following: >> >> 1. IO Error while trying to get the size of the >> Directory:java.io.FileNotFoundException: SOME_RAMDOM_FILE* (.doc, .pos, >> .fnm, .si, .nv, .gen extensions)* >> 2. SolrException.java:120 null:org.apache.sorl.common.SolrException: Core >> with name 'BUCKET NAME' already exists. >> 3. SolrException.java:120 null:org.eclipse.jetty.io.EofException >> 4. Server refused connection >> 5. IOException occured when talking to server >> >> >> >> >> Steve >> >> On 5 March 2015 at 08:50, Steve Garon <steve.ga...@gmail.com> wrote: >> >>> Yes I do have yz_events,handle_info in crash log. Tons of them actually >>> and they have a big ass strack trace attached to each of them. >>> >>> It would be hard for me to provide you with logs. If you have specific >>> questions you want answers to I'd be happy to help though. >>> >>> >>> Steve >>> >>> On 4 March 2015 at 14:20, Zeeshan Lakhani <zlakh...@basho.com> wrote: >>> >>>> Hey Steve, >>>> >>>> Sorry to see you’re having new issues. >>>> >>>> We’ll have the fix for “space in the key” out soon; it’s currently >>>> under review. And, I know that this issue is unrelated. >>>> >>>> I have a few different routes/thoughts for you, but are you seeing >>>> anything related to `yz_events,handle_info` in your crash logs? Also, can >>>> you gist/pastebin me your solr logs? I’d like to seem if it correlates with >>>> something we’re currently looking at. >>>> >>>> Thanks. >>>> >>>> Zeeshan Lakhani >>>> programmer | >>>> software engineer at @basho | >>>> org. member/founder of @papers_we_love | paperswelove.org >>>> twitter => @zeeshanlakhani >>>> >>>> On Mar 4, 2015, at 10:39 AM, Steve Garon <steve.ga...@gmail.com> wrote: >>>> >>>> Hey all, >>>> >>>> We were having the "space in the key" bug in our cluster so we went >>>> through the whole dataset, backing it up to json file and removing the >>>> spaces in the keys. Then we trashed our whole cluster and restart from >>>> scratch reimporting the whole data. Everything worked like a charm for two >>>> weeks but this weekend, not sure what happened but AAE died again. >>>> >>>> I have two issues now: >>>> 1. AAE is trying to recreate an index that already exists and crashes >>>> with an "Already exists" error ... I get this in my error log every 15 >>>> seconds >>>> 2. AAE crashes while iterating through entropy data with a request >>>> timeout error every hour followed by tons of failed to index objects with >>>> request timeout as well. Stack trace looks like this: >>>> >>>> *[error] emulator Error in process <025931.7166> on node 'riak@IP' with >>>> exit value: {function_clause,[{yz_entropy,iterate_entropy_data,[<<11 >>>> bytes>>,[{continuation,<<159 >>>> bytes>>},{limit,100},{partition,12}],#Fun<yz_index_hashtree.5.46188917,{error,{error,req_timeout}}],[{file,"src/yz_entropy.erl"},{line,44}]},{yz_index_hashtree,'-fold_keys/2-lc$^0/1-0-',3,[{file... >>>> **(TRUNCATED)* >>>> >>>> *[error] <0.1371.0>@yz_kv:index:215 failed to index object >>>> {{<<"TYPE">>,<<"BUCKET">>},<<"KEY">>} with error {"Failed to index >>>> docs",{error,req_timeout}} because ... **(REPEATED MULTIPLE TIME FOR >>>> DIFFERENT KEYS)* >>>> >>>> >>>> I tried clearing the yz anti-entropy tree and reinitialising the >>>> yz_entropy_mgr with no luck. Anything I can do to fix this? >>>> >>>> Oh FYI, I cannot insert data with spaces in the key anymore because we >>>> are using a wrapper on top of riak that will prevent us to do so therefore >>>> my issues are not related to this for sure. >>>> >>>> These are some config changes that may be good to know for more context. >>>> >>>> We added this to ibrowse.conf: >>>> {dest, "localhost", 8093, 100, 1000, []}. >>>> >>>> Jetty is set with minthread 80, acceptors 80, and is using the NIO >>>> connector. >>>> >>>> All our solr buckets have filterCache disabled with softcommits set to >>>> 10s instead of 1. >>>> >>>> Our riak.conf has background_manager turned on with AAE and handoff >>>> using it. >>>> >>>> >>>> Thanks, >>>> >>>> Steve >>>> _______________________________________________ >>>> riak-users mailing list >>>> riak-users@lists.basho.com >>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >>>> >>>> >>>> >>> >> >> > >
_______________________________________________ riak-users mailing list riak-users@lists.basho.com http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com