I'm seeing 9X throughput with 1000 docs/batch vs 1 doc/batch, with a single thread, so it's certainly worth it.
Thanks, Peter On Fri, Nov 7, 2014 at 2:18 PM, Erick Erickson <erickerick...@gmail.com> wrote: > And Walter has also been around for a _long_ time ;) > > (sorry, couldn't resist).... > > Erick > > On Fri, Nov 7, 2014 at 11:12 AM, Walter Underwood <wun...@wunderwood.org> > wrote: > > Yes, I implemented exactly that fallback for Solr 1.2 at Netflix. > > > > It isn’t to hard if the code is structured for it; retry with a batch > size of 1. > > > > wunder > > > > On Nov 7, 2014, at 11:01 AM, Erick Erickson <erickerick...@gmail.com> > wrote: > > > >> Yeah, this has been an ongoing issue for a _long_ time. Basically, > >> you can't. So far, people have essentially written fallback logic to > >> index the docs of a failing packet one at a time and report it. > >> > >> I'd really like better reporting back, but we haven't gotten there yet. > >> > >> Best, > >> Erick > >> > >> On Fri, Nov 7, 2014 at 8:25 AM, Peter Keegan <peterlkee...@gmail.com> > wrote: > >>> How are folks handling Solr exceptions that occur during batch > indexing? > >>> Solr stops parsing the docs stream when an error occurs (e.g. a doc > with a > >>> missing mandatory field), and stops indexing the batch. The bad > document is > >>> not identified, so it would be hard for the client to recover by > skipping > >>> over it. > >>> > >>> Peter > > >