Have you tried modifying the timeout parameters? See: http://wiki.apache.org/solr/SolrReplication, the "Slave" section..
Best Erick On Mon, May 14, 2012 at 10:30 AM, Rohit <ro...@in-rev.com> wrote: > The size of index is about 300GB, I am seeing the following error in the > logs, > > java.net.SocketTimeoutException: Read timed out > at java.net.SocketInputStream.socketRead0(Native Method) > at java.net.SocketInputStream.read(SocketInputStream.java:129) > at java.io.BufferedInputStream.fill(BufferedInputStream.java:218) > at java.io.BufferedInputStream.read(BufferedInputStream.java:237) > at > org.apache.commons.httpclient.ChunkedInputStream.getChunkSizeFromInputStream > (ChunkedInputStream.java:250) > at > org.apache.commons.httpclient.ChunkedInputStream.nextChunk(ChunkedInputStrea > m.java:221) > at > org.apache.commons.httpclient.ChunkedInputStream.read(ChunkedInputStream.jav > a:176) > at java.io.FilterInputStream.read(FilterInputStream.java:116) > at > org.apache.commons.httpclient.AutoCloseInputStream.read(AutoCloseInputStream > .java:108) > at > org.apache.solr.common.util.FastInputStream.refill(FastInputStream.java:68) > at > org.apache.solr.common.util.FastInputStream.read(FastInputStream.java:97) > at > org.apache.solr.common.util.FastInputStream.readFully(FastInputStream.java:1 > 22) > at > org.apache.solr.common.util.FastInputStream.readFully(FastInputStream.java:1 > 17) > at > org.apache.solr.handler.SnapPuller$FileFetcher.fetchPackets(SnapPuller.java: > 943) > at > org.apache.solr.handler.SnapPuller$FileFetcher.fetchFile(SnapPuller.java:904 > ) > at > org.apache.solr.handler.SnapPuller.downloadIndexFiles(SnapPuller.java:545) > at > org.apache.solr.handler.SnapPuller.fetchLatestIndex(SnapPuller.java:295) > at > org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:2 > 68) > at > org.apache.solr.handler.ReplicationHandler$1.run(ReplicationHandler.java:149 > ) > May 14, 2012 1:45:46 PM org.apache.solr.handler.ReplicationHandler doFetch > SEVERE: SnapPull failed > org.apache.solr.common.SolrException: Unable to download _vvyv.fdt > completely. Downloaded 200278016!=208644265 > at > org.apache.solr.handler.SnapPuller$FileFetcher.cleanup(SnapPuller.java:1038) > at > org.apache.solr.handler.SnapPuller$FileFetcher.fetchFile(SnapPuller.java:918 > ) > at > org.apache.solr.handler.SnapPuller.downloadIndexFiles(SnapPuller.java:545) > at > org.apache.solr.handler.SnapPuller.fetchLatestIndex(SnapPuller.java:295) > at > org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:2 > 68) > at > org.apache.solr.handler.ReplicationHandler$1.run(ReplicationHandler.java:149 > ) > > > Actually the replication starts, but is never able to complete and then > restarts again. > > Regards, > Rohit > Mobile: +91-9901768202 > About Me: http://about.me/rohitg > > > -----Original Message----- > From: Erick Erickson [mailto:erickerick...@gmail.com] > Sent: 14 May 2012 18:00 > To: solr-user@lucene.apache.org > Subject: Re: Relicating a large solr index > > What do your logs show? Solr replication should be robust. > How large is "large"? > > You might review: > http://wiki.apache.org/solr/UsingMailingLists > > Best > Erick > > On Mon, May 14, 2012 at 3:11 AM, Rohit <ro...@in-rev.com> wrote: >> Hi, >> >> >> >> I have a large solr index which needs to be replicated, solr >> replication start but then keeps breaking and starting from 0. Is >> there another way to achieve this, I was thinking of using >> scp to copy the index from master to slave and then enable replication, > will this work? >> >> >> >> >> Regards, >> >> Rohit >> >> >> > >