I'll have to check on the commit situation. We have been pushing data from
SharePoint the last week or so. Would that somehow block the documents
moving between the solr instances?

I'll try another version tomorrow. Thanks for the suggestions.

On Mon, Feb 27, 2012 at 5:34 PM, Mark Miller <markrmil...@gmail.com> wrote:

> Hmmm...all of that looks pretty normal...
>
> Did a commit somehow fail on the other machine? When you view the stats
> for the update handler, are there a lot of pending adds for on of the
> nodes? Do the commit counts match across nodes?
>
> You can also query an individual node with distrib=false to check that.
>
> If you build is a month old, I'd honestly recommend you try upgrading as
> well.
>
> - Mark
>
> On Feb 27, 2012, at 3:34 PM, Matthew Parker wrote:
>
> > Here is most of the cluster state:
> >
> > Connected to Zookeeper
> > localhost:2181, localhost: 2182, localhost:2183
> >
> > /(v=0 children=7) ""
> >   /CONFIGS(v=0, children=1)
> >      /CONFIGURATION(v=0 children=25)
> >             <<<<< all the configuration files, velocity info, xslt, etc.
> >>>>>
> >  /NODE_STATES(v=0 children=4)
> >     MACHINE1:8083_SOLR (v=121)"[{"shard_id":"shard1",
> > "state":"active","core":"","collection":"collection1","node_name:"..."
> >     MACHINE1:8082_SOLR (v=101)"[{"shard_id":"shard2",
> > "state":"active","core":"","collection":"collection1","node_name:"..."
> >     MACHINE1:8081_SOLR (v=92)"[{"shard_id":"shard1",
> > "state":"active","core":"","collection":"collection1","node_name:"..."
> >     MACHINE1:8084_SOLR (v=73)"[{"shard_id":"shard2",
> > "state":"active","core":"","collection":"collection1","node_name:"..."
> >  /ZOOKEEPER (v-0 children=1)
> >     QUOTA(v=0)
> >
> >
> /CLUSTERSTATE.JSON(V=272)"{"collection1":{"shard1":{MACHINE1:8081_solr_":{shard_id":"shard1","leader":"true","..."
> >  /LIVE_NODES (v=0 children=4)
> >     MACHINE1:8083_SOLR(ephemeral v=0)
> >     MACHINE1:8082_SOLR(ephemeral v=0)
> >     MACHINE1:8081_SOLR(ephemeral v=0)
> >     MACHINE1:8084_SOLR(ephemeral v=0)
> >  /COLLECTIONS (v=1 children=1)
> >     COLLECTION1(v=0 children=2)"{"configName":"configuration1"}"
> >         LEADER_ELECT(v=0 children=2)
> >             SHARD1(V=0 children=1)
> >                 ELECTION(v=0 children=2)
> >
> > 87186203314552835-MACHINE1:8081_SOLR_-N_0000000096(ephemeral v=0)
> >
> > 87186203314552836-MACHINE1:8083_SOLR_-N_0000000084(ephemeral v=0)
> >             SHARD2(v=0 children=1)
> >                 ELECTION(v=0 children=2)
> >
> > 231301391392833539-MACHINE1:8084_SOLR_-N_0000000085(ephemeral v=0)
> >
> > 159243797356740611-MACHINE1:8082_SOLR_-N_0000000084(ephemeral v=0)
> >         LEADERS (v=0 children=2)
> >             SHARD1 (ephemeral
> > v=0)"{"core":"","node_name":"MACHINE1:8081_solr","base_url":"
> > http://MACHINE1:8081/solr"}";
> >             SHARD2 (ephemeral
> > v=0)"{"core":"","node_name":"MACHINE1:8082_solr","base_url":"
> > http://MACHINE1:8082/solr"}";
> >  /OVERSEER_ELECT (v=0 children=2)
> >     ELECTION (v=0 children=4)
> >         231301391392833539-MACHINE1:8084_SOLR_-N_0000000251(ephemeral
> v=0)
> >         87186203314552835-MACHINE1:8081_SOLR_-N_0000000248(ephemeral v=0)
> >         159243797356740611-MACHINE1:8082_SOLR_-N_0000000250(ephemeral
> v=0)
> >         87186203314552836-MACHINE1:8083_SOLR_-N_0000000249(ephemeral v=0)
> >     LEADER (emphemeral
> > v=0)"{"id":"87186203314552835-MACHINE1:8081_solr-n_000000248"}"
> >
> >
> >
> > On Mon, Feb 27, 2012 at 2:47 PM, Mark Miller <markrmil...@gmail.com>
> wrote:
> >
> >>
> >> On Feb 27, 2012, at 2:22 PM, Matthew Parker wrote:
> >>
> >>> Thanks for your reply Mark.
> >>>
> >>> I believe the build was towards the begining of the month. The
> >>> solr.spec.version is 4.0.0.2012.01.10.38.09
> >>>
> >>> I cannot access the clusterstate.json contents. I clicked on it a
> couple
> >> of
> >>> times, but nothing happens. Is that stored on disk somewhere?
> >>
> >> Are you using the new admin UI? That has recently been updated to work
> >> better with cloud - it had some troubles not too long ago. If you are,
> you
> >> should trying using the old admin UI's zookeeper page - that should show
> >> the cluster state.
> >>
> >> That being said, there has been a lot of bug fixes over the past month -
> >> so you may just want to update to a recent version.
> >>
> >>>
> >>> I configured a custom request handler to calculate an unique document
> id
> >>> based on the file's url.
> >>>
> >>> On Mon, Feb 27, 2012 at 1:13 PM, Mark Miller <markrmil...@gmail.com>
> >> wrote:
> >>>
> >>>> Hey Matt - is your build recent?
> >>>>
> >>>> Can you visit the cloud/zookeeper page in the admin and send the
> >> contents
> >>>> of the clusterstate.json node?
> >>>>
> >>>> Are you using a custom index chain or anything out of the ordinary?
> >>>>
> >>>>
> >>>> - Mark
> >>>>
> >>>> On Feb 27, 2012, at 12:26 PM, Matthew Parker wrote:
> >>>>
> >>>>> TWIMC:
> >>>>>
> >>>>> Environment
> >>>>> =========
> >>>>> Apache SOLR rev-1236154
> >>>>> Apache Zookeeper 3.3.4
> >>>>> Windows 7
> >>>>> JDK 1.6.0_23.b05
> >>>>>
> >>>>> I have built a SOLR Cloud instance with 4 nodes using the embeded
> Jetty
> >>>>> servers.
> >>>>>
> >>>>> I created a 3 node zookeeper ensemble to manage the solr
> configuration
> >>>> data.
> >>>>>
> >>>>> All the instances run on one server so I've had to move ports around
> >> for
> >>>>> the various applications.
> >>>>>
> >>>>> I start the 3 zookeeper nodes.
> >>>>>
> >>>>> I started the first instance of solr cloud with the parameter to have
> >> two
> >>>>> shards.
> >>>>>
> >>>>> The start the remaining 3 solr nodes.
> >>>>>
> >>>>> The system comes up fine. No errors thrown.
> >>>>>
> >>>>> I can view the solr cloud console and I can see the SOLR
> configuration
> >>>>> files managed by ZooKeeper.
> >>>>>
> >>>>> I published data into the SOLR Cloud instances from SharePoint using
> >>>> Apache
> >>>>> Manifold 0.4-incubating. Manifold is setup to publish the data into
> >>>>> collection1, which is the only collection defined in the cluster.
> >>>>>
> >>>>> When I query the data from collection1 as per the solr wiki, the
> >> results
> >>>>> are inconsistent. Sometimes all the results are there, other times
> >>>> nothing
> >>>>> comes back at all.
> >>>>>
> >>>>> It seems to be having an issue auto replicating the data across the
> >>>> cloud.
> >>>>>
> >>>>> Is there some specific setting I might have missed? Based upon what I
> >>>> read,
> >>>>> I thought that SOLR cloud would take care of distributing and
> >> replicating
> >>>>> the data automatically. Do you have to tell it what shard to publish
> >> the
> >>>>> data into as well?
> >>>>>
> >>>>> Any help would be appreciated.
> >>>>>
> >>>>> Thanks,
> >>>>>
> >>>>> Matt
> >>>>>
> >>>>> ------------------------------
> >>>>> This e-mail and any files transmitted with it may be proprietary.
> >>>> Please note that any views or opinions presented in this e-mail are
> >> solely
> >>>> those of the author and do not necessarily represent those of Apogee
> >>>> Integration.
> >>>>
> >>>> - Mark Miller
> >>>> lucidimagination.com
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>
> >>> Matt
> >>>
> >>> ------------------------------
> >>> This e-mail and any files transmitted with it may be proprietary.
> >> Please note that any views or opinions presented in this e-mail are
> solely
> >> those of the author and do not necessarily represent those of Apogee
> >> Integration.
> >>
> >> - Mark Miller
> >> lucidimagination.com
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >
> > ------------------------------
> > This e-mail and any files transmitted with it may be proprietary.
>  Please note that any views or opinions presented in this e-mail are solely
> those of the author and do not necessarily represent those of Apogee
> Integration.
>
> - Mark Miller
> lucidimagination.com
>
>
>
>
>
>
>
>
>
>
>
>

------------------------------
This e-mail and any files transmitted with it may be proprietary.  Please note 
that any views or opinions presented in this e-mail are solely those of the 
author and do not necessarily represent those of Apogee Integration.

Reply via email to