Hi Bernd,

Yet, another noob question.

Consider that my conf directory for creating a collection is _default. Suppose
now I made changes to managed-schema and conf.xml, How do I upload it to
external zookeeper at 2181 port?

Can you please give me the command that uploads altered config.xml and
managed-schema to zookeeper?

Thanks.


On Fri, Aug 2, 2019 at 11:53 AM Bernd Fehling <
bernd.fehl...@uni-bielefeld.de> wrote:

>
> to 1) yes, because -Djute.maxbuffer is going to JAVA as a start parameter.
>
> to 2) I don't know because i never use internal zookeeper
>
> to 3) the configs are located at solr/server/solr/configsets/
>        - choose one configset, make your changes and upload it to zookeeper
>        - when creating a new collection choose your uploaded config
>        - whenever you change something at your config you have to upload
> it to zookeeper
>
> I don't know which Solr version you are using, but a good starting point
> with solr cloud is
> http://lucene.apache.org/solr/guide/6_6/solrcloud.html
>
> Regards
> Bernd
>
>
>
> Am 02.08.19 um 07:59 schrieb Salmaan Rashid Syed:
> > Hi Bernd,
> >
> > Sorry for noob questions.
> >
> > 1) What do you mean by restart? Do you mean that I shoud issue ./bin/solr
> > stop -all?
> >
> > And then issue these commands,
> >
> > bin/solr restart -cloud -s example/cloud/node1/solr -p 8983
> >
> > bin/solr restart -c -p 7574 -z localhost:9983 -s example/cloud/node2/solr
> >
> >
> > 2) Where can I find solr internal Zookeeper folder for issuing this
> command
> > SERVER_JVMFLAGS="$SERVER_JVMFLAGS -Djute.maxbuffer=10000000"?
> >
> >
> > 3) Where can I find schema.xml and config.xmo files for Solr Cloud Cores
> to
> > make changes in schema and configuration? Or do I have to make chages in
> > the directory that contains managed-schema and config.xml files with
> which
> > I initialized and created collections? And then the solr will pick them
> up
> > from there when it restarts?
> >
> >
> > Regards,
> >
> > Salmaan
> >
> >
> >
> > On Thu, Aug 1, 2019 at 5:40 PM Bernd Fehling <
> bernd.fehl...@uni-bielefeld.de>
> > wrote:
> >
> >>
> >>
> >> Am 01.08.19 um 13:57 schrieb Salmaan Rashid Syed:
> >>> After I make the -Djute.maxbuffer changes to Solr, deployed in
> >> production,
> >>> Do I need to restart the solr to be able to add synonyms >1MB?
> >>
> >> Yes, you have to restart Solr.
> >>
> >>
> >>>
> >>> Or, Was this supposed to be done before putting Solr to production
> ever?
> >>> Can we make chages when the Solr is running in production?
> >>
> >> It depends on your system. In my cloud with 5 shards and 3 replicas I
> can
> >> take one by one offline, stop, modify and start again without problems.
> >>
> >>
> >>>
> >>> Thanks.
> >>>
> >>> Regards,
> >>> Salmaan
> >>>
> >>>
> >>>
> >>> On Tue, Jul 30, 2019 at 4:53 PM Bernd Fehling <
> >>> bernd.fehl...@uni-bielefeld.de> wrote:
> >>>
> >>>> You have to increase the -Djute.maxbuffer for large configs.
> >>>>
> >>>> In Solr bin/solr/solr.in.sh use e.g.
> >>>> SOLR_OPTS="$SOLR_OPTS -Djute.maxbuffer=10000000"
> >>>> This will increase maxbuffer for zookeeper on solr side to 10MB.
> >>>>
> >>>> In Zookeeper zookeeper/conf/zookeeper-env.sh
> >>>> SERVER_JVMFLAGS="$SERVER_JVMFLAGS -Djute.maxbuffer=10000000"
> >>>>
> >>>> I have a >10MB Thesaurus and use 30MB for jute.maxbuffer, works
> perfect.
> >>>>
> >>>> Regards
> >>>>
> >>>>
> >>>> Am 30.07.19 um 13:09 schrieb Salmaan Rashid Syed:
> >>>>> Hi Solr Users,
> >>>>>
> >>>>> I have a very big synonym file (>5MB). I am unable to start Solr in
> >> cloud
> >>>>> mode as it throws an error message stating that the synonmys file is
> >>>>> too large. I figured out that the zookeeper doesn't take a file
> greater
> >>>>> than 1MB size.
> >>>>>
> >>>>> I tried to break down my synonyms file to smaller chunks less than
> 1MB
> >>>>> each. But, I am not sure about how to include all the filenames into
> >> the
> >>>>> Solr schema.
> >>>>>
> >>>>> Should it be seperated by commas like synonyms = "__1_synonyms.txt,
> >>>>> __2_synonyms.txt, __3synonyms.txt"
> >>>>>
> >>>>> Or is there a better way of doing that? Will the bigger file when
> >> broken
> >>>>> down to smaller chunks will be uploaded to zookeeper as well.
> >>>>>
> >>>>> Please help or please guide me to relevant documentation regarding
> >> this.
> >>>>>
> >>>>> Thank you.
> >>>>>
> >>>>> Regards.
> >>>>> Salmaan.
> >>>>>
> >>>>
> >>>
> >>
> >
>

Reply via email to