So, basically, your proposed changes, broken into tangible gobbets of work:

   - Expand ambari to manage the remaining sensor-specific configs
   - Refactor the push calls to zookeeper (in ConfigurationUtils, I think)
   to push to ambari and take a reason
      - Question remains about whether ambari can do the push to zookeeper
      or whether ConfigurationUtils has to push to zookeeper as well as update
      ambari.
   - Refactor the middleware that Ryan submitted to have the API calls take
   a reason
   - Refactor the management UI to pass in a reason
   - Refactor the Stellar Management functions CONFIG_PUT to accept a reason

Just so we can evaluate it and I can ensure I haven't overlooked some
important point.  Please tell me if Ambari cannot do the things we're
suggesting it can do.

Casey

On Fri, Jan 13, 2017 at 9:15 AM, David Lyle <dlyle65...@gmail.com> wrote:

> That's exactly correct, Casey. Basically, an expansion of what we're
> currently doing with global.json, enrichment.properties and
> elasticsearch.properties.
>
> -D...
>
>
> On Fri, Jan 13, 2017 at 9:12 AM, Casey Stella <ceste...@gmail.com> wrote:
>
> > I would suggest not having Ambari replace zookeeper.  I think the
> proposal
> > is to have Ambari replace the editable store (like the JSON files on
> > disk).  Zookeeper woudl be the source of truth for the running topologies
> > and ambari would be sync'd to it.
> >
> > Correct if I misspeak, dave or matt.
> >
> > Casey
> >
> > On Fri, Jan 13, 2017 at 9:09 AM, Nick Allen <n...@nickallen.org> wrote:
> >
> > > Ambari seems like a logical choice.
> > >
> > > *>> It doesn’t natively integrate Zookeeper storage of configs, but
> there
> > > is a natural place to specify copy to/from Zookeeper for the files
> > > desired.*
> > >
> > > How would Ambari interact with Zookeeper in this scenario?  Would
> Ambari
> > > replace Zookeeper completely? Or would Zookeeper act as the persistence
> > > tier under Ambari?
> > >
> > >
> > >
> > >
> > > On Thu, Jan 12, 2017 at 9:24 PM, Matt Foley <ma...@apache.org> wrote:
> > >
> > > > Mike, could you try again on the image, please, making sure it is a
> > > simple
> > > > format (gif, png, or jpeg)?  It got munched, at least in my viewer.
> > > Thanks.
> > > >
> > > > Casey, responding to some of the questions you raised:
> > > >
> > > > I’m going to make a rather strong statement:  We already have a
> service
> > > > “to intermediate and handle config update/retrieval”.
> > > > Furthermore, it:
> > > > - Correctly handles the problems of distributed services running on
> > > > multi-node clusters.  (That’s a HARD problem, people, and we
> shouldn’t
> > > try
> > > > to reinvent the wheel.)
> > > > - Correctly handles Kerberos security. (That’s kinda hard too, or at
> > > least
> > > > a lot of work.)
> > > > - It does automatic versioning of configurations, and allows viewing,
> > > > comparing, and reverting historical configs
> > > > - It has a capable REST API for all those things.
> > > > It doesn’t natively integrate Zookeeper storage of configs, but there
> > is
> > > a
> > > > natural place to specify copy to/from Zookeeper for the files
> desired.
> > > >
> > > > It is Ambari.  And we should commit to it, rather than try to
> re-create
> > > > such features.
> > > > Because it has a good REST API, it is perfectly feasible to implement
> > > > Stellar functions that call it.
> > > > GUI configuration tools can also use the Ambari APIs, or better yet
> be
> > > > integrated in an “Ambari View”. (Eg, see the “Yarn Capacity Scheduler
> > > > Configuration Tool” example in the Ambari documentation, under “Using
> > > > Ambari Views”.)
> > > >
> > > > Arguments are: Parsimony, Sufficiency, Not reinventing the wheel, and
> > Not
> > > > spending weeks and weeks of developer time over the next year
> > reinventing
> > > > the wheel while getting details wrong multiple times…
> > > >
> > > > Okay, off soapbox.
> > > >
> > > > Casey asked what the config update behavior of Ambari is, and how it
> > will
> > > > interact with changes made from outside Ambari.
> > > > The following is from my experience working with the Ambari Mpack for
> > > > Metron.  I am not otherwise an Ambari expert, so tomorrow I’ll get it
> > > > reviewed by an Ambari development engineer.
> > > >
> > > > Ambari-server runs on one node, and Ambari-agent runs on each of all
> > the
> > > > nodes.
> > > > Ambari-server has a private set of py, xml, and template files, which
> > > > together are used both to generate the Ambari configuration GUI, with
> > > > defaults, and to generate configuration files (of any needed
> filetype)
> > > for
> > > > the various Stack components.
> > > > Ambari-server also has a database where it stores the schema related
> to
> > > > these files, so even if you reach in and edit Ambari’s files, it will
> > > Error
> > > > out if the set of parameters or parameter names changes.  The
> > historical
> > > > information about configuration changes is also stored in the db.
> > > > For each component (and in the case of Metron, for each topology),
> > there
> > > > is a python file which controls the logic for these actions, among
> > > others:
> > > > - Install
> > > > - Start / stop / restart / status
> > > > - Configure
> > > >
> > > > It is actually up to this python code (which we wrote for the Metron
> > > > Mpack) what happens in each of these API calls.  But the current
> code,
> > > and
> > > > I believe this is typical of Ambari-managed components, performs a
> > > > “Configure” action whenever you press the “Save” button after
> changing
> > a
> > > > component config in Ambari, and also on each Install and Start or
> > > Restart.
> > > >
> > > > The Configure action consists of approximately the following sequence
> > > (see
> > > > disclaimer above :-)
> > > > - Recreate the generated config files, using the template files and
> the
> > > > actual configuration most recently set in Ambari
> > > > o Note this is also under the control of python code that we wrote,
> and
> > > > this is the appropriate place to push to ZK if desired.
> > > > - Propagate those config files to each Ambari-agent, with a command
> to
> > > set
> > > > them locally
> > > > - The ambari-agents on each node receive the files and write them to
> > the
> > > > specified locations on local storage
> > > >
> > > > Ambari-server then whines that the updated services should be
> > restarted,
> > > > but does not initiate that action itself (unless of course the
> > initiating
> > > > action was a Start command from the administrator).
> > > >
> > > > Make sense?  It’s all quite straightforward in concept, there’s just
> an
> > > > awful lot of stuff wrapped around that to make it all go smoothly and
> > > > handle the problems when it doesn’t.
> > > >
> > > > There’s additional complexity in that the Ambari-agent also caches
> (on
> > > > each node) both the template files and COMPILED forms of the python
> > files
> > > > (.pyc) involved in transforming them.  The pyc files incorporate some
> > > > amount of additional info regarding parameter values, but I’m not
> sure
> > of
> > > > the form.  I don’t think that changes the above in any practical way
> > > unless
> > > > you’re trying to cheat Ambari by reaching in and editing its files
> > > > directly.  In that case, you also need to whack the pyc files (on
> each
> > > > node) to force the data to be reloaded from Ambari-server.  Best
> > solution
> > > > is don’t cheat.
> > > >
> > > > Also, there may be circumstances under which the Ambari-agent will
> > detect
> > > > changes and re-write the latest version it knows of the config files,
> > > even
> > > > without a Save or Start action at the Ambari-server.  I’m not sure of
> > > this
> > > > and need to check with Ambari developers.  It may no longer happen,
> > altho
> > > > I’m pretty sure change detection/reversion was a feature of early
> > > versions
> > > > of Ambari.
> > > >
> > > > Hope this helps,
> > > > --Matt
> > > >
> > > > ================================================
> > > > From: Michael Miklavcic <michael.miklav...@gmail.com>
> > > > Reply-To: "dev@metron.incubator.apache.org"
> > > <dev@metron.incubator.apache.
> > > > org>
> > > > Date: Thursday, January 12, 2017 at 3:59 PM
> > > > To: "dev@metron.incubator.apache.org" <dev@metron.incubator.apache.
> org
> > >
> > > > Subject: Re: [DISCUSS] Ambari Metron Configuration Management
> > > consequences
> > > > and call to action
> > > >
> > > > Hi Casey,
> > > >
> > > > Thanks for starting this thread. I believe you are correct in your
> > > > assessment of the 4 options for updating configs in Metron. When
> using
> > > more
> > > > than one of these options we can get into a split-brain scenario. A
> > basic
> > > > example is updating the global config on disk and using the
> > > > zk_load_configs.sh. Later, if a user decides to restart Ambari, the
> > > cached
> > > > version stored by Ambari (it's in the MySQL or other database backing
> > > > Ambari) will be written out to disk in the defined config directory,
> > and
> > > > subsequently loaded using the zk_load_configs.sh under the hood. Any
> > > global
> > > > configuration modified outside of Ambari will be lost at this point.
> > This
> > > > is obviously undesirable, but I also like the purpose and utility
> > exposed
> > > > by the multiple config management interfaces we currently have
> > > available. I
> > > > also agree that a service would be best.
> > > >
> > > > For reference, here's my understanding of the current configuration
> > > > loading mechanisms and their deps.
> > > >
> > > > <image>
> > > >
> > > > Mike
> > > >
> > > >
> > > > On Thu, Jan 12, 2017 at 3:08 PM, Casey Stella <ceste...@gmail.com>
> > > wrote:
> > > >
> > > > In the course of discussion on the PR for METRON-652
> > > > <https://github.com/apache/incubator-metron/pull/415> something
> that I
> > > > should definitely have understood better came to light and I thought
> > that
> > > > it was worth bringing to the attention of the community to get
> > > > clarification/discuss is just how we manage configs.
> > > >
> > > > Currently (assuming the management UI that Ryan Merriman submitted)
> > > configs
> > > > are managed/adjusted via a couple of different mechanism.
> > > >
> > > >    - zk_load_utils.sh: pushed and pulled from disk to zookeeper
> > > >    - Stellar REPL: pushed and pulled via the CONFIG_GET/CONFIG_PUT
> > > > functions
> > > >    - Ambari: initialized via the zk_load_utils script and then some
> of
> > > them
> > > >    are managed directly (global config) and some indirectly
> > > > (sensor-specific
> > > >    configs).
> > > >       - NOTE: Upon service restart, it may or may not overwrite
> changes
> > > on
> > > >       disk or on zookeeper.  *Can someone more knowledgeable than me
> > > about
> > > >       this describe precisely the semantics that we can expect on
> > > > service restart
> > > >       for Ambari? What gets overwritten on disk and what gets updated
> > > > in ambari?*
> > > >    - The Management UI: manages some of the configs. *RYAN: Which
> > configs
> > > >    do we support here and which don't we support here?*
> > > >
> > > > As you can see, we have a mishmash of mechanisms to update and manage
> > the
> > > > configuration for Metron in zookeeper.  In the beginning the approach
> > was
> > > > just to edit configs on disk and push/pull them via zk_load_utils.
> > > Configs
> > > > could be historically managed using source control, etc.  As we got
> > more
> > > > and more components managing the configs, we haven't taken care that
> > they
> > > > they all work with each other in an expected way (I believe these are
> > > > true..correct me if I'm wrong):
> > > >
> > > >    - If configs are modified in the management UI or the Stellar REPL
> > and
> > > >    someone forgets to pull the configs from zookeeper to disk, before
> > > they
> > > > do
> > > >    a push via zk_load_utils, they will clobber the configs in
> zookeeper
> > > > with
> > > >    old configs.
> > > >    - If the global config is changed on disk and the ambari service
> > > >    restarts, it'll get reset with the original global config.
> > > >    - *Ryan, in the management UI, if someone changes the zookeeper
> > > configs
> > > >    from outside, are those configs reflected immediately in the UI?*
> > > >
> > > >
> > > > It seems to me that we have a couple of options here:
> > > >
> > > >    - A service to intermediate and handle config update/retrieval and
> > > >    tracking historical changes so these different mechanisms can use
> a
> > > > common
> > > >    component for config management/tracking and refactor the existing
> > > >    mechanisms to use that service
> > > >    - Standardize on exactly one component to manage the configs and
> > > regress
> > > >    the others (that's a verb, right?   nicer than delete.)
> > > >
> > > > I happen to like the service approach, myself, but I wanted to put it
> > up
> > > > for discussion and hopefully someone will volunteer to design such a
> > > thing.
> > > >
> > > > To frame the debate, I want us to keep in mind a couple of things
> that
> > > may
> > > > or may not be relevant to the discussion:
> > > >
> > > >    - We will eventually be moving to support kerberos so there should
> > at
> > > >    least be a path to use kerberos for any solution IMO
> > > >    - There is value in each of the different mechanisms in place now.
> > If
> > > >    there weren't, then they wouldn't have been created.  Before we
> try
> > to
> > > > make
> > > >    this a "there can be only one" argument, I'd like to hear very
> good
> > > >    arguments.
> > > >
> > > > Finally, I'd appreciate if some people might answer the questions I
> > have
> > > in
> > > > bold there.  Hopefully this discussion, if nothing else happens, will
> > > > result in fodder for proper documentation of the ins and outs of each
> > of
> > > > the components bulleted above.
> > > >
> > > > Best,
> > > >
> > > > Casey
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > >
> > >
> > > --
> > > Nick Allen <n...@nickallen.org>
> > >
> >
>

Reply via email to