Ah, in answer to my own question: http://reductivelabs.com/trac/puppet/wiki/ConfigurationReference
--ignorecache should do the trick. On Thu, Sep 17, 2009 at 7:14 AM, Pete Emerson <pemer...@gmail.com> wrote: > With this solution would I need to clear a cache? If I do two puppet > runs right after each other, doesn't puppet cache the recipes for a > period of time? If so, what do I need to do to wipe that local cache > out? > > Pete > > On Tue, Sep 15, 2009 at 2:42 PM, Pete Emerson <pemer...@gmail.com> wrote: >> >> Silviu, I think it's a pretty good solution, though. >> >> I'm actually contemplating writing a simple job scheduler that would >> eliminate this problem, but wanted to make sure that I'm not missing >> something obvious like a built-in queuing system or something like >> that. >> >> On Sep 15, 2:14 pm, Silviu Paragina <sil...@paragina.ro> wrote: >>> Now I realize that this is not so portable :-?? you could try creating a >>> simple pp file and run it with puppet (not puppetd) which would essentially >>> do the same thing. >>> >>> Silviu >>> >>> On Wed, 16 Sep 2009 00:02:07 +0300, Silviu Paragina <sil...@paragina.ro> >>> wrote: >>> >>> > The error message gives you the solution, check for the existence of >>> > /var/lib/puppet/state/puppetdlock. >>> >>> > My solution would be >>> >>> > invoke-rc.d puppet stop >>> > #or /etc/init.d/puppet or whatever >>> > while [ -f /var/lib/puppet/state/puppetdlock ] >>> > do >>> > sleep 1 >>> > done >>> >>> > #do your stuff >>> >>> > Silviu >>> >>> > On Tue, 15 Sep 2009 13:05:37 -0700, Pete Emerson <pemer...@gmail.com> >>> > wrote: >>> >> I'm using puppet (0.24, working on the 0.25 migration) to do rolling >>> >> upgrades across our datacenter. >>> >>> >> I'm running puppet as a daemon. >>> >>> >> In order to change an application version, I modify a database, which >>> >> in turn modifies the data that my puppet_node_classifier presents. I >>> >> then ssh to the nodes that I want to upgrade and force a puppet run >>> >> with puppetd --server=foo --test --report. >>> >>> >> The problem I'm running into is that on a regular basis a node is >>> >> already in the process of doing an update, and so I get back a message >>> >> like this: >>> >>> >> Lock file /var/lib/puppet/state/puppetdlock exists; skipping catalog run >>> >>> >> I can avoid this in some fashion by detecting this return result and >>> >> re-sshing into the node to run puppetd again, but this doesn't seem >>> >> very elegant. What are other people doing to avoid this sort of >>> >> situation? >>> >>> >> Pete >> >> >> > --~--~---------~--~----~------------~-------~--~----~ You received this message because you are subscribed to the Google Groups "Puppet Users" group. To post to this group, send email to puppet-users@googlegroups.com To unsubscribe from this group, send email to puppet-users+unsubscr...@googlegroups.com For more options, visit this group at http://groups.google.com/group/puppet-users?hl=en -~----------~----~----~----~------~----~------~--~---