Is there a protocol you all follow to be safe with incremental propagation? Like doing Full Sync, every now and then, maybe once a day or once a week?
Can the ulog or God forbid the database get clobbered corrupted and get replicated everywhere?? that would be huge problem, but I cant imaging this to be such a high risk. Have backups in place of database, daily I believe. Maybe should be making hourly backups just in case. Pondering away… Tareq > On Feb 20, 2020, at 2:04 PM, Tareq Alrashid <ta...@qerat.com> wrote: > > Well! we had a theory based on facts, that the last 1 update from the master > KDC blew all other replicas out of the water! > So I performed a "kproplog -R” got reset log and perform a Full Sync. > Restarted kprop and what do you know, the incrementals started working and > kprop never crashed again! > > Now we to wait on RHEL to see if coredump provides some insights on what > lightening mysterious event took place for that last password change action > by one insomniac at 02:11am that can cause this crash to ever happen again? > > Could it be the 3 second interval poll being too frequent?...etc. we have RCA > to come up with. > > Shared this in case others have to deal with, or in case someone did > experience this and can shed some light. > > As always any and all are welcome to chime in. > > Thank you, > Tareq > >> On Feb 20, 2020, at 10:19 AM, Tareq Alrashid <ta...@qerat.com >> <mailto:ta...@qerat.com>> wrote: >> >> RHEL 7.7 >> >> 2am all of a sudden all my replicas kpropd process crashed with a core dump. >> Nothing has changed! >> >> systemctl status kprop.service >> >> kprop.service: main process exited, code=dumped, status=6/ABRT >> Feb 20 09:35:13 kerb-replica systemd[1]: Unit kprop.service entered failed >> state. >> Feb 20 09:35:13 kerb-replica systemd[1]: kprop.service failed. >> >> >> kdc.conf on replica >> …. >> iprop_port = 754 >> iprop_slave_poll = 3s >> } >> >> Any ideas would be most appreciate, as we are in the weeds now looking at >> what has happened. >> >> Thank you in advance >> >> Tareq >> >>> On Jan 12, 2020, at 9:45 PM, TAREQ ALRASHID <ta...@qerat.com >>> <mailto:ta...@qerat.com>> wrote: >>> >>> Since my last message, I came across the documentation part were it >>> mentions the iprop_slave_poll. >>> We’re running Kerberos 5 release 1.15.1! - Will make the proper change and >>> the test results should make more sense! >>> >>> Thanks Greg. >>> >>> >>> >>>> On Jan 12, 2020, at 5:54 PM, Greg Hudson <ghud...@mit.edu >>>> <mailto:ghud...@mit.edu>> wrote: >>>> >>>> On 1/10/20 8:22 PM, Tareq Alrashid wrote: >>>>> Maybe I am missing something but changing the kdc.conf to any value... >>>>> >>>>> iprop_replica_poll=1s >>>>> or even... >>>>> iprop_replica_poll = 0.016666666666667m >>>>> (for 1s= 1/60min!) >>>>> >>>>> Based on tailing the kadmind.log, it is showing the replica polling >>>>> every 2m!? >>>> >>>> If you are running a release prior to 1.17, you need to use the old name >>>> . (The old name still works in 1.17 as well.) >>>> >>>> Also make sure to set the value on the machine running kpropd (not the >>>> master KDC where kadmind is run), and to restart kpropd. >>>> >>>> I don't think the delta time format supports floating point values, but >>>> "1s" or just "1" should work. >>>> >>>>> Final question if there is any negative impact for having replicas poll >>>>> at often as one second or maybe it is best to be at higher numbers of >>>>> seconds? >>>> Polling every second will cause a little bit of work on the replica and >>>> the master KDC each second, and use a little bit of network traffic. >>>> With today's computers and networks it's probably going to have much >>>> impact. >>> >> > ________________________________________________ Kerberos mailing list Kerberos@mit.edu https://mailman.mit.edu/mailman/listinfo/kerberos