I was hoping to get an opinion on this. Maybe it's something that's been seen before.
I see this in the kadmind log: Feb 7 17:00:10 server.realm.com kadmind[17670]: [ID 702911 local0.notice] Request: kadm5_create_principal, bar...@realm.com, Cannot lock database, client=f...@realm.com, service=kadmin/server.realm....@realm.com, addr=1.2.3.4 The create_principal failed...we even check in our provisioning client: Feb 7 17:00:39 server.realm.com kadmind[17670]: [ID 702911 local0.notice] Request: kadm5_get_principal, bar...@realm.com, Principal does not exist, client=f...@realm.com, service=kadmin/server.realm....@realm.com, addr=1.2.3.4 BUT the propagation log says this: Update Entry Update serial # : 150028 Update operation : Add Update principal : bar...@realm.com Update size : 524 Update committed : True Update time stamp : Thu Feb 7 17:00:00 2013 Attributes changed : 12 Attribute flags Maximum ticket life Maximum renewable life Principal expiration Password expiration Principal Key data Password last changed Modifying principal Modification time TL data Length Note the timestamps...the update entry appeared 10s earlier than the kadmind log error! There was no previous create_principal request 10s earlier. Later on, we attempt to create the principal again with success reported this time... Feb 7 17:00:56 server.realm.com kadmind[17670]: [ID 702911 local0.notice] Request: kadm5_create_principal, bar...@realm.com, success, client=f...@realm.com, service=kadmin/server.realm....@realm.com, addr=1.2.3.4 And we get another update entry (same timestamp this time): Update Entry Update serial # : 150030 Update operation : Add Update principal : bar...@realm.com Update size : 524 Update committed : True Update time stamp : Thu Feb 7 17:00:56 2013 Attributes changed : 12 Attribute flags Maximum ticket life Maximum renewable life Principal expiration Password expiration Principal Key data Password last changed Modifying principal Modification time TL data Length The fun really starts when the incremental propagation kicks in. We have 7 secondary kdc's and in the case that I am investigating; 4 of them got Update 150028 right away, on its own, with no problems. The other 3 had Update 150028 bundled with one or more additional updates and the kpropd on these three servers started core dumping over and over again until we forced a full resync. Please let me know if this is a known bug. If it is not, maybe the code changes from 1.8 to 1.11 have made this behavior stop. It would be nice to be able to reliably reproduce this so we can then test post-1.8.3 versions of the code. So far, this has his us twice in the last couple of years. jd
________________________________________________ Kerberos mailing list Kerberos@mit.edu https://mailman.mit.edu/mailman/listinfo/kerberos