Hello,
there is an app, written in Python, which stores few bytes of
datas in a single file. The application uses threads. Every
thread can modify the file, but only one at a time. I'm using a
lock file to prevent the multiple access.
Here is the lock method:
while True:
try:
fl = os.open(self.lockfile, os.O_CREAT | os.O_EXCL | os.O_RDWR)
except OSError, e:
if e.errno != errno.EEXIST:
raise
time.sleep(0.2)
continue
except:
syslog.syslog(syslog.LOG_DEBUG, "Sync error: " +
str(sys.exc_info()[1]))
else:
break
This works as well for me - about 3-4 weeks. After some weeks, I
got this error:
OSError: [Errno 24] Too many open files: '/var/spool/myapp/queue.lock'
Today the app had been restarted, about 3-4 hours ago. Now I see
under the proc/PID/fd:
lrwx------ 1 root root 64 márc 31 16.45 5 ->
/var/spool/myapp/queue.lock (deleted)
there are about 50 deleted FD's. After few weeks the process
reaches the number if max fd's.
How can I prevent or avoid this issue? What's the correct way to
handle the lockfile in Python?
Thanks,
Ervin
--
https://mail.python.org/mailman/listinfo/python-list