Hi

I'm getting this INFO message from losetup I'm using in my scripts:
if more info is needed from my side just ask.

(haven't seen this with 2.6.24..., using T61, 64bit kernel)

Bye

Zdenek

 =======================================================
 [ INFO: possible circular locking dependency detected ]
 2.6.25-rc1 #29
 -------------------------------------------------------
 losetup/26595 is trying to acquire lock:
  (&bdev->bd_mutex){--..}, at: [<ffffffff810e6258>] __blkdev_put+0x38/0x1e0

 but task is already holding lock:
  (&lo->lo_ctl_mutex){--..}, at: [<ffffffff883d097e>] lo_ioctl+0x4e/0xaf0 [loop]

 which lock already depends on the new lock.


 the existing dependency chain (in reverse order) is:

 -> #1 (&lo->lo_ctl_mutex){--..}:
        [<ffffffff81066f80>] __lock_acquire+0xf70/0x1000
        [<ffffffff883cf384>] lo_open+0x34/0x50 [loop]
        [<ffffffff810670b8>] lock_acquire+0xa8/0xf0
        [<ffffffff883cf384>] lo_open+0x34/0x50 [loop]
        [<ffffffff812f17ac>] mutex_lock_nested+0xcc/0x370
        [<ffffffff883cf384>] lo_open+0x34/0x50 [loop]
        [<ffffffff883cf384>] lo_open+0x34/0x50 [loop]
        [<ffffffff810e67c9>] do_open+0xa9/0x350
        [<ffffffff812f3b80>] _spin_unlock+0x30/0x60
        [<ffffffff810e6aad>] blkdev_open+0x3d/0x90
        [<ffffffff810b61b0>] __dentry_open+0x110/0x280
        [<ffffffff810e6a70>] blkdev_open+0x0/0x90
        [<ffffffff810b63e4>] nameidata_to_filp+0x44/0x60
        [<ffffffff810b644a>] do_filp_open+0x4a/0x50
        [<ffffffff810b603d>] get_unused_fd_flags+0xed/0x140
        [<ffffffff810b64c6>] do_sys_open+0x76/0x100
        [<ffffffff810b657b>] sys_open+0x1b/0x20
        [<ffffffff8100c8ae>] tracesys+0xdc/0xe1
        [<ffffffffffffffff>] 0xffffffffffffffff

 -> #0 (&bdev->bd_mutex){--..}:
        [<ffffffff81066e8d>] __lock_acquire+0xe7d/0x1000
        [<ffffffff810658d9>] trace_hardirqs_on+0x139/0x1a0
        [<ffffffff810e6258>] __blkdev_put+0x38/0x1e0
        [<ffffffff810670b8>] lock_acquire+0xa8/0xf0
        [<ffffffff810e6258>] __blkdev_put+0x38/0x1e0
        [<ffffffff812f17ac>] mutex_lock_nested+0xcc/0x370
        [<ffffffff810e6258>] __blkdev_put+0x38/0x1e0
        [<ffffffff81014908>] native_sched_clock+0x98/0xa0
        [<ffffffff810e6258>] __blkdev_put+0x38/0x1e0
        [<ffffffff810e640b>] blkdev_put+0xb/0x10
        [<ffffffff810e6438>] blkdev_close+0x28/0x40
        [<ffffffff810b9530>] __fput+0xd0/0x1b0
        [<ffffffff810b962d>] fput+0x1d/0x30
        [<ffffffff883d091d>] loop_clr_fd+0x1ad/0x1c0 [loop]
        [<ffffffff883d0a9f>] lo_ioctl+0x16f/0xaf0 [loop]
        [<ffffffff810cccdf>] __d_lookup+0xef/0x180
        [<ffffffff81014908>] native_sched_clock+0x98/0xa0
        [<ffffffff81014908>] native_sched_clock+0x98/0xa0
        [<ffffffff8106634d>] __lock_acquire+0x33d/0x1000
        [<ffffffff8106634d>] __lock_acquire+0x33d/0x1000
        [<ffffffff810639a6>] get_lock_stats+0x16/0x60
        [<ffffffff81014908>] native_sched_clock+0x98/0xa0
        [<ffffffff810639a6>] get_lock_stats+0x16/0x60
        [<ffffffff810639fe>] put_lock_stats+0xe/0x30
        [<ffffffff812f2049>] mutex_unlock+0x9/0x10
        [<ffffffff8116731d>] blkdev_driver_ioctl+0x8d/0xa0
        [<ffffffff811675ad>] blkdev_ioctl+0x27d/0x7e0
        [<ffffffff81014908>] native_sched_clock+0x98/0xa0
        [<ffffffff8106634d>] __lock_acquire+0x33d/0x1000
        [<ffffffff81014908>] native_sched_clock+0x98/0xa0
        [<ffffffff810639a6>] get_lock_stats+0x16/0x60
        [<ffffffff810e5a7b>] block_ioctl+0x1b/0x20
        [<ffffffff810c67b1>] vfs_ioctl+0x31/0xa0
        [<ffffffff810c6aa3>] do_vfs_ioctl+0x283/0x2f0
        [<ffffffff810c6ba9>] sys_ioctl+0x99/0xa0
        [<ffffffff8100c8ae>] tracesys+0xdc/0xe1
        [<ffffffffffffffff>] 0xffffffffffffffff

 other info that might help us debug this:

 1 lock held by losetup/26595:
  #0:  (&lo->lo_ctl_mutex){--..}, at: [<ffffffff883d097e>]
lo_ioctl+0x4e/0xaf0 [loop]

 stack backtrace:
 Pid: 26595, comm: losetup Not tainted 2.6.25-rc1 #29

 Call Trace:
  [<ffffffff81065d54>] print_circular_bug_tail+0x84/0x90
  [<ffffffff81066e8d>] __lock_acquire+0xe7d/0x1000
  [<ffffffff810658d9>] ? trace_hardirqs_on+0x139/0x1a0
  [<ffffffff810e6258>] ? __blkdev_put+0x38/0x1e0
  [<ffffffff810670b8>] lock_acquire+0xa8/0xf0
  [<ffffffff810e6258>] ? __blkdev_put+0x38/0x1e0
  [<ffffffff812f17ac>] mutex_lock_nested+0xcc/0x370
  [<ffffffff810e6258>] ? __blkdev_put+0x38/0x1e0
  [<ffffffff81014908>] ? native_sched_clock+0x98/0xa0
  [<ffffffff810e6258>] __blkdev_put+0x38/0x1e0
  [<ffffffff810e640b>] blkdev_put+0xb/0x10
  [<ffffffff810e6438>] blkdev_close+0x28/0x40
  [<ffffffff810b9530>] __fput+0xd0/0x1b0
  [<ffffffff810b962d>] fput+0x1d/0x30
  [<ffffffff883d091d>] :loop:loop_clr_fd+0x1ad/0x1c0
  [<ffffffff883d0a9f>] :loop:lo_ioctl+0x16f/0xaf0
  [<ffffffff810cccdf>] ? __d_lookup+0xef/0x180
  [<ffffffff81014908>] ? native_sched_clock+0x98/0xa0
  [<ffffffff81014908>] ? native_sched_clock+0x98/0xa0
  [<ffffffff8106634d>] ? __lock_acquire+0x33d/0x1000
  [<ffffffff8106634d>] ? __lock_acquire+0x33d/0x1000
  [<ffffffff810639a6>] ? get_lock_stats+0x16/0x60
  [<ffffffff81014908>] ? native_sched_clock+0x98/0xa0
  [<ffffffff810639a6>] ? get_lock_stats+0x16/0x60
  [<ffffffff810639fe>] ? put_lock_stats+0xe/0x30
  [<ffffffff812f2049>] ? mutex_unlock+0x9/0x10
  [<ffffffff8116731d>] blkdev_driver_ioctl+0x8d/0xa0
  [<ffffffff811675ad>] blkdev_ioctl+0x27d/0x7e0
  [<ffffffff81014908>] ? native_sched_clock+0x98/0xa0
  [<ffffffff8106634d>] ? __lock_acquire+0x33d/0x1000
  [<ffffffff81014908>] ? native_sched_clock+0x98/0xa0
  [<ffffffff810639a6>] ? get_lock_stats+0x16/0x60
  [<ffffffff810e5a7b>] block_ioctl+0x1b/0x20
  [<ffffffff810c67b1>] vfs_ioctl+0x31/0xa0
  [<ffffffff810c6aa3>] do_vfs_ioctl+0x283/0x2f0
  [<ffffffff810c6ba9>] sys_ioctl+0x99/0xa0
  [<ffffffff8100c8ae>] tracesys+0xdc/0xe1
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to