It sort of looks like the out of space triggered the issue. Libvirt shows

2013-10-03 15:38:57.414+0000: 2710: error : virCommandWait:2348 :
internal error Child process (/bin/umount
/mnt/5230667e-9c58-3ff6-983c-5fc2a72df669) unexpected exit status 32:
error writing /etc/mtab.tmp: No space left on device

So that entry of the filesystem being mounted is orphaned in
/etc/mtab, since it can't be removed. That seems to be the source of
why 'df' shows the thing mounted when it isn't, at least.


On Thu, Oct 3, 2013 at 11:45 AM, Indra Pramana <in...@sg.or.id> wrote:
> Hi Marcus and all,
>
> I also find some strange and interesting error messages from the libvirt
> logs:
>
> http://pastebin.com/5ByfNpAf
>
> Looking forward to your reply, thank you.
>
> Cheers.
>
>
>
> On Fri, Oct 4, 2013 at 1:38 AM, Indra Pramana <in...@sg.or.id> wrote:
>
>> Hi Marcus,
>>
>> Good day to you, and thank you for your e-mail. See my reply inline.
>>
>> On Fri, Oct 4, 2013 at 12:29 AM, Marcus Sorensen <shadow...@gmail.com>wrote:
>>
>>> Are you using the release artifacts, or your own 4.2 build?
>>>
>>
>> [Indra:] We are using the release artifacts from below repo since we are
>> using Ubuntu:
>>
>> deb http://cloudstack.apt-get.eu/ubuntu precise 4.2
>>
>>
>>> When you rebooted the host, did the problem go away or come back the
>>> same?
>>
>>
>> [Indra:] When I rebooted the host, the problem go away for a while, but it
>> will come back again after some time. It will come randomly at the time
>> when we need to create a new instance on that host, or start an existing
>> stopped instance.
>>
>>
>>> You may want to look at 'virsh pool-list' to see if libvirt is
>>> mounting/registering the secondary storage.
>>>
>>
>> [Indra:] This is the result of the virsh pool-list command:
>>
>> root@hv-kvm-02:/var/log/libvirt# virsh pool-list
>> Name                 State      Autostart
>> -----------------------------------------
>> 301071ac-4c1d-4eac-855b-124126da0a38 active     no
>> 5230667e-9c58-3ff6-983c-5fc2a72df669 active     no
>> d433809b-01ea-3947-ba0f-48077244e4d6 active     no
>>
>> Strange thing is that none of my secondary storage IDs are there. Could it
>> be that the ID might have changed during Cloudstack upgrade? Here is the
>> list of my secondary storage (there are two of them) even though they are
>> on the same NFS server:
>>   c02da448-b9f4-401b-b8d5-83e8ead5cfde nfs://
>> 10.237.11.31/mnt/vol1/sec-storage NFS
>> 5937edb6-2e95-4ae2-907b-80fe4599ed87 nfs://
>> 10.237.11.31/mnt/vol1/sec-storage2 NFS
>>
>>> Is this happening on multiple hosts, the same way?
>>
>>
>> [Indra:] Yes, it is happening on all the two hosts that I have, the same
>> way.
>>
>>
>>> You may want to
>>> look at /etc/mtab, if the system reports it's mounted, though it's
>>> not, it might be in there. Look at /proc/mounts as well.
>>>
>>
>> [Indra:] Please find result of df, /etc/mtab and /proc/mounts below. The
>> "ghost" mount point is on df and /etc/mtab, but not on /proc/mounts.
>>
>> root@hv-kvm-02:/etc# df
>>
>> Filesystem                                              1K-blocks    Used
>> Available Use% Mounted on
>> /dev/sda1                                                 4195924
>> 4192372         0 100% /
>>
>> udev                                                    132053356       4
>> 132053352   1% /dev
>> tmpfs                                                    52825052     704
>> 52824348   1% /run
>>
>> none                                                         5120
>> 0      5120   0% /run/lock
>> none                                                    132062620       0
>> 132062620   0% /run/shm
>> cgroup                                                  132062620       0
>> 132062620   0% /sys/fs/cgroup
>> /dev/sda6                                                10650544
>> 2500460   7609056  25% /home
>> 10.237.11.31:/mnt/vol1/sec-storage2/template/tmpl/2/288   4195924
>> 4192372         0 100% /mnt/5230667e-9c58-3ff6-983c-5fc2a72df669
>>
>> root@hv-kvm-02:/etc# cat /etc/mtab
>> /dev/sda1 / ext4 rw,errors=remount-ro 0 0
>> proc /proc proc rw,noexec,nosuid,nodev 0 0
>> sysfs /sys sysfs rw,noexec,nosuid,nodev 0 0
>> none /sys/fs/fuse/connections fusectl rw 0 0
>> none /sys/kernel/debug debugfs rw 0 0
>> none /sys/kernel/security securityfs rw 0 0
>> udev /dev devtmpfs rw,mode=0755 0 0
>> devpts /dev/pts devpts rw,noexec,nosuid,gid=5,mode=0620 0 0
>> tmpfs /run tmpfs rw,noexec,nosuid,size=10%,mode=0755 0 0
>> none /run/lock tmpfs rw,noexec,nosuid,nodev,size=5242880 0 0
>> none /run/shm tmpfs rw,nosuid,nodev 0 0
>> cgroup /sys/fs/cgroup tmpfs rw,relatime,mode=755 0 0
>> cgroup /sys/fs/cgroup/cpuset cgroup rw,relatime,cpuset 0 0
>> cgroup /sys/fs/cgroup/cpu cgroup rw,relatime,cpu 0 0
>> cgroup /sys/fs/cgroup/cpuacct cgroup rw,relatime,cpuacct 0 0
>> cgroup /sys/fs/cgroup/memory cgroup rw,relatime,memory 0 0
>> cgroup /sys/fs/cgroup/devices cgroup rw,relatime,devices 0 0
>> cgroup /sys/fs/cgroup/freezer cgroup rw,relatime,freezer 0 0
>> cgroup /sys/fs/cgroup/blkio cgroup rw,relatime,blkio 0 0
>> cgroup /sys/fs/cgroup/perf_event cgroup rw,relatime,perf_event 0 0
>> /dev/sda6 /home ext4 rw 0 0
>> rpc_pipefs /run/rpc_pipefs rpc_pipefs rw 0 0
>> binfmt_misc /proc/sys/fs/binfmt_misc binfmt_misc rw,noexec,nosuid,nodev 0 0
>> 10.237.11.31:/mnt/vol1/sec-storage2/template/tmpl/2/288
>> /mnt/5230667e-9c58-3ff6-983c-5fc2a72df669 nfs rw,addr=10.237.11.31 0 0
>>
>> rootfs / rootfs rw 0 0
>> sysfs /sys sysfs rw,nosuid,nodev,noexec,relatime 0 0
>> proc /proc proc rw,nosuid,nodev,noexec,relatime 0 0
>> udev /dev devtmpfs rw,relatime,size=132053356k,nr_inodes=33013339,mode=755
>> 0 0
>> devpts /dev/pts devpts
>> rw,nosuid,noexec,relatime,gid=5,mode=620,ptmxmode=000 0 0
>> tmpfs /run tmpfs rw,nosuid,relatime,size=52825052k,mode=755 0 0
>> /dev/disk/by-uuid/d21f4676-075d-44be-a378-92ef6aaf2496 / ext4
>> rw,relatime,errors=remount-ro,data=ordered 0 0
>> none /sys/fs/fuse/connections fusectl rw,relatime 0 0
>> none /sys/kernel/debug debugfs rw,relatime 0 0
>> cgroup /sys/fs/cgroup tmpfs rw,relatime,mode=755 0 0
>> none /sys/kernel/security securityfs rw,relatime 0 0
>> none /run/lock tmpfs rw,nosuid,nodev,noexec,relatime,size=5120k 0 0
>> none /run/shm tmpfs rw,nosuid,nodev,relatime 0 0
>> cgroup /sys/fs/cgroup/cpuset cgroup rw,relatime,cpuset 0 0
>> cgroup /sys/fs/cgroup/cpu cgroup rw,relatime,cpu 0 0
>> cgroup /sys/fs/cgroup/cpuacct cgroup rw,relatime,cpuacct 0 0
>> cgroup /sys/fs/cgroup/memory cgroup rw,relatime,memory 0 0
>> cgroup /sys/fs/cgroup/devices cgroup rw,relatime,devices 0 0
>> cgroup /sys/fs/cgroup/freezer cgroup rw,relatime,freezer 0 0
>> cgroup /sys/fs/cgroup/blkio cgroup rw,relatime,blkio 0 0
>> cgroup /sys/fs/cgroup/perf_event cgroup rw,relatime,perf_event 0 0
>> /dev/sda6 /home ext4 rw,relatime,data=ordered 0 0
>> rpc_pipefs /run/rpc_pipefs rpc_pipefs rw,relatime 0 0
>> binfmt_misc /proc/sys/fs/binfmt_misc binfmt_misc
>> rw,nosuid,nodev,noexec,relatime 0 0
>>
>> If I tried to un-mount (umount) the "ghost" mount point, it will fail with
>> below error message saying that it is not found in /proc/mounts:
>>
>> root@hv-kvm-02:/etc# umount /mnt/5230667e-9c58-3ff6-983c-5fc2a72df669
>> /mnt/5230667e-9c58-3ff6-983c-5fc2a72df669 was not found in /proc/mounts
>> /mnt/5230667e-9c58-3ff6-983c-5fc2a72df669 was not found in /proc/mounts
>>
>> Do you have any idea what could be the problem? Is it libvirt issue or
>> CloudStack issue?
>>
>> Any help on this is greatly appreciated. Since I am upgrading a production
>> setup, we only have 1-2 hours left to resolve the problem, otherwise we
>> will be forced to revert back to 4.1.1 (for the fourth time). :(
>>
>> Urgently looking forward to your reply, thank you.
>>
>> Cheers.
>>
>>
>>
>>
>>>
>>> On Thu, Oct 3, 2013 at 9:53 AM, Indra Pramana <in...@sg.or.id> wrote:
>>> > Dear all,
>>> >
>>> > We face a major problem after upgrading to 4.2.0. Mounting from KVM
>>> hosts
>>> > to secondary storage seems to fail, every time a new VM instance is
>>> > created, it will use up the / (root) partition of the KVM hosts instead.
>>> >
>>> > Here is the df result:
>>> >
>>> > ====
>>> > root@hv-kvm-02:/home/indra# df
>>> > Filesystem                                              1K-blocks
>>>  Used
>>> > Available Use% Mounted on
>>> > /dev/sda1                                                 4195924
>>> > 4195924         0 100% /
>>> > udev                                                    132053356
>>> 4
>>> > 132053352   1% /dev
>>> > tmpfs                                                    52825052
>>> 440
>>> > 52824612   1% /run
>>> > none                                                         5120
>>> > 0      5120   0% /run/lock
>>> > none                                                    132062620
>>> 0
>>> > 132062620   0% /run/shm
>>> > cgroup                                                  132062620
>>> 0
>>> > 132062620   0% /sys/fs/cgroup
>>> > /dev/sda6                                                10650544
>>> 2500424
>>> > 7609092  25% /home
>>> > 10.237.11.31:/mnt/vol1/sec-storage2/template/tmpl/2/288   4195924
>>> > 4195924         0 100% /mnt/5230667e-9c58-3ff6-983c-5fc2a72df669
>>> > ====
>>> >
>>> > The strange thing is that df shows that it seems to be mounted, but it's
>>> > actually not mounted. If you noticed, the total capacity of the mount
>>> point
>>> > is exactly the same as the capacity of the / (root) partition. By right
>>> it
>>> > should show 7 TB instead of just 4 GB.
>>> >
>>> > This caused VM creation to be error due to out of disk space. This also
>>> > affect the KVM operations since the / (root) partition becomes full,
>>> and we
>>> > can only release the space after we reboot the KVM host.
>>> >
>>> > Anyone experience this problem before? We are at loss on how to resolve
>>> the
>>> > problem.
>>> >
>>> > Looking forward to your reply, thank you.
>>> >
>>> > Cheers.
>>>
>>
>>

Reply via email to