On Fri, Sep 02, 2022 at 09:06:27AM +0100, Richard W.M. Jones wrote:
> On Thu, Sep 01, 2022 at 12:14:40PM +0100, Richard W.M. Jones wrote:
> > On Thu, Sep 01, 2022 at 04:01:39PM +0800, Ming Lei wrote:
> > > Maybe you should use one nbd disk, which has the closest code path
> > > with nbdublk.
> > 
> > Good idea - I have started a heavy test using nbd.ko as the
> > backing.  Let's see what happens after 12 hours or more.
> 
> This has been running overnight (using nbd.ko) and there are no
> problems.
> 
> Since I have seen 2 failures with ublk, I think there is a problem
> with the kernel driver, although it happens very rarely.
> 
> My tests were done using 6.0.0-rc3 (dcf8e5633e).

I believe it is still too early to say it is one ublk driver's issue.

The log shows that the following BUG_ON() is triggered, and ublk driver
is far away with this issue, and zero zspace->isolated is more like
one zsmalloc issue, and I highly suggest to report to zsmalloc/zram
community.

static void dec_zspage_isolation(struct zspage *zspage)
{
        VM_BUG_ON(zspage->isolated == 0);
        zspage->isolated--;
}

The effect from ublk could be that swap is more frequently or easy to
trigger since the ublk io code path needs to pin page from ublksrv's 
vm space, and nbd socket could consume more memory too.

BTW, I did run same kernel build workload on ublk-loop for ~20 hours,
and it did survive without any issue triggered. And all ublk userspace
block devices share totally same ublk driver io code path.

Also maybe you can try to run the same workload by replacing zram
swap disk with other swap disk, such as virito-scsi/..., then see if
it can survive.



Thanks,
Ming
_______________________________________________
Libguestfs mailing list
Libguestfs@redhat.com
https://listman.redhat.com/mailman/listinfo/libguestfs

Reply via email to