On 9/19/22 8:48 AM, Peter Krempa wrote:
On Wed, Aug 31, 2022 at 13:40:45 -0500, Jonathon Jongsma wrote:
After a bit of a lengthy delay, this is the second version of this patch
series. See https://bugzilla.redhat.com/show_bug.cgi?id=2016527 for more
information about the goal, but the summary is that RHEL does not want to ship
the qemu storage plugins for curl and ssh.  Handling them outside of the qemu
process provides several advantages such as reduced attack surface and
stability.

IMO it's also worthy noting that it increases complexity of the setup
and potentially also resource usage.

A quick summary of the code:

[...]

Open questions
  - selinux: I need some help from people more familiar with selinux to figure
    out what is needed here. When selinux is enforcing, I get a failure to
    launch nbdkit to serve the disks. I suspect we need a new context and policy
    for /usr/sbin/nbdkit that allows it to transition to the appropriate selinux
    context. The current context (on fedora) is "system_u:object_r:bin_t:s0".
    When I (temporarily) change the context to something like qemu_exec_t,
    I am able to start nbdkit and the domain launches.

Is the problem with starting the 'nbdkit' process itself or with the
socket?

The problem seems to be with the nbdkit process itself. Because I use qemuSecurityCommandRun() to launch nbdkit, it sets the security label for the nbkit process to be the security label for the domain and then attempts to launch it. Presumably there is no rule allowing the binaries with the bin_t context to transition to the label for the domain. But as I said, my selinux knowledge is quite shallow, so I may need some help figuring out the proper way to do this.


At least in case of the socket we must make sure that no other process
can acess it especially once you pass authentication to nbdkit to avoid
any kind of backdoor to authenticated storage.

nbdkit does have a --selinux-label argument which will set the label of the socket.


Few more open questions:
  - What if 'nbdkit' crashes
     With an integrated block layer, all of the VM crashes. Now when we
     have a separated access to disks (this is also an issue for use of
     the qemu-storage-daemon) if any of the helper processes crash we get
     into a new situation.

     I think we'll need to think about:
         - adding an event for any of the helper processes failing
         - adding a lifecycle action for it (e.g. pause qemu if nbdkit
           dies)
         - think about possible recovery of the situation

Indeed. Seems like something re-usable that could also be used for qemu-storage-daemon might be useful. I'll look into it.


  - resource pinning
    For now all the resources are integral to the qemu process so
    emulator and iothread pinning can be used to steer which cpus the
    disk should use. With us adding new possibly cpu intensive processes
    we'll probably need to consider how to handle them more generally and
    manage their resources.

Good point. I hadn't really thought about this.


  - Integration with qemu storage daemon used for a VM

    With the attempt to rewrite QSD in other languages it will possibly
    make sense to run it instead of the native qemu block layer (e.g.
    take advantage of memory safe languages). So we should also think
    about how these two will be able to coexist.

The last point is more of a future-work thing to consider, but the
first two points should be considered for the final release of this
feature. Specifically because in your current design it replaces the
in-qemu driver even in cases when it is compiled into qemu thus also for
existing users. Alternatively it would have to be opt-in.

As far as I can tell, this is currently no good way to determine (via capabilities or otherwise) whether e.g. the curl driver is compiled into a particular qemu binary. If I've missed something, please let me know. My recollection of previous discussions was that I should try to use nbdkit if available and then fall back to trying the built-in qemu driver approach. But we could change this strategy if there's a good way to do it. When you talk about opt-in, do you mean per-domain? e.g. in the xml? or some other mechanism?


Known shortcomings
  - creating disks (in ssh) still isn't supported. I wanted to send out the
    patch series anyway since it's been delayed too long already.

That shouldn't be a problem, there's plenty protocols where we don't
support creating the storage. Creating storage is needed only for
snapshots so we can simply refuse to do it in the first place.


Reply via email to