On Mon, 2020-08-10 at 19:36 -0400, Chuck Lever wrote: > > On Aug 10, 2020, at 11:35 AM, James Bottomley > > <james.bottom...@hansenpartnership.com> wrote: > > On Sun, 2020-08-09 at 13:16 -0400, Mimi Zohar wrote: > > > On Sat, 2020-08-08 at 13:47 -0400, Chuck Lever wrote: [...] > > > > The first priority (for me, anyway) therefore is getting the > > > > ability to move IMA metadata between NFS clients and servers > > > > shoveled into the NFS protocol, but that's been blocked for > > > > various legal reasons. > > > > > > Up to now, verifying remote filesystem file integrity has been > > > out of scope for IMA. With fs-verity file signatures I can at > > > least grasp how remote file integrity could possibly work. I > > > don't understand how remote file integrity with existing IMA > > > formats could be supported. You might want to consider writing a > > > whitepaper, which could later be used as the basis for a patch > > > set cover letter. > > > > I think, before this, we can help with the basics (and perhaps we > > should sort them out before we start documenting what we'll do). > > Thanks for the help! I just want to emphasize that documentation > (eg, a specification) will be critical for remote filesystems. > > If any of this is to be supported by a remote filesystem, then we > need an unencumbered description of the new metadata format rather > than code. GPL-encumbered formats cannot be contributed to the NFS > standard, and are probably difficult for other filesystems that are > not Linux-native, like SMB, as well.
I don't understand what you mean by GPL encumbered formats. The GPL is a code licence not a data or document licence. The way the spec process works in Linux is that we implement or evolve a data format under a GPL implementaiton, but that implementation doesn't implicate the later standardisation of the data format and people are free to reimplement under any licence they choose. > > The first basic is that a merkle tree allows unit at a time > > verification. First of all we should agree on the unit. Since we > > always fault a page at a time, I think our merkle tree unit should > > be a page not a block. > > Remote filesystems will need to agree that the size of that unit is > the same everywhere, or the unit size could be stored in the per-file > metadata. > > > > Next, we should agree where the check gates for the per page > > accesses should be ... definitely somewhere in readpage, I suspect > > and finally we should agree how the merkle tree is presented at the > > gate. I think there are three ways: > > > > 1. Ahead of time transfer: The merkle tree is transferred and > > verified > > at some time before the accesses begin, so we already have a > > verified copy and can compare against the lower leaf. > > 2. Async transfer: We provide an async mechanism to transfer the > > necessary components, so when presented with a unit, we check > > the > > log n components required to get to the root > > 3. The protocol actually provides the capability of 2 (like the > > SCSI > > DIF/DIX), so to IMA all the pieces get presented instead of > > IMA > > having to manage the tree > > A Merkle tree is potentially large enough that it cannot be stored in > an extended attribute. In addition, an extended attribute is not a > byte stream that you can seek into or read small parts of, it is > retrieved in a single shot. Well you wouldn't store the tree would you, just the head hash. The rest of the tree can be derived from the data. You need to distinguish between what you *must* have to verify integrity (the head hash, possibly signed) and what is nice to have to speed up the verification process. The choice for the latter is cache or reconstruct depending on the resources available. If the tree gets cached on the server, that would be a server implementation detail invisible to the client. > For this reason, the idea was to save only the signature of the > tree's root on durable storage. The client would retrieve that > signature possibly at open time, and reconstruct the tree at that > time. Right that's the integrity data you must have. > Or the tree could be partially constructed on-demand at the time each > unit is to be checked (say, as part of 2. above). Whether it's reconstructed or cached can be an implementation detail. You clearly have to reconstruct once, but whether you have to do it again depends on the memory available for caching and all the other resource calls in the system. > The client would have to reconstruct that tree again if memory > pressure caused some or all of the tree to be evicted, so perhaps an > on-demand mechanism is preferable. Right, but I think that's implementation detail. Probably what we need is a way to get the log(N) verification hashes from the server and it's up to the client whether it caches them or not. > > There are also a load of minor things like how we get the head > > hash, which must be presented and verified ahead of time for each > > of the above 3. > > Also, changes to a file's content and its tree signature are not > atomic. If a file is mutable, then there is the period between when > the file content has changed and when the signature is updated. > Some discussion of how a client is to behave in those situations will > be necessary. For IMA, if you write to a checked file, it gets rechecked the next time the gate (open/exec/mmap) is triggered. This means you must complete the update and have the new integrity data in-place before triggering the check. I think this could apply equally to a merkel tree based system. It's a sort of Doctor, Doctor it hurts when I do this situation. James