Package: git-lfs Version: 3.0.2-1 Severity: wishlist X-Debbugs-Cc: none, Barak A. Pearlmutter <ba...@pearlmutter.net>
I have a repo whose only remote is on a gitlab instance. I'm using git-lfs to manage large binary files in this repo. The remote goes down. Now "git add foo.pdf / git commit", when *.pdf files are tracked by lfs, freezes! Waits forever for the remote when trying to transfer the big blobs. This violates what I consider a central concept of git, namely that operations are local unless you explicitly fetch or push. It means you cannot work with lfs while offline, like on an aeroplane, or even (as above) when the gitlab instance is offline for maintenance. There is also a potential security issue. Users might reasonably assume they can safely do "git add/commit/rebase" operations locally, with intermediate steps exposing secret information that is later removed before doing a push. Nope! Anyway: I *wish* git-lfs allowed remote operation, like git-annex does. It seems like it should be technically possible to wait until an lfs-tracked file (well, its https://git-lfs.github.com/spec/v1 smudge stub) is actually pushed before transferring the associated big binary blob. Or at the very least, giving up and remembering to try again later if there's a big binary blob transfer problem. -- System Information: Versions of packages git-lfs depends on: ii git 1:2.34.1-1 ii libc6 2.33-5