"Poison BL." <poiso...@gmail.com> writes:

> On Sat, Apr 29, 2017 at 3:24 PM, lee <l...@yagibdah.de> wrote:
>
>> Mick <michaelkintz...@gmail.com> writes:
>>
>> > On Tuesday 25 Apr 2017 16:45:37 Alan McKinnon wrote:
>> >> On 25/04/2017 16:29, lee wrote:
>> >> > Hi,
>> >> >
>> >> > since the usage of FTP seems to be declining, what is a replacement
>> >> > which is at least as good as FTP?
>> >> >
>> >> > I'm aware that there's webdav, but that's very awkward to use and
>> >> > missing features.
>> >>
>> >> Why not stick with ftp?
>> >> Or, put another way, why do you feel you need to use something else?
>> >>
>> >> There's always dropbox
>> >
>> >
>> > Invariably all web hosting ISPs offer ftp(s) for file upload/download.
>> If you
>> > pay a bit more you should be able to get ssh/scp/sftp too.  Indeed, many
>> ISPs
>> > throw in scp/sftp access as part of their basic package.
>> >
>> > Webdav(s) offers the same basic upload/download functionality, so I am
>> not
>> > sure what you find awkward about it, although I'd rather use lftp
>> instead of
>> > cadaver any day. ;-)
>> >
>> > As Alan mentioned, with JavaScript'ed web pages these days there are many
>> > webapp'ed ISP offerings like Dropbox and friends.
>> >
>> > What is the use case you have in mind?
>>
>> transferring large amounts of data and automatization in processing at
>> least some of it, without involving a 3rd party
>>
>> "Large amounts" can be "small" like 100MB --- or over 50k files in 12GB,
>> or even more.  The mirror feature of lftp is extremely useful for such
>> things.
>>
>> I wouldn't ever want having to mess around with web pages to figure out
>> how to do this.  Ftp is plain and simple.  So you see why I'm explicitly
>> asking for a replacement which is at least as good as ftp.
>>
>>
>> --
>> "Didn't work" is an error.
>>
>>
> Half petabyte datasets aren't really something I'd personally *ever* trust
> ftp with in the first place.

Why not?  (12GB are nowhere close to half a petabyte ...)

> That said, it depends entirely on the network
> you're working with. Are you pushing this data in/out of the network your
> machines live in, or are you working primarily internally? If internal,
> what're the network side capabilities you have? Since you're likely already
> using something on the order of CEPH or Gluster to back the datasets where
> they sit, just working with it all across network from that storage would
> be my first instinct.

The data would come in from suppliers.  There isn't really anything
going on atm but fetching data once a month which can be like 100MB or
12GB or more.  That's because ppl don't use ftp ...

> How often does it need moved in/out of your facility, and is there no way
> to break up the processing into smaller chunks than a 0.6PB mass of files?
> Distribute out the smaller pieces with rsync, scp, or the like, operate on
> them, and pull back in the results, rather than trying to shift around the
> entire set. There's a reason Amazon will send a physical truck to a site to
> import large datasets into glacier... ;)

Amazon has trucks?  Perhaps they do in other countries.  Here, amazon is
just another web shop.  They might have some delivery vans, but I've
never seen one, so I doubt it.  And why would anyone give them their
data?  There's no telling what they would do with it.


-- 
"Didn't work" is an error.

Reply via email to