On Fri, Sep 20, 2019 at 11:10 AM Paul Emmerich <paul.emmerich(a)croit.io> wrote:
Probably easiest if you get a tape library that supports S3. You might
even have some luck with radosgw's cloud sync module (but I wouldn't
count on it, Octopus should improve things, though)
Just intercepting PUT requests isn't that easy because of multi-part
stuff and load balancing. I.e., if you upload a large file you should
be sending it in chunks and each chunk should go to a different
server, that makes any "simple" solutions pretty messy.
I wasn't aware of any library being S3 aware, usually it's been part
of the backup software. Do you have any suggestions for multi PB
libraries that have the S3 feature?
The idea with the PUT was not to intercept them in the path, but to
basically have RGW log access to LogStash, then a job would run to
find all the objects that were PUT within a time frame, then read the
objects off the cluster and write them to tape. Maybe that's not as
easy as I'm thinking either.
----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1