We are making hourly snapshots of ~400 rbd drives in one (spinning-rust)
cluster. The snapshots are made one by one.
Total size of the base images is around 80TB. The entire process takes a
few minutes.
We do not experience any problems doing this.
Op do 30 jan. 2020 om 15:30 schreef Adam Boyhan <adamb(a)medent.com>om>:
We are looking to role out a all flash Ceph cluster as
storage for our
cloud solution. The OSD's will be on slightly slower Micron 5300 PRO's,
with WAL/DB on Micron 7300 MAX NVMe's.
My main concern with Ceph being able to fit the bill is its snapshot
abilities.
For each RBD we would like the following snapshots
8x 30 minute snapshots (latest 4 hours)
With our current solution (HPE Nimble) we simply pause all write IO on the
10 minute mark for roughly 2 seconds and then we take a snapshot of the
entire Nimble volume. Each VM within the Nimble volume is sitting on a
Linux Logical Volume so its easy for us to take one big snapshot and only
get access to a specific clients data.
Are there any options for automating managing/retention of snapshots
within Ceph besides some bash scripts? Is there anyway to take snapshots of
all RBD's within a pool at a given time?
Is there anyone successfully running with this many snapshots? If anyone
is running a similar setup, would love to hear how your doing it.
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io