Hello,
* 2 x Xeon Silver 4212 (12C/24T)
I would choose single cpu AMD EPYC systems for lower price with better performance. Supermicro does have some good systems for AMD as well.
* 16 x 10 TB nearline SAS HDD (8 bays for future needs)
Don't waste money here as well. No real gain. Invest it better in more or faster (ssd) disks.
* 4 x 40G QSFP+
With 24x spinning media, even a single 40G link will be enough. No gain for a lot of money again.
* 2 x 40G per server for ceph network (LACP/VPC for HA)
* 2 x 40G per server for public network (LACP/VPC for HA)
Use vlans if you really want to separate the networks. Most of the time we see new customers coming in with problems on such configurations and we don't suggest tu configure Ceph that way from our experience.
* ZFS on RBD, exposed via samba shares (cluster with failover)
Maybe, just maybe think about just using samba on top of cephfs to export the data. No need for all the overhead and possible bugs you would encounter.
* We're used to run mons and mgrs daemons on a few of our OSD nodes, without any issue so far : is this a bad idea for a big cluster ?
We always do so and never had a problem with it. Just make sure the MON has enough resources for your workload.
* We thought using cache tiering on an SSD pool, but a large part of the PB is used on a daily basis, so we expect the cache to be not so effective and really expensive ?
Tend to be error prone and we saw a lot of cluster meltdowns in the last 7 years due to cache tiering. Just go for an all flash cluster use db/wal devices to improve performance.
* Could a 2x10G network be enough ?
Yes ;), but maybe on recovery workloads it will slow down the recovery a bit. However I don't believe that it will be a problem in your mentioned szenario.
* ZFS on Ceph ? Any thoughts ?
just don't ;)
* What about CephFS ? We'd like to use RBD diff for backups but it looks impossible to use snapshot diff with Cephfs ?
If you do have questions or want some consulting to get the best Ceph cluster for the job. Please feel free to contact us.