For now the storage new cluster used is more than old cluster whereas
the sync progress is about 50% to go. Besides, the amount of objects on
pool 'rgw.buckets.data' is larger too.
I'm not sure if the space of new cluster is enough or not for the whole
data.
Zhenshi Zhou <deaderzzs(a)gmail.com> 于2020年5月12日周二 上午11:40写道:
Hi,
I deployed a multisite in order to sync data from a mimic cluster zone
to a nautilus cluster zone. The data sync well at present. However,
I check the cluster status and I find something strange. The data in my
new cluster seems larger than that in old ones. The data is far from full
synced while the space used is nearly the same. Does that normal?
'ceph df ' on old cluster:
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
82 TiB 41 TiB 41 TiB 50.37
POOLS:
NAME ID USED %USED
MAX AVAIL OBJECTS
.rgw.root 1 6.0 KiB 0
10 TiB 19
default.rgw.control 2 0 B 0
10 TiB 8
default.rgw.meta 3 3.5 KiB 0
10 TiB 19
default.rgw.log 4 8.4 KiB 0
10 TiB 1500
default.rgw.buckets.index 5 0 B 0
10 TiB 889
default.rgw.buckets.non-ec 6 0 B 0
10 TiB 497
default.rgw.buckets.data 7 14 TiB 56.96
10 TiB 3968545
testpool 8 0 B 0
10 TiB 0
'ceph df ' on new cluster:
RAW STORAGE:
CLASS SIZE AVAIL USED RAW USED %RAW USED
hdd 137 TiB 98 TiB 38 TiB 38 TiB 28.02
TOTAL 137 TiB 98 TiB 38 TiB 38 TiB 28.02
POOLS:
POOL ID STORED OBJECTS
USED %USED MAX AVAIL
.rgw.root 1 6.4 KiB
21 3.8 MiB 0 26 TiB
shubei.rgw.control 13 0 B 8
0 B 0 26 TiB
shubei.rgw.meta 14 4.1 KiB 20
3.2 MiB 0 26 TiB
shubei.rgw.log 15 9.9 MiB 1.64k
47 MiB 0 26 TiB
default.rgw.meta 16 0 B 0
0 B 0 26 TiB
shubei.rgw.buckets.index 17 2.7 MiB 889
2.7 MiB 0 26 TiB
shubei.rgw.buckets.data 18 11 TiB 2.90M
33 TiB 29.37 26 TiB
'radosgw-admin sync status' on new cluster:
realm bde4bb56-fbca-4ef8-a979-935dbf109b78 (new-oriental)
zonegroup d25ae683-cdb8-4227-be45-ebaf0aed6050 (beijing)
zone 313c8244-fe4d-4d46-bf9b-0e33e46be041 (shubei)
metadata sync syncing
full sync: 0/64 shards
incremental sync: 64/64 shards
metadata is caught up with master
data sync source: f70a5eb9-d88d-42fd-ab4e-d300e97094de (oldzone)
syncing
full sync: 106/128 shards
full sync: 350 buckets to sync
incremental sync: 22/128 shards
data is behind on 115 shards
behind shards:
[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,23,24,25,26,27,28,29,30,32,35,37,38,39,40,41,42,43,44,45,46,47,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,96,97,98,99,100,101,102,103,104,105,107,108,109,110,111,112,113,114,116,118,119,120,121,122,123,124,125,126,127]
oldest incremental change not applied: 2020-05-11
10:46:41.0.60179s [80]
5 shards are recovering
recovering shards: [21,31,95,104,106]