Manuel, find the output of ceph osd df tree command:
# ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME
-7 84.00099 - 85.1TiB 41.6TiB 43.6TiB 48.82 1.00 - root root
-5 12.00000 - 13.1TiB 5.81TiB 7.29TiB 44.38 0.91 - rack
rack1
-1 12.00000 - 13.1TiB 5.81TiB 7.29TiB 44.38 0.91 -
node cpn01
0 nvme 1.00000 1.00000 1.09TiB 496GiB 621GiB 44.40 0.91 143
osd.0
1 nvme 1.00000 1.00000 1.09TiB 489GiB 629GiB 43.72 0.90 141
osd.1
2 nvme 1.00000 1.00000 1.09TiB 537GiB 581GiB 48.03 0.98 155
osd.2
3 nvme 1.00000 1.00000 1.09TiB 474GiB 644GiB 42.40 0.87 137
osd.3
4 nvme 1.00000 1.00000 1.09TiB 532GiB 586GiB 47.57 0.97 153
osd.4
5 nvme 1.00000 1.00000 1.09TiB 476GiB 642GiB 42.60 0.87 137
osd.5
6 nvme 1.00000 1.00000 1.09TiB 467GiB 650GiB 41.82 0.86 135
osd.6
7 nvme 1.00000 1.00000 1.09TiB 544GiB 574GiB 48.65 1.00 157
osd.7
8 nvme 1.00000 1.00000 1.09TiB 482GiB 636GiB 43.12 0.88 139
osd.8
9 nvme 1.00000 1.00000 1.09TiB 458GiB 660GiB 40.96 0.84 133
osd.9
10 nvme 1.00000 1.00000 1.09TiB 514GiB 604GiB 45.97 0.94 148
osd.10
11 nvme 1.00000 1.00000 1.09TiB 484GiB 633GiB 43.34 0.89 140
osd.11
-6 12.00099 - 13.1TiB 6.02TiB 7.08TiB 45.98 0.94 - rack
rack2
-2 12.00099 - 13.1TiB 6.02TiB 7.08TiB 45.98 0.94 -
node cpn02
12 nvme 1.00000 1.00000 1.09TiB 499GiB 619GiB 44.61 0.91 144
osd.12
13 nvme 1.00000 1.00000 1.09TiB 561GiB 557GiB 50.19 1.03 162
osd.13
14 nvme 1.00000 1.00000 1.09TiB 577GiB 541GiB 51.60 1.06 167
osd.14
15 nvme 1.00000 1.00000 1.09TiB 546GiB 572GiB 48.84 1.00 158
osd.15
16 nvme 1.00000 1.00000 1.09TiB 537GiB 580GiB 48.07 0.98 155
osd.16
17 nvme 1.00000 1.00000 1.09TiB 508GiB 610GiB 45.41 0.93 147
osd.17
18 nvme 1.00000 1.00000 1.09TiB 490GiB 628GiB 43.86 0.90 142
osd.18
19 nvme 1.00000 1.00000 1.09TiB 534GiB 584GiB 47.76 0.98 155
osd.19
20 nvme 1.00000 1.00000 1.09TiB 467GiB 651GiB 41.80 0.86 134
osd.20
21 nvme 1.00000 1.00000 1.09TiB 447GiB 671GiB 40.01 0.82 129
osd.21
22 nvme 1.00099 1.00000 1.09TiB 561GiB 556GiB 50.21 1.03 162
osd.22
23 nvme 1.00000 1.00000 1.09TiB 441GiB 677GiB 39.45 0.81 127
osd.23
-15 12.00000 - 13.1TiB 5.92TiB 7.18TiB 45.20 0.93 - rack
rack3
-3 12.00000 - 13.1TiB 5.92TiB 7.18TiB 45.20 0.93 -
node cpn03
24 nvme 1.00000 1.00000 1.09TiB 500GiB 617GiB 44.77 0.92 145
osd.24
25 nvme 1.00000 1.00000 1.09TiB 462GiB 655GiB 41.37 0.85 133
osd.25
26 nvme 1.00000 1.00000 1.09TiB 446GiB 672GiB 39.88 0.82 129
osd.26
27 nvme 1.00000 1.00000 1.09TiB 565GiB 553GiB 50.54 1.04 162
osd.27
28 nvme 1.00000 1.00000 1.09TiB 513GiB 605GiB 45.89 0.94 148
osd.28
29 nvme 1.00000 1.00000 1.09TiB 554GiB 564GiB 49.55 1.01 160
osd.29
30 nvme 1.00000 1.00000 1.09TiB 527GiB 591GiB 47.12 0.97 152
osd.30
31 nvme 1.00000 1.00000 1.09TiB 484GiB 634GiB 43.31 0.89 140
osd.31
32 nvme 1.00000 1.00000 1.09TiB 505GiB 612GiB 45.21 0.93 146
osd.32
33 nvme 1.00000 1.00000 1.09TiB 551GiB 567GiB 49.28 1.01 159
osd.33
34 nvme 1.00000 1.00000 1.09TiB 498GiB 620GiB 44.52 0.91 143
osd.34
35 nvme 1.00000 1.00000 1.09TiB 457GiB 660GiB 40.93 0.84 132
osd.35
-16 12.00000 - 13.1TiB 6.00TiB 7.10TiB 45.77 0.94 - rack
rack4
-4 12.00000 - 13.1TiB 6.00TiB 7.10TiB 45.77 0.94 -
node cpn04
36 nvme 1.00000 1.00000 1.09TiB 540GiB 578GiB 48.29 0.99 156
osd.36
37 nvme 1.00000 1.00000 1.09TiB 517GiB 601GiB 46.25 0.95 149
osd.37
38 nvme 1.00000 1.00000 1.09TiB 519GiB 599GiB 46.42 0.95 149
osd.38
39 nvme 1.00000 1.00000 1.09TiB 457GiB 661GiB 40.85 0.84 132
osd.39
40 nvme 1.00000 1.00000 1.09TiB 527GiB 590GiB 47.17 0.97 152
osd.40
41 nvme 1.00000 1.00000 1.09TiB 537GiB 581GiB 48.01 0.98 155
osd.41
42 nvme 1.00000 1.00000 1.09TiB 522GiB 596GiB 46.68 0.96 151
osd.42
43 nvme 1.00000 1.00000 1.09TiB 459GiB 658GiB 41.09 0.84 132
osd.43
44 nvme 1.00000 1.00000 1.09TiB 550GiB 568GiB 49.17 1.01 158
osd.44
45 nvme 1.00000 1.00000 1.09TiB 570GiB 548GiB 51.00 1.04 164
osd.45
46 nvme 1.00000 1.00000 1.09TiB 451GiB 667GiB 40.32 0.83 130
osd.46
47 nvme 1.00000 1.00000 1.09TiB 492GiB 626GiB 44.03 0.90 142
osd.47
-20 12.00000 - 10.9TiB 5.77TiB 5.15TiB 52.84 1.08 - rack
rack5
-19 12.00000 - 10.9TiB 5.77TiB 5.15TiB 52.84 1.08 -
node cpn05
48 nvme 1.00000 1.00000 931GiB 551GiB 380GiB 59.19 1.21 159
osd.48
49 nvme 1.00000 1.00000 931GiB 469GiB 462GiB 50.39 1.03 136
osd.49
50 nvme 1.00000 1.00000 931GiB 548GiB 384GiB 58.83 1.20 158
osd.50
51 nvme 1.00000 1.00000 931GiB 380GiB 551GiB 40.83 0.84 109
osd.51
52 nvme 1.00000 1.00000 931GiB 487GiB 445GiB 52.24 1.07 141
osd.52
53 nvme 1.00000 1.00000 931GiB 503GiB 429GiB 53.98 1.11 146
osd.53
54 nvme 1.00000 1.00000 931GiB 479GiB 452GiB 51.47 1.05 139
osd.54
55 nvme 1.00000 1.00000 931GiB 522GiB 410GiB 55.99 1.15 150
osd.55
56 nvme 1.00000 1.00000 931GiB 571GiB 360GiB 61.31 1.26 165
osd.56
57 nvme 1.00000 1.00000 931GiB 405GiB 527GiB 43.46 0.89 117
osd.57
58 nvme 1.00000 1.00000 931GiB 456GiB 475GiB 48.97 1.00 132
osd.58
59 nvme 1.00000 1.00000 931GiB 535GiB 396GiB 57.45 1.18 154
osd.59
-23 12.00000 - 10.9TiB 5.98TiB 4.93TiB 54.79 1.12 - rack
rack6
-24 12.00000 - 10.9TiB 5.98TiB 4.93TiB 54.79 1.12 -
node cpn06
60 nvme 1.00000 1.00000 931GiB 499GiB 432GiB 53.61 1.10 144
osd.60
61 nvme 1.00000 1.00000 931GiB 447GiB 485GiB 47.94 0.98 129
osd.61
62 nvme 1.00000 1.00000 931GiB 505GiB 426GiB 54.24 1.11 146
osd.62
63 nvme 1.00000 1.00000 931GiB 563GiB 368GiB 60.47 1.24 162
osd.63
64 nvme 1.00000 1.00000 931GiB 605GiB 326GiB 65.01 1.33 175
osd.64
65 nvme 1.00000 1.00000 931GiB 476GiB 455GiB 51.15 1.05 138
osd.65
66 nvme 1.00000 1.00000 931GiB 461GiB 471GiB 49.44 1.01 133
osd.66
67 nvme 1.00000 1.00000 931GiB 483GiB 448GiB 51.86 1.06 140
osd.67
68 nvme 1.00000 1.00000 931GiB 521GiB 411GiB 55.92 1.15 151
osd.68
69 nvme 1.00000 1.00000 931GiB 481GiB 450GiB 51.69 1.06 139
osd.69
70 nvme 1.00000 1.00000 931GiB 506GiB 426GiB 54.29 1.11 146
osd.70
71 nvme 1.00000 1.00000 931GiB 576GiB 355GiB 61.87 1.27 166
osd.71
-27 12.00000 - 10.9TiB 6.06TiB 4.85TiB 55.56 1.14 - rack
rack7
-28 12.00000 - 10.9TiB 6.06TiB 4.85TiB 55.56 1.14 -
node cpn07
72 nvme 1.00000 1.00000 931GiB 554GiB 378GiB 59.43 1.22 160
osd.72
73 nvme 1.00000 1.00000 931GiB 443GiB 489GiB 47.52 0.97 128
osd.73
74 nvme 1.00000 1.00000 931GiB 450GiB 481GiB 48.33 0.99 130
osd.74
75 nvme 1.00000 1.00000 931GiB 529GiB 403GiB 56.77 1.16 153
osd.75
76 nvme 1.00000 1.00000 931GiB 489GiB 443GiB 52.48 1.08 141
osd.76
77 nvme 1.00000 1.00000 931GiB 571GiB 360GiB 61.32 1.26 165
osd.77
78 nvme 1.00000 1.00000 931GiB 474GiB 458GiB 50.87 1.04 137
osd.78
79 nvme 1.00000 1.00000 931GiB 536GiB 395GiB 57.58 1.18 155
osd.79
80 nvme 1.00000 1.00000 931GiB 492GiB 440GiB 52.79 1.08 142
osd.80
81 nvme 1.00000 1.00000 931GiB 511GiB 421GiB 54.84 1.12 148
osd.81
82 nvme 1.00000 1.00000 931GiB 563GiB 368GiB 60.48 1.24 162
osd.82
83 nvme 1.00000 1.00000 931GiB 599GiB 332GiB 64.32 1.32 173
osd.83
TOTAL 85.1TiB 41.6TiB 43.6TiB 48.82
MIN/MAX VAR: 0.81/1.33 STDDEV: 6.30
Is there any documentation or scripts regarding manually distribution
for rebalancing
disk with minimal I/O impact on clients io?
On Tue, Feb 4, 2020 at 10:25 AM EDH - Manuel Rios <mriosfer(a)easydatahost.com>
wrote:
With “ceph osd df tree” will be clear but right now I
can see that some
%USE osd between 44% and 65%.
Ceph osd df tree give also the balance at host level.
Do you have balancer enabled ?No “perfect” distribution cause that you
cant use the full space.
In our case we gain space manually rebalancing disk, that cause some
objects moves to other osd but you can so fast space available.
Regards
*De:* German Anders <yodasbunker(a)gmail.com>
*Enviado el:* martes, 4 de febrero de 2020 14:20
*Para:* EDH - Manuel Rios <mriosfer(a)easydatahost.com>
*CC:* ceph-users(a)ceph.com
*Asunto:* Re: [ceph-users] Doubt about AVAIL space on df
Hi Manuel,
Sure thing:
# ceph osd df
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS
0 nvme 1.00000 1.00000 1.09TiB 496GiB 622GiB 44.35 0.91 143
1 nvme 1.00000 1.00000 1.09TiB 488GiB 630GiB 43.63 0.89 141
2 nvme 1.00000 1.00000 1.09TiB 537GiB 581GiB 48.05 0.99 155
3 nvme 1.00000 1.00000 1.09TiB 473GiB 644GiB 42.36 0.87 137
4 nvme 1.00000 1.00000 1.09TiB 531GiB 587GiB 47.52 0.97 153
5 nvme 1.00000 1.00000 1.09TiB 476GiB 642GiB 42.55 0.87 137
6 nvme 1.00000 1.00000 1.09TiB 467GiB 651GiB 41.77 0.86 135
7 nvme 1.00000 1.00000 1.09TiB 543GiB 574GiB 48.61 1.00 157
8 nvme 1.00000 1.00000 1.09TiB 481GiB 636GiB 43.08 0.88 139
9 nvme 1.00000 1.00000 1.09TiB 457GiB 660GiB 40.92 0.84 133
10 nvme 1.00000 1.00000 1.09TiB 513GiB 604GiB 45.92 0.94 148
11 nvme 1.00000 1.00000 1.09TiB 484GiB 634GiB 43.29 0.89 140
12 nvme 1.00000 1.00000 1.09TiB 498GiB 620GiB 44.57 0.91 144
13 nvme 1.00000 1.00000 1.09TiB 560GiB 557GiB 50.13 1.03 162
14 nvme 1.00000 1.00000 1.09TiB 576GiB 542GiB 51.55 1.06 167
15 nvme 1.00000 1.00000 1.09TiB 545GiB 572GiB 48.78 1.00 158
16 nvme 1.00000 1.00000 1.09TiB 537GiB 581GiB 48.02 0.98 155
17 nvme 1.00000 1.00000 1.09TiB 507GiB 611GiB 45.36 0.93 147
18 nvme 1.00000 1.00000 1.09TiB 490GiB 628GiB 43.86 0.90 142
19 nvme 1.00000 1.00000 1.09TiB 533GiB 584GiB 47.72 0.98 155
20 nvme 1.00000 1.00000 1.09TiB 467GiB 651GiB 41.75 0.86 134
21 nvme 1.00000 1.00000 1.09TiB 447GiB 671GiB 39.97 0.82 129
22 nvme 1.00099 1.00000 1.09TiB 561GiB 557GiB 50.16 1.03 162
23 nvme 1.00000 1.00000 1.09TiB 441GiB 677GiB 39.46 0.81 127
24 nvme 1.00000 1.00000 1.09TiB 500GiB 618GiB 44.72 0.92 145
25 nvme 1.00000 1.00000 1.09TiB 462GiB 656GiB 41.30 0.85 133
26 nvme 1.00000 1.00000 1.09TiB 445GiB 672GiB 39.85 0.82 129
27 nvme 1.00000 1.00000 1.09TiB 564GiB 554GiB 50.45 1.03 162
28 nvme 1.00000 1.00000 1.09TiB 512GiB 605GiB 45.84 0.94 148
29 nvme 1.00000 1.00000 1.09TiB 553GiB 565GiB 49.49 1.01 160
30 nvme 1.00000 1.00000 1.09TiB 526GiB 592GiB 47.07 0.97 152
31 nvme 1.00000 1.00000 1.09TiB 484GiB 633GiB 43.34 0.89 140
32 nvme 1.00000 1.00000 1.09TiB 504GiB 613GiB 45.13 0.93 146
33 nvme 1.00000 1.00000 1.09TiB 550GiB 567GiB 49.23 1.01 159
34 nvme 1.00000 1.00000 1.09TiB 497GiB 620GiB 44.51 0.91 143
35 nvme 1.00000 1.00000 1.09TiB 457GiB 661GiB 40.88 0.84 132
36 nvme 1.00000 1.00000 1.09TiB 539GiB 578GiB 48.25 0.99 156
37 nvme 1.00000 1.00000 1.09TiB 516GiB 601GiB 46.19 0.95 149
38 nvme 1.00000 1.00000 1.09TiB 518GiB 600GiB 46.35 0.95 149
39 nvme 1.00000 1.00000 1.09TiB 456GiB 662GiB 40.81 0.84 132
40 nvme 1.00000 1.00000 1.09TiB 527GiB 591GiB 47.13 0.97 152
41 nvme 1.00000 1.00000 1.09TiB 536GiB 581GiB 47.98 0.98 155
42 nvme 1.00000 1.00000 1.09TiB 521GiB 597GiB 46.62 0.96 151
43 nvme 1.00000 1.00000 1.09TiB 459GiB 659GiB 41.05 0.84 132
44 nvme 1.00000 1.00000 1.09TiB 549GiB 569GiB 49.12 1.01 158
45 nvme 1.00000 1.00000 1.09TiB 569GiB 548GiB 50.95 1.04 164
46 nvme 1.00000 1.00000 1.09TiB 450GiB 668GiB 40.28 0.83 130
47 nvme 1.00000 1.00000 1.09TiB 491GiB 626GiB 43.97 0.90 142
48 nvme 1.00000 1.00000 931GiB 551GiB 381GiB 59.13 1.21 159
49 nvme 1.00000 1.00000 931GiB 469GiB 463GiB 50.34 1.03 136
50 nvme 1.00000 1.00000 931GiB 548GiB 384GiB 58.78 1.21 158
51 nvme 1.00000 1.00000 931GiB 380GiB 552GiB 40.79 0.84 109
52 nvme 1.00000 1.00000 931GiB 486GiB 445GiB 52.20 1.07 141
53 nvme 1.00000 1.00000 931GiB 502GiB 429GiB 53.93 1.11 146
54 nvme 1.00000 1.00000 931GiB 479GiB 452GiB 51.42 1.05 139
55 nvme 1.00000 1.00000 931GiB 521GiB 410GiB 55.93 1.15 150
56 nvme 1.00000 1.00000 931GiB 570GiB 361GiB 61.25 1.26 165
57 nvme 1.00000 1.00000 931GiB 404GiB 527GiB 43.43 0.89 117
58 nvme 1.00000 1.00000 931GiB 455GiB 476GiB 48.89 1.00 132
59 nvme 1.00000 1.00000 931GiB 535GiB 397GiB 57.39 1.18 154
60 nvme 1.00000 1.00000 931GiB 499GiB 433GiB 53.56 1.10 144
61 nvme 1.00000 1.00000 931GiB 446GiB 485GiB 47.92 0.98 129
62 nvme 1.00000 1.00000 931GiB 505GiB 427GiB 54.18 1.11 146
63 nvme 1.00000 1.00000 931GiB 563GiB 369GiB 60.39 1.24 162
64 nvme 1.00000 1.00000 931GiB 605GiB 326GiB 64.99 1.33 175
65 nvme 1.00000 1.00000 931GiB 476GiB 455GiB 51.10 1.05 138
66 nvme 1.00000 1.00000 931GiB 460GiB 471GiB 49.38 1.01 133
67 nvme 1.00000 1.00000 931GiB 483GiB 449GiB 51.82 1.06 140
68 nvme 1.00000 1.00000 931GiB 520GiB 411GiB 55.86 1.15 151
69 nvme 1.00000 1.00000 931GiB 481GiB 450GiB 51.64 1.06 139
70 nvme 1.00000 1.00000 931GiB 505GiB 426GiB 54.24 1.11 146
71 nvme 1.00000 1.00000 931GiB 576GiB 356GiB 61.81 1.27 166
72 nvme 1.00000 1.00000 931GiB 552GiB 379GiB 59.30 1.22 160
73 nvme 1.00000 1.00000 931GiB 442GiB 489GiB 47.47 0.97 128
74 nvme 1.00000 1.00000 931GiB 450GiB 482GiB 48.28 0.99 130
75 nvme 1.00000 1.00000 931GiB 529GiB 403GiB 56.77 1.16 153
76 nvme 1.00000 1.00000 931GiB 488GiB 443GiB 52.44 1.08 141
77 nvme 1.00000 1.00000 931GiB 570GiB 361GiB 61.25 1.26 165
78 nvme 1.00000 1.00000 931GiB 473GiB 458GiB 50.79 1.04 137
79 nvme 1.00000 1.00000 931GiB 536GiB 396GiB 57.54 1.18 155
80 nvme 1.00000 1.00000 931GiB 491GiB 440GiB 52.74 1.08 142
81 nvme 1.00000 1.00000 931GiB 510GiB 421GiB 54.78 1.12 148
82 nvme 1.00000 1.00000 931GiB 563GiB 369GiB 60.42 1.24 162
83 nvme 1.00000 1.00000 931GiB 599GiB 333GiB 64.28 1.32 173
TOTAL 85.1TiB 41.5TiB 43.6TiB 48.77
MIN/MAX VAR: 0.81/1.33 STDDEV: 6.30
Thanks in advance,
Best regards,
On Tue, Feb 4, 2020 at 10:15 AM EDH - Manuel Rios <
mriosfer(a)easydatahost.com> wrote:
Hi German,
Can you post , ceph osd df tree ?
Looks like your usage distribution is not perfect and that's why you got
less space than real.
Regards
-----Mensaje original-----
De: German Anders <yodasbunker(a)gmail.com>
Enviado el: martes, 4 de febrero de 2020 14:00
Para: ceph-users(a)ceph.com
Asunto: [ceph-users] Doubt about AVAIL space on df
Hello Everyone,
I would like to understand if this output is right:
*# ceph df*
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
85.1TiB 43.7TiB 41.4TiB 48.68
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
volumes 13 13.8TiB 64.21 7.68TiB 3620495
I only have (1) pool called 'volumes' which is using 13.8TiB (we have a
replica of 3) so it's actually using 41,4TiB and that would be the RAW
USED, at this point is fine, but, then it said on the GLOBAL section that
the AVAIL space is 43.7TiB and the %RAW USED is only 48.68%.
So if I use the 7.68TiB of MAX AVAIL and the pool goes up to 100% of
usage, that would not lead to the total space of the cluster, right? I mean
were are those 43.7TiB of AVAIL space?
I'm using Luminous 12.2.12 release.
Sorry if it's a silly question or if it has been answered before.
Thanks in advance,
Best regards,
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io To unsubscribe send an
email to ceph-users-leave(a)ceph.io