>>And it seems smartctl on our seagate
ST4000NM0034 drives do not give us
data on total bytes written or read
If it's a SAS device, it's not always obvious where to find this information.
You can use Seagate's openseachest toolset.
For any (SAS/SATA, HDD/SSD) device, the --deviceInfo will give you
some of the info you are looking for; e.g.
sudo ./openSeaChest_Info -d /dev/sg1 --deviceInfo | grep Total
Total Bytes Read (TB): 82.46
Total Bytes Written (TB): 311.56
On Tue, Feb 11, 2020 at 3:10 AM lists <lists(a)merit.unu.edu> wrote:
>
> Hi Joe and Mehmet!
>
> Thanks for your responses!
>
> The requested outputs at the end of the message.
>
> But to make my question more clear:
>
> What we are actually after, is not about CURRENT usage of our OSDs, but
> stats on total GBs written in the cluster, per OSD, and read/write ratio.
>
> With those numbers, we would be able to identify suitable replacement
> SSDs for our current HDDs, and select specifically for OUR typical use.
> (taking into account endurance, speed, price, etc, etc)
>
> And it seems smartctl on our seagate ST4000NM0034 drives do not give us
> data on total bytes written or read. (...or are we simply not looking in
> the right place..?)
>
> Requested outputs below:
>
> > root@node1:~# ceph osd df tree
> > ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME
> > -1 87.35376 - 87.3TiB 49.1TiB 38.2TiB 56.22 1.00 - root default
> > -2 29.11688 - 29.1TiB 16.4TiB 12.7TiB 56.23 1.00 - host
node1
> > 0 hdd 3.64000 1.00000 3.64TiB 2.01TiB 1.62TiB 55.34 0.98 137 osd.0
> > 1 hdd 3.64000 1.00000 3.64TiB 2.09TiB 1.54TiB 57.56 1.02 141 osd.1
> > 2 hdd 3.63689 1.00000 3.64TiB 1.92TiB 1.72TiB 52.79 0.94 128 osd.2
> > 3 hdd 3.64000 1.00000 3.64TiB 2.07TiB 1.57TiB 56.90 1.01 143 osd.3
> > 12 hdd 3.64000 1.00000 3.64TiB 2.15TiB 1.48TiB 59.18 1.05 138
osd.12
> > 13 hdd 3.64000 1.00000 3.64TiB 1.99TiB 1.64TiB 54.80 0.97 131
osd.13
> > 14 hdd 3.64000 1.00000 3.64TiB 1.93TiB 1.70TiB 53.13 0.94 127
osd.14
> > 15 hdd 3.64000 1.00000 3.64TiB 2.19TiB 1.45TiB 60.10 1.07 143
osd.15
> > -3 29.12000 - 29.1TiB 16.4TiB 12.7TiB 56.22 1.00 - host
node2
> > 4 hdd 3.64000 1.00000 3.64TiB 2.11TiB 1.53TiB 57.97 1.03 142 osd.4
> > 5 hdd 3.64000 1.00000 3.64TiB 1.97TiB 1.67TiB 54.11 0.96 134 osd.5
> > 6 hdd 3.64000 1.00000 3.64TiB 2.12TiB 1.51TiB 58.40 1.04 142 osd.6
> > 7 hdd 3.64000 1.00000 3.64TiB 1.97TiB 1.66TiB 54.28 0.97 128 osd.7
> > 16 hdd 3.64000 1.00000 3.64TiB 2.00TiB 1.64TiB 54.90 0.98 133
osd.16
> > 17 hdd 3.64000 1.00000 3.64TiB 2.33TiB 1.30TiB 64.14 1.14 153
osd.17
> > 18 hdd 3.64000 1.00000 3.64TiB 1.97TiB 1.67TiB 54.07 0.96 132
osd.18
> > 19 hdd 3.64000 1.00000 3.64TiB 1.89TiB 1.75TiB 51.93 0.92 124
osd.19
> > -4 29.11688 - 29.1TiB 16.4TiB 12.7TiB 56.22 1.00 - host
node3
> > 8 hdd 3.64000 1.00000 3.64TiB 1.79TiB 1.85TiB 49.24 0.88 123 osd.8
> > 9 hdd 3.64000 1.00000 3.64TiB 2.17TiB 1.47TiB 59.72 1.06 144 osd.9
> > 10 hdd 3.64000 1.00000 3.64TiB 2.40TiB 1.24TiB 65.88 1.17 157
osd.10
> > 11 hdd 3.64000 1.00000 3.64TiB 2.06TiB 1.58TiB 56.64 1.01 133
osd.11
> > 20 hdd 3.64000 1.00000 3.64TiB 2.19TiB 1.45TiB 60.23 1.07 148
osd.20
> > 21 hdd 3.64000 1.00000 3.64TiB 1.74TiB 1.90TiB 47.80 0.85 115
osd.21
> > 22 hdd 3.64000 1.00000 3.64TiB 2.05TiB 1.59TiB 56.27 1.00 138
osd.22
> > 23 hdd 3.63689 1.00000 3.64TiB 1.96TiB 1.67TiB 54.01 0.96 130
osd.23
> > TOTAL 87.3TiB 49.1TiB 38.2TiB 56.22
> > MIN/MAX VAR: 0.85/1.17 STDDEV: 4.08
> > root@node1:~# ceph osd status
> > +----+------+-------+-------+--------+---------+--------+---------+-----------+
> > | id | host | used | avail | wr ops | wr data | rd ops | rd data | state |
> > +----+------+-------+-------+--------+---------+--------+---------+-----------+
> > | 0 | node1 | 2061G | 1663G | 38 | 5168k | 3 | 1491k | exists,up
|
> > | 1 | node1 | 2143G | 1580G | 4 | 1092k | 9 | 2243k | exists,up
|
> > | 2 | node1 | 1965G | 1758G | 20 | 3643k | 5 | 1758k | exists,up
|
> > | 3 | node1 | 2119G | 1605G | 17 | 99.5k | 4 | 3904k | exists,up
|
> > | 4 | node2 | 2158G | 1565G | 12 | 527k | 1 | 2632k | exists,up
|
> > | 5 | node2 | 2014G | 1709G | 15 | 239k | 0 | 889k | exists,up
|
> > | 6 | node2 | 2174G | 1549G | 11 | 1677k | 5 | 1931k | exists,up
|
> > | 7 | node2 | 2021G | 1702G | 2 | 597k | 0 | 1638k | exists,up
|
> > | 8 | node3 | 1833G | 1890G | 4 | 564k | 4 | 5595k | exists,up
|
> > | 9 | node3 | 2223G | 1500G | 6 | 1124k | 10 | 4864k | exists,up
|
> > | 10 | node3 | 2453G | 1270G | 8 | 1257k | 3 | 1447k | exists,up
|
> > | 11 | node3 | 2109G | 1614G | 14 | 2889k | 3 | 1449k | exists,up
|
> > | 12 | node1 | 2204G | 1520G | 17 | 1596k | 4 | 1806k | exists,up
|
> > | 13 | node1 | 2040G | 1683G | 15 | 2526k | 0 | 819k | exists,up
|
> > | 14 | node1 | 1978G | 1745G | 11 | 1713k | 8 | 3489k | exists,up
|
> > | 15 | node1 | 2238G | 1485G | 25 | 5151k | 5 | 2715k | exists,up
|
> > | 16 | node2 | 2044G | 1679G | 2 | 43.3k | 1 | 3371k | exists,up
|
> > | 17 | node2 | 2388G | 1335G | 14 | 1736k | 9 | 5315k | exists,up
|
> > | 18 | node2 | 2013G | 1710G | 8 | 1907k | 2 | 2004k | exists,up
|
> > | 19 | node2 | 1934G | 1790G | 15 | 2115k | 4 | 3248k | exists,up
|
> > | 20 | node3 | 2243G | 1481G | 15 | 3292k | 1 | 1763k | exists,up
|
> > | 21 | node3 | 1780G | 1944G | 8 | 1636k | 0 | 86.4k | exists,up
|
> > | 22 | node3 | 2095G | 1628G | 23 | 5012k | 4 | 1654k | exists,up
|
> > | 23 | node3 | 2011G | 1712G | 9 | 1662k | 1 | 2457k | exists,up
|
> > +----+------+-------+-------+--------+---------+--------+---------+-----------+
>
> Thanks!
>
> MJ
> _______________________________________________
> ceph-users mailing list -- ceph-users(a)ceph.io
> To unsubscribe send an email to ceph-users-leave(a)ceph.io