Hello,
I was looking for an official announcement for Octopus release, as the
latest update (back in Q3/2019) on the subject said it was scheduled for
March 1st.
Any updates on that?
BR,
--
Alex Chalkias
*Product Manager*
alex.chalkias(a)canonical.com
+33 766599367
*Canonical | **Ubuntu*
Hello all,
I'm maintaining a small Nautilus 12 OSD cluster (36TB raw). My mon nodes have the mgr/mds collocated/stacked with the mon. Each are allocated 10gb of RAM.
During a recent single disk failure and corresponding recovery, I noticed my mgr/mon's were starting to get OOM killed/restarted every 5ish hours - the mgr using around 6.5GB on all my nodes. My monitoring shows an interesting sawtooth pattern with network usage (100MB/s at max), disk storage usage, and disk IO (up to 300MB/s against SSD's at max) usage increasing in parallel with memory usage.
I know the docs for hardware recommendations say:
> Monitor and manager daemon memory usage generally scales with the size of the cluster. For small clusters, 1-2 GB is generally sufficient. For large clusters, you should provide more (5-10 GB).
Now, I would like to think my cluster is on the small size of things, so I was hoping 10gb is enough for the mgr and mon (my OSD nodes are only allocated 32GB of ram), but that assumption appears to be false.
So I was wondering how mgr's (and to a lesser extent mon's) are expected to scale in terms of memory. Is it the osd count, or the osd's size, number of pg's, etc.? And if there's a way to limit the amount of RAM used by these mgr's (it seems the mon_osd_cache_size and rocksdb_cache_size settings are for mons if I'm not mistaken).
Regards,
Mark
I have some osd's having a bluefs label formatted like this:
{
"/dev/sdc2": {
"osd_uuid": "cfb2eaa3-1811-4108-b9bb-aad49555246c",
"size": 4000681099264,
"btime": "2017-07-14 14:58:09.627614",
"description": "main",
"require_osd_release": "14"
}
}
And I have some osd's having a bluefs label formatted like this:
{
"/dev/sdd2": {
"osd_uuid": "d8912a1b-696c-4668-9337-c740ec47e0d0",
"size": 8001457295360,
"btime": "2018-06-01 18:27:47.760695",
"description": "main",
"bluefs": "1",
"ceph_fsid": "0f1701f5-453a-4a3b-928d-f652a2bbbcb0",
"kv_backend": "rocksdb",
"magic": "ceph osd volume v026",
"mkfs_done": "yes",
"ready": "ready",
"require_osd_release": "14",
"whoami": "18"
}
}
I assume this is a version difference? What are the implications of
having this? How to upgrade?
Hi
We have a new mimic (13.2.6, will upgrade to nautilus next month) cluster
where the rados gateway pool has currently many more objects per PG then
the other pools. This leads to a warning with ceph status
1 pools have many more objects per pg than average
I tries to get rid of this warning by setting on the commandline
ceph config set mgr mon_pg_warn_max_object_skew 0
I recall that it leaded to HEALTH_OK but some time later it was at
HEALTH_WARN again with the same message. I tried setting the value to -1
as well, restarting mon and mgr daemons. Unfortunately the message stays.
Setting in ceph.conf does not help either
I know that in the future the number of objects per PG increase for the
other pools but in the meantime monitoring for another incoming health
warn issue is hard to automate so we aim to always have a HEALTH_OK state
Anyone any luck in getting rid of the warning by setting
mon_pg_warn_max_object_skew?
Thx
Marcel
Hey guys,
I'm trying to solve some Lost+Found errors, but when I try to run the
"scan_link" command, it crashes.
Any tip?
Cheers!
ceph cluster version 13.2.6 (7b695f835b03642f85998b2ae7b6dd093d9fbce4)
mimic (stable)
cephfs-data-scan version 14.2.7
(3d58626ebeec02d8385a4cefb92c6cbc3a45bfe8) nautilus (stable)
root@deployer:/etc/ceph# cephfs-data-scan scan_links
2020-03-02 01:52:42.876 7fda150fa700 -1 NetHandler create_socket
couldn't create socket (97) Address family not supported by protocol
2020-03-02 01:52:43.664 7fda23d26f40 -1 datascan.scan_links: Error
getting omap from '10000af78b3.00000000': (2) No such file or directory
Error ((2) No such file or directory)
root@deployer:/etc/ceph# cephfs-data-scan --version
ceph version 14.2.7 (3d58626ebeec02d8385a4cefb92c6cbc3a45bfe8) nautilus
(stable)
We have some inexplicable situation.
We have ceph cluster on 14.2.4.
About 14 nodes with 12 disks (4Tb) in each node.
But command ceph df returns for us next report:
# ceph df
RAW STORAGE:
CLASS SIZE AVAIL USED RAW USED %RAW USED
hdd 611 TiB 155 TiB 455 TiB 456 TiB 74.60
TOTAL 611 TiB 155 TiB 455 TiB 456 TiB 74.60
POOLS:
POOL ID STORED OBJECTS USED %USED MAX AVAIL
poolera01 20 238 TiB 84.59M 362 TiB 76.39 75 TiB
poolera01md 21 17 GiB 58.33k 17 GiB 0.01 37 TiB
poolera02dt 25 71 TiB 24.56M 92 TiB 45.02 90 TiB
poolera02md 26 749 MiB 42.23k 1.2 GiB 0 37 TiB
poolera01 - pool Erasure-coded 4+2
poolera02dt - pool Erasure-coded 8+2
poolera01md and poolera02md - both replicated size 3
We are using 2 cephfs.
It seems, that all our pools can only utilize about 111TiB raw capacity. But we have 155 TiB!
We have enabled balancer with mode upmap
# ceph balancer status
{
"active": true,
"plans": [],
"mode": "upmap"
}
Can someone explain why ceph df show less MAX AVAIL than possible from AVAIL 155 TiB?
I did not have time to convert all drives to lvm yet, so I would like to
stick to the use of the partition until I have time to change
everything.
-----Original Message-----
Sent: 01 March 2020 18:17
Subject: Re: [ceph-users] Is it ok to add a luminous ceph-disk osd to
nautilus still?
So use ceph-volume. ??
The nautilus release notes explain why.
> On Mar 1, 2020, at 9:02 AM, Marc Roos <M.Roos(a)f1-outsourcing.eu>
wrote:
>
>
> ceph-disk is not available in Nautilus.elease
>
> why scrub first? It is a new disk not having any data yet. Scrubbing
> is verifying pg's not?
>
> I just created a vm on the ceph node where I want to add this osd. Did
> a passthru of the disk and installed a few rpm's with nodeps to get
> the ceph-disk command.
>
>
>
> -----Original Message-----
> Sent: 01 March 2020 17:47
> Subject: Re: [ceph-users] Is it ok to add a luminous ceph-disk osd to
> nautilus still?
>
> Ensure that it gets scrubbed at least once by Luminous first. But how
> and why are you doing this ? Why not use Nautilus binaries ?
>
>>> On Mar 1, 2020, at 8:36 AM, Marc Roos <M.Roos(a)f1-outsourcing.eu>
>> wrote:
>>
>>
>> If I create and osd with luminous 12.0.3 binaries, can I just add it
>> to an existing Nautilus cluster?
>>
>> I sort of did this already, just wondered if there are any drawbacks.
>>
>>
>> [@test2 software]# ceph-disk prepare --bluestore --zap-disk /dev/sdb
>> Creating new GPT entries.
>> GPT data structures destroyed! You may now partition the disk using
>> fdisk or other utilities.
>> Creating new GPT entries.
>> The operation has completed successfully.
>> Setting name!
>> partNum is 0
>> REALLY setting name!
>> The operation has completed successfully.
>> Setting name!
>> partNum is 1
>> REALLY setting name!
>> The operation has completed successfully.
>> The operation has completed successfully.
>> meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=6400
>> blks
>> = sectsz=512 attr=2, projid32bit=1
>> = crc=1 finobt=0, sparse=0
>> data = bsize=4096 blocks=25600,
imaxpct=25
>> = sunit=0 swidth=0 blks
>> naming =version 2 bsize=4096 ascii-ci=0 ftype=1
>> log =internal log bsize=4096 blocks=864, version=2
>> = sectsz=512 sunit=0 blks,
> lazy-count=1
>> realtime =none extsz=4096 blocks=0, rtextents=0
>> The operation has completed successfully.
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users(a)ceph.io To unsubscribe send an
>> email to ceph-users-leave(a)ceph.io
>
>
ceph-disk is not available in Nautilus.
why scrub first? It is a new disk not having any data yet. Scrubbing is
verifying pg's not?
I just created a vm on the ceph node where I want to add this osd. Did a
passthru of the disk and installed a few rpm's with nodeps to get the
ceph-disk command.
-----Original Message-----
Sent: 01 March 2020 17:47
Subject: Re: [ceph-users] Is it ok to add a luminous ceph-disk osd to
nautilus still?
Ensure that it gets scrubbed at least once by Luminous first. But how
and why are you doing this ? Why not use Nautilus binaries ?
> On Mar 1, 2020, at 8:36 AM, Marc Roos <M.Roos(a)f1-outsourcing.eu>
wrote:
>
>
> If I create and osd with luminous 12.0.3 binaries, can I just add it
> to an existing Nautilus cluster?
>
> I sort of did this already, just wondered if there are any drawbacks.
>
>
> [@test2 software]# ceph-disk prepare --bluestore --zap-disk /dev/sdb
> Creating new GPT entries.
> GPT data structures destroyed! You may now partition the disk using
> fdisk or other utilities.
> Creating new GPT entries.
> The operation has completed successfully.
> Setting name!
> partNum is 0
> REALLY setting name!
> The operation has completed successfully.
> Setting name!
> partNum is 1
> REALLY setting name!
> The operation has completed successfully.
> The operation has completed successfully.
> meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=6400
> blks
> = sectsz=512 attr=2, projid32bit=1
> = crc=1 finobt=0, sparse=0
> data = bsize=4096 blocks=25600, imaxpct=25
> = sunit=0 swidth=0 blks
> naming =version 2 bsize=4096 ascii-ci=0 ftype=1
> log =internal log bsize=4096 blocks=864, version=2
> = sectsz=512 sunit=0 blks,
lazy-count=1
> realtime =none extsz=4096 blocks=0, rtextents=0
> The operation has completed successfully.
>
> _______________________________________________
> ceph-users mailing list -- ceph-users(a)ceph.io To unsubscribe send an
> email to ceph-users-leave(a)ceph.io
If I create and osd with luminous 12.0.3 binaries, can I just add it to
an existing Nautilus cluster?
I sort of did this already, just wondered if there are any drawbacks.
[@test2 software]# ceph-disk prepare --bluestore --zap-disk /dev/sdb
Creating new GPT entries.
GPT data structures destroyed! You may now partition the disk using
fdisk or
other utilities.
Creating new GPT entries.
The operation has completed successfully.
Setting name!
partNum is 0
REALLY setting name!
The operation has completed successfully.
Setting name!
partNum is 1
REALLY setting name!
The operation has completed successfully.
The operation has completed successfully.
meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=6400
blks
= sectsz=512 attr=2, projid32bit=1
= crc=1 finobt=0, sparse=0
data = bsize=4096 blocks=25600, imaxpct=25
= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0 ftype=1
log =internal log bsize=4096 blocks=864, version=2
= sectsz=512 sunit=0 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
The operation has completed successfully.