I did not have time to convert all drives to lvm yet, so I would like to
stick to the use of the partition until I have time to change
everything.
-----Original Message-----
Sent: 01 March 2020 18:17
Subject: Re: [ceph-users] Is it ok to add a luminous ceph-disk osd to
nautilus still?
So use ceph-volume. ??
The nautilus release notes explain why.
On Mar 1, 2020, at 9:02 AM, Marc Roos
<M.Roos(a)f1-outsourcing.eu>
wrote:
ceph-disk is not available in Nautilus.elease
why scrub first? It is a new disk not having any data yet. Scrubbing
is verifying pg's not?
I just created a vm on the ceph node where I want to add this osd. Did
a passthru of the disk and installed a few rpm's
with nodeps to get
the ceph-disk command.
-----Original Message-----
Sent: 01 March 2020 17:47
Subject: Re: [ceph-users] Is it ok to add a luminous ceph-disk osd to
nautilus still?
Ensure that it gets scrubbed at least once by Luminous first. But how
and why are you doing this ? Why not use Nautilus
binaries ?
>> On Mar 1, 2020, at 8:36 AM, Marc Roos <M.Roos(a)f1-outsourcing.eu>
> wrote:
>
>
> If I create and osd with luminous 12.0.3 binaries, can I just add it
> to an existing Nautilus cluster?
>
> I sort of did this already, just wondered if there are any drawbacks.
>
>
> [@test2 software]# ceph-disk prepare --bluestore --zap-disk /dev/sdb
> Creating new GPT entries.
> GPT data structures destroyed! You may now partition the disk using
> fdisk or other utilities.
> Creating new GPT entries.
> The operation has completed successfully.
> Setting name!
> partNum is 0
> REALLY setting name!
> The operation has completed successfully.
> Setting name!
> partNum is 1
> REALLY setting name!
> The operation has completed successfully.
> The operation has completed successfully.
> meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=6400
> blks
> = sectsz=512 attr=2, projid32bit=1
> = crc=1 finobt=0, sparse=0
> data = bsize=4096 blocks=25600,
imaxpct=25
=
sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0 ftype=1
log =internal log bsize=4096 blocks=864, version=2
= sectsz=512 sunit=0 blks,
lazy-count=1
realtime =none extsz=4096
blocks=0, rtextents=0
The operation has completed successfully.
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io To unsubscribe send an
email to ceph-users-leave(a)ceph.io