Hello,
After two months of the "ceph try and error game", I finally managed to get an Octopuss cluster up and running.
The unconventional thing about it is, it's just for hot backups, no virtual machines on there.
All the nodes are without any caching ssd's, just plain hdd's.
At the moment there are eight of them with a total of 50TB. We are planning to go up to 25 and bigger disks so we end on 300TB-400TB
I decided to go with cephfs, because I don't have any experience in things like S3 and I need to read the same file system from more than one client.
I made one cephfs with a replicated pool.
On there I added erasure-coded pools to save some Storage.
To add those pools, I did it with the setfattr command like this:
setfattr -n ceph.dir.layout.pool -v ec_data_server1 /cephfs/nfs/server1
Some of our servers cannot use cephfs (old kernels, special OS's) so I have to use nfs.
This is set up with the included ganesha-nfs.
Exported is the /cephfs/nfs folder and clients can mount folders below this.
There are two final questions:
- Was it right to go with the way of "mounting" pools with setfattr, or should I have used multiple cephfs?
First I was thinking about using multiple cephfs but there are warnings everywhere. The deeper I got in, the more it seems I would have been fine with multiple cephfs.
- Is there a way I don't know, but it would be easier?
I still don't know much about Rest, S3, RBD etc... so there may be a better way
Other remarks are desired.
Thanks in advance,
Simon
Hello,
I am new to CEPH and on a few test servers attempting to setup and
learn a test ceph system.
I started off the install with the "Cephadm" option and it uses podman
containers.
Followed steps here:
https://docs.ceph.com/docs/master/cephadm/install/
I ran the bootstrap, added remote hosts, added monitors and everything
is looking good.
Now I would like to add OSDs...
On the bootstrapped server i did a :
ceph-volume lvm prepare --data /dev/sda6
and then the "activate" and "ceph orch daemon add osd (etc)" to add
it and it works...
But now I am ready to add OSDs on the remote nodes. I am not able to
find documentation or examples on how to do :
ceph-volume lvm prepare & activate steps on the remote hosts.
How do we prepare & activate the remote hosts disks?
Thank you very much for your input,
Cheers
Steve
Hello,
I have a ceph cluster (nautilus 14.2.8) with 2 filesystems and 3 mds.
mds1 is managing fs1
mds2 manages fs2
mds3 is standby
I want to completely remove fs1.
It seems that the command to use is ceph fs rm fs1 --yes-i-really-mean-it
and then delete the data and metadata pools with ceph osd pool delete
but in many threads I noticed that you must shutdown the mds before
running ceph fs rm.
Is it still the case ?
What happens in my configuration (I have 2 fs) ? If I stop mds1, the
mds3 will take the management. If I stop mds3 what will mds2 do (try to
manage the 2 fs or continue only with fs2) ?
Thanks for your advices.
F.
Hi,
I am currently facing the problem that our Ceph Cluster running Nautilus
is only listening on msgrv2 and we are not sure why.
This stops us from using block devices via rbd or mounting ceph via the
kernel module.
Attached[0] you can find the output of 'cat /etc/ceph/ceph.conf', 'ceph
mon dump' and 'ceph config dump'.
I already asked on IRC and was told that I probably have more success on
the mailing list so hopefully someone here also encountered that issue
and can help us out.
Kind regards,
Julian Fölsch
--
Julian Fölsch
Arbeitsgemeinschaft Dresdner Studentennetz (AG DSN)
Stellvertretender Schatzmeister
Telefon: +49 351 271816 69
Mobil: +49 152 22915871
Fax: +49 351 46469685
Email: julian.foelsch(a)agdsn.de
Studierendenrat der TU Dresden
Helmholtzstr. 10
01069 Dresden
Hi,
a lot of our OSD have crashed a few hours ago because of a failed assertion:
/build/ceph-15.2.3/src/osd/ECUtil.h: 34: FAILED ceph_assert(stripe_width
% stripe_size == 0)
Full output here:
https://pastebin.com/D1SXzKsK
All OSDs are on bluestore and run 15.2.3.
I think I messed up when I tried to change an existing EC profile (using
--force) for an active EC pool.
I already tried to delete the pool and the EC profile and start the OSDs
but they keep crashing with the same assertion.
Is there a way to at least find out what the values are for stripe_width
and stripe_size?
Regards,
Michael
Hi,
I'm new to radosgw (learned more about the MDS than I care to...), and it
seems like the buckets and objects created by one user cannot be accessed
by another user.
Is there a way to make any content created by User A accessible (read-only)
by User B?
From the documentation it looks like this is handled as an S3 permission
but I'm not finding an easy/obvious way to do this.
Any help would be appreciated. Thanks in advance!
Hi,
I've installed my ceph cluster with ceph-ansible a few months ago. I've
just added one monitor and one rgw at that time.
So I have 3 nodes, from which one is monitor and rgw and two others only
OSD.
Now I want to add the other two nodes as monitor and rgw.
Can I just modify the ansible host file and re-run the site.yml?
I've done some modification in Storage classes, I've added some OSD and
uploaded a lot of data up to now. Is it safe to re-run ansible site.yml
playbook?
I don't want to end with a fresh new cluster! :D
Thanks a lot,
Khodayar
I have 3 ceph clusters on nautilus 14.2.9 (same configuration through
puppet). 2 of them are autmatically sharding rados buckets, one of them is
not
When I do
radosgw-admin reshard stale-instances list on the cluster where it does
not work I get:
reshard stale-instances list
Resharding disabled in a multisite env, stale instances unlikely from
resharding
These instances may not be safe to delete.
Use --yes-i-really-mean-it to force displaying these instances.
The other 2 clusters don't give this warning. They are all single site.
The out put of realm list, zonegroup list and zone list for the cluster
that fails auto sharding are as follows
realm list
{
"default_info": "e724bd71-31eb-45c8-a456-151f6a5aa8b5",
"realms": [
"backup"
]
}
zonegroup list
{
"default_info": "ce4329ae-2bc8-4117-9b82-271022b223fa",
"zonegroups": [
"dc3"
]
}
zone list
{
"default_info": "7f9bebd6-a9cf-4006-83b1-ff99391aacc0",
"zones": [
"dc3-r1"
]
}
mgr configuration (default value)
rgw_dynamic_resharding true
How does ceph determine whether it is single site or multi site? How can I
force the automated bucket sharding??
Any help would be much appreciated
Marcel