Hi,
We are playing a little bit with OSD specs on a test cluster, and we ended up having nodes that match more than 1 OSD spec. (currently 4 or 5).
And there is something we did not get yet. Is there any order cephadm will apply the sepcs? Are the specs sorted in any way inside cephadm?
We understand that for a specific spec, cephadm will try to match nodes by host, label and then host_pattern. Our question is more at spec level, and the order cephadm will "loop" the specs.
I hope I was clear enough.
Thanks,
Luis Domingues
Proton AG
hi friends
i have deployed my first node in cluster. we dont have direct internet on
my server so i have to set proxy for that.i set it /etc/environment
/etc/profile but i get bellow error
2023-09-26 17:09:38,254 7f04058b4b80 DEBUG
--------------------------------------------------------------------------------
cephadm ['--image', 'quay.io/ceph/ceph:v17', 'pull']
2023-09-26 17:09:38,302 7f04058b4b80 INFO Pulling container image
quay.io/ceph/ceph:v17...
2023-09-26 17:09:42,083 7f04058b4b80 INFO Non-zero exit code 125 from
/usr/bin/podman pull quay.io/ceph/ceph:v17
2023-09-26 17:09:42,083 7f04058b4b80 INFO /usr/bin/podman: stderr Trying to
pull quay.io/ceph/ceph:v17...
2023-09-26 17:09:42,084 7f04058b4b80 INFO /usr/bin/podman: stderr
time="2023-09-26T17:09:38+03:30" level=warning msg="Failed, retrying in 1s
... (1/3). Error: initializing source docker://quay.io/ceph/ceph:v17:
pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp
34.228.154.221:443: connect: connection refused"
2023-09-26 17:09:42,084 7f04058b4b80 INFO /usr/bin/podman: stderr
time="2023-09-26T17:09:39+03:30" level=warning msg="Failed, retrying in 1s
... (2/3). Error: initializing source docker://quay.io/ceph/ceph:v17:
pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp
3.220.246.53:443: connect: connection refused"
2023-09-26 17:09:42,084 7f04058b4b80 INFO /usr/bin/podman: stderr
time="2023-09-26T17:09:40+03:30" level=warning msg="Failed, retrying in 1s
... (3/3). Error: initializing source docker://quay.io/ceph/ceph:v17:
pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp
18.213.60.205:443: connect: connection refused"
2023-09-26 17:09:42,084 7f04058b4b80 INFO /usr/bin/podman: stderr Error:
initializing source docker://quay.io/ceph/ceph:v17: pinging container
registry quay.io: Get "https://quay.io/v2/": dial tcp 34.231.182.47:443:
connect: connection refused
2023-09-26 17:09:42,084 7f04058b4b80 ERROR ERROR: Failed command:
/usr/bin/podman pull quay.io/ceph/ceph:v17
would you please help.
thanks,
Hello Folks
We've recently performed an upgrade on our Cephadm cluster, transitioning
from Ceph Quiency to Reef. However, following the manual implementation of
a read balancer in the Reef cluster, we've experienced a significant
slowdown in client I/O operations within the Ceph cluster, affecting both
client bandwidth and overall cluster performance.
This slowdown has resulted in unresponsiveness across all virtual machines
within the cluster, despite the fact that the cluster exclusively utilizes
SSD storage."
Kindly guide us to move forward.
Regards
Mosharaf Hossain
We have a storage node that is failing, but the disks themselves are not. What is the recommended procedure for replacing the host itself without destroying the OSDs or losing data?
This cluster is running ceph 16.2.11 using ceph orchestrator with docker containers on Ubuntu 20.04 (focal).
thank you,
Wyllys Ingersoll
Hello,
First, sorry for my english...
Since a few weeks, I receive every day notifies with HEALTH ERR in my ceph. The notifies are related to inconssistent pgs and ever are on same osd.
I ran smartctl test to the disk osd assigned and the result is "passed".
Should replace the disk by other new?
Regards!
Hey ceph users!
I'd like to compile the different lib-xyz from ceph (rados, rbd,
cephfs) with ENABLE_SHARED=OFF (I want static libraries).
Since few days I am struggling building the whole ceph repo on debian
12.
Is there any way to build only the libraries? I dont need ceph, but only
the client side libraries.
Regards,
Arnaud.
Hey all,
Since the upgrade to Ceph 16.2.14, I keep seeing the following warning:
10 client(s) laggy due to laggy OSDs
ceph health detail shows it as:
[WRN] MDS_CLIENTS_LAGGY: 10 client(s) laggy due to laggy OSDs
mds.***(mds.3): Client *** is laggy; not evicted because some
OSD(s) is/are laggy
more of this...
When I restart the client(s) or the affected MDS daemons, the message
goes away and then comes back after a while. ceph osd perf does not list
any laggy OSDs (a few with 10-60ms ping, but overwhelmingly < 1ms), so
I'm on a total loss what this even means.
I have never seen this message before nor was I able to find anything
about it. Do you have any idea what this message actually means and how
I can get rid of it?
Thanks
Janek
Hey all,
We will be having a Ceph science/research/big cluster call on Wednesday
September 27th. If anyone wants to discuss something specific they can
add it to the pad linked below. If you have questions or comments you
can contact me.
This is an informal open call of community members mostly from
hpc/htc/research environments where we discuss whatever is on our minds
regarding ceph. Updates, outages, features, maintenance, etc...there is
no set presenter but I do attempt to keep the conversation lively.
NOTE: The change to using Jitsi for the meeting. We are no longer using
the bluejeans meeting links. The ceph calendar event does not yet
reflect this and has the wrong day as well.
Pad URL:
https://pad.ceph.com/p/Ceph_Science_User_Group_20230927
Virtual event details:
September 27th, 2023
14:00 UTC
4pm Central European
9am Central US
Description: Main pad for discussions:
https://pad.ceph.com/p/Ceph_Science_User_Group_Index
Meetings will be recorded and posted to the Ceph Youtube channel.
To join the meeting on a computer or mobile phone:
https://meet.jit.si/ceph-science-wg
Kevin
--
Kevin Hrpcek
NASA VIIRS Atmosphere SIPS/TROPICS
Space Science & Engineering Center
University of Wisconsin-Madison
Hi,
while I don't currently use rgw, I still am curious about consistency
guarantees.
Usually, S3 has strong read-after-write consistency guarantees (for
requests that do not overlap). According to
https://docs.ceph.com/en/latest/dev/radosgw/bucket_index/
in Ceph this is also true for per-object ACLs.
Is there also a strong consistency guarantee for (bucket) policies? The
documentation at
https://docs.ceph.com/en/latest/radosgw/bucketpolicy/
apparently does not say anything about this.
How would multiple rgw instances synchronize a policy change? Is this
effective immediate with strong consistency or is there some propagation
delay (hopefully on with some upper bound)?
Best regards
Matthias