Help us cross the exabyte mark by opting-in today! We can also learn that the average cluster size grows over timeĬeph telemetry is on an opt-in basis. One of the things telemetry helps us to understand is version adoption Since the telemetry service was launched. With almost 2,500 clusters reporting nearly 1 exabyte of total storageĬapacity, we can see a steady growth rate We are excited to share with you the latest statistics from our Ceph public We would greatly appreciate any assistance or insights you could provide on these matters. ![]() Furthermore, we are wondering how severe traffic package loss and latency must be to cause OSD process crashes, and how the monitoring system determines that an OSD is offline. ![]() We have heard about potential issues with rockdb 6 causing OSD process failures, and we are curious about how to check the rockdb version. Additionally, we are wondering if a 2% packet loss can significantly affect Ceph's performance and even cause the OSD process to fail sometimes. We suspect that this may be a network issue, but we are unsure of how Ceph detects such long latency. When running the 'ceph -s' command, we observed a slow OSD heartbeat on the back and front, with the longest latency being 2250.54ms. We also noticed a large number of UDP port 5405 packets and the 'corosync' process utilizing a significant amount of CPU. Upon checking, we found that the ping is around 0.1ms, and there is occasional 2% packet loss when using flood ping, but not consistently. We are encountering a slow OSD heartbeat and have not been able to identify any network traffic issues. We are experiencing with Ceph after deploying it by PVE with the network backed by a 10G Cisco switch with VPC feature on.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |