Ceph: Frequent Node Reboots

Submitted by gpmidi on Sun, 01/10/2021 - 15:19

A few of my Ceph nodes with 60-ish disks each were experiencing frequent reboots. Turns out kernel.nmi_watchdog was rebooting it due to disks holding it up under very high load. By turning it off via `echo "kernel.nmi_watchdog=0" > /etc/sysctl.d/99-watchdog.conf` the problem was solved. Although I suspect there are better ways to tune NMI Watchdog to fix this. I'm being lazy. 


Ceph: Networking Between Hosts gpmidi Sat, 01/09/2021 - 14:55

Hosts in my cluster all have two 10GbE links to a switch bonded with LACP. That gives them an effective link of 20GbE normally. But really, even 5Gbps would be enough for most hosts. The kernel in most boxes combined with drivers and such isn't going to be able to push 20GbE, much less 40GbE. So faster isn't worth it. 

Note: Don't bother with IPoIB. You're better off with 10GbE x2 via LACP. Although if it's a choice between 40Gbps IB and IPoIB or a couple of 1GbE lines, go IPoIB. 

Ceph: What Drive Sizes To Use

Submitted by gpmidi on Sat, 01/09/2021 - 14:49

Drive Counts

  • 43 2TB
  • 1 2.5TB
  • 23  3TB
  • 42 8TB (In Ceph)
  • 30 8TB (Not In Ceph)


Why these drives? They're the main data drives I've had and used for years. The only ones I'm going to remove soon are a subset of the 2TB drives with high spin times. Some of them have more than 8.5 years of spin time. I'll probably remove any disk with more than 6 years of spin time as a preventive measure. 


Ceph: OSD Journal/WAL (& Sometimes Database)

Submitted by gpmidi on Sat, 01/09/2021 - 14:41

Once GREAT option for improving OSD performance with spinning disk, esp slow disk, is to use a redundant array of SSDs for the Bluestore Journal. If you've got SSD space to spare you could even put the RocksDB on SSD too. But that needs a LOT more space. The Journal only needs a couple of GiB per OSD. RocksDB needs a LOT more


Ceph: osd_memory_target

Submitted by gpmidi on Sat, 01/09/2021 - 14:37

A couple of months back I changed the value of "osd_memory_target" for all of my OSDs from 4GiB to 1.5GiB. That change has stopped all RAM related issues on my cluster. While I suspect but can't prove a small performance drop, it's well worth it in my case. 


Two i6000 Tape Libraries gpmidi Wed, 01/06/2021 - 10:14

In a home no less ;)