mastodon.ie is one of the many independent Mastodon servers you can use to participate in the fediverse.
Irish Mastodon - run from Ireland, we welcome all who respect the community rules and members.

Administered by:

Server stats:

1.6K
active users

#ceph

6 posts6 participants0 posts today
Michael<p>Putting the data roots for each buildah container into a separate directory on the shared CephFS volume did work and reduced the overall build time for the FluentD container from 23 minutes to 15 minutes. So the shared dir was part of the problem. But it's still rather slow, and I'm still pretty sure it's because of the VFS driver and CephFS. So let's see whether I can come up with something better.</p><p><a href="https://social.mei-home.net/tags/HomeLab" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HomeLab</span></a> <a href="https://social.mei-home.net/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a> <a href="https://social.mei-home.net/tags/buildah" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>buildah</span></a></p>
Heinlein Support<p>Du brauchst dringend viel <a href="https://social.heinlein-support.de/tags/Speicherplatz" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Speicherplatz</span></a>? Dann kommst Du im Open Source-Bereich an <a href="https://social.heinlein-support.de/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a> eigentlich nicht vorbei. Egal, ob als Daten-Backup, für Virtualisierungs-Projekte, als Fileserver oder als Ersatz für Amazon S3 – Ceph bietet als <a href="https://social.heinlein-support.de/tags/Storage" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Storage</span></a>-System vielfältige Einsatzszenarien. Und das Beste: bei guter Pflege und mit fundierter Schulung ist das System quasi unkaputtbar. </p><p>In unserer Schulung lernst Du, wie man ein Ceph-Cluster aufbaut und betreibt:<br><a href="https://www.heinlein-support.de/schulung/ceph-grundlagen" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">heinlein-support.de/schulung/c</span><span class="invisible">eph-grundlagen</span></a></p><p><a href="https://social.heinlein-support.de/tags/CephSchulung" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>CephSchulung</span></a></p>
JPL<p>Ich habe jetzt ca. eine Personenwoche damit verbracht, die Voraussetzungen zum Einsatz von Software X zu schaffen, nur um dann festzustellen dass X Mist ist und wir es vermutlich doch nicht einsetzen - was man auch aus der Doku hätte erahnen können. </p><p>Ich habe interessante Dinge dabei gelernt, z.B. über <a href="https://norden.social/tags/Kubernetes" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Kubernetes</span></a> auf Multi-Homed Bare-Metal Nodes. Und über <a href="https://norden.social/tags/Rook" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Rook</span></a> <a href="https://norden.social/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a>. Ich sollte mich also wahrscheinlich freuen, bin aber dennoch gerade sehr enttäuscht.</p><p><a href="https://norden.social/tags/SysAdmin" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>SysAdmin</span></a></p>
Neustradamus :xmpp: :linux:<p><a href="https://mastodon.social/tags/ProxmoxVE" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ProxmoxVE</span></a> 9.0 has been released (<a href="https://mastodon.social/tags/Proxmox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Proxmox</span></a> / <a href="https://mastodon.social/tags/VirtualEnvironment" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>VirtualEnvironment</span></a> / <a href="https://mastodon.social/tags/Virtualization" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Virtualization</span></a> / <a href="https://mastodon.social/tags/VirtualMachine" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>VirtualMachine</span></a> / <a href="https://mastodon.social/tags/VM" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>VM</span></a> / <a href="https://mastodon.social/tags/Linux" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Linux</span></a> / <a href="https://mastodon.social/tags/Debian" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Debian</span></a> / <a href="https://mastodon.social/tags/Trixie" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Trixie</span></a> / <a href="https://mastodon.social/tags/DebianTrixie" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DebianTrixie</span></a> / <a href="https://mastodon.social/tags/QEMU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>QEMU</span></a> / <a href="https://mastodon.social/tags/LXC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>LXC</span></a> / <a href="https://mastodon.social/tags/KVM" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>KVM</span></a> / <a href="https://mastodon.social/tags/ZFS" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ZFS</span></a> / <a href="https://mastodon.social/tags/OpenZFS" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>OpenZFS</span></a> / <a href="https://mastodon.social/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a>) <a href="https://proxmox.com/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="">proxmox.com/</span><span class="invisible"></span></a></p>
Rachel<p><span>ok looking a bit closer I think I see how it works! But I have some thoughts and a decision/question...<br><br>Each ceph pool gets automatically split into some number of PGs, and each PG gets split up by the crush rules.<br><br>So the PGs of my rbd pool all get split up between three hosts, but </span><i>which</i><span> three is chosen such that the total data usage is spread evenly across nodes/disks based on the weight.<br><br>For some reason the big hdd ec pool only has 1 PG so far so that is just using 6 of the drives, as it adds PGs it should spread out to the 8 drives just fine<br><br>But now I am thinking:<br><br>Do I continue with OSD failure domain, or do I switch to 2+1 EC with 4 hosts for this pool?<br><br>Basically everyone suggests not using OSD failure domain, but the mgr/etc data is replicated on the SSDs and with 8 drives it could re-balance (it will be a LONG time till I fill this, or even get close to 50% used)<br><br>Meanwhile with 3+1 and node failure domain I have the same capacity.<br><br></span><a href="https://hachyderm.io/@willglynn" class="u-url mention" rel="nofollow noopener" target="_blank">@willglynn@hachyderm.io</a> any thoughts? It could be a while until I can add more nodes/disks so no suggestions of filling a rack with 20 more servers ;) <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Homelab" rel="nofollow noopener" target="_blank">#Homelab</a></p>
Rachel<p><span>Ok so I'm seeing some curious issues with this ceph cluster and either I have a config issue, or a core misunderstanding that I really should address before I build out further <br><br>I have 8 hdd disks across 4 nodes, and in rook they're dust assigned to a 4+2 EC pool. <br><br>I had thought it would spread the data over all 8 drives eventually <br><br>But in </span><code>ceph pg dump</code><span> I see two drives missing. <br><br>* maybe it just grabs the he first 6 that it sees and leaves those two alone? Would it swap in drives as a hot spare? <br>* Or maybe I had a hiccup and they didn't get assigned for some other reason?<br>* Maybe I should reset it and re-create with 5+3 EC? <br>* The two drives in question have basically no data usage but they're not errored out<br></span><a href="https://transitory.social/tags/Homelab" rel="nofollow noopener" target="_blank">#Homelab</a> <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Minilab" rel="nofollow noopener" target="_blank">#Minilab</a></p>
Rachel<p><span>All drives have been added physically, now to figure out how the heck I'm going to build the pools and then transfer data over<br><br>Current plan:<br>* I have 8 disks, so I'm thinking 4+2 EC pool with the HDDs only, ODD failure domain because I just don't have enough hosts<br>* I think I can update Ceph and tell it to add the new HDDs as blank OSDs pretty easily<br>* Then I can add a new pool ? Maybe I'm a little fuzzy on terms. Goal is to run CephFS on this pool. <br>* Then I'll create the base CephFS volumes. I'm thinking I'll create a new namespace and add the PVCs there first. I'll back things up via this namespace. <br>* Then I can create new CephFS vols/PVCs in each namespace that also needs access<br>* I should be able to do some of the basic copy operations by just more bring NFS and CephFS at the same time in a pod? Maybe. That or I expose CephFS outside of the cluster and mount it directly on the NAS itself. <br><br></span><a href="https://transitory.social/tags/Homelab" rel="nofollow noopener" target="_blank">#Homelab</a> <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Minilab" rel="nofollow noopener" target="_blank">#Minilab</a></p>
Métaphysicien Douteux<p>Hey les copains <a href="https://framapiaf.org/tags/sysadmin" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>sysadmin</span></a> grands sorciers de <a href="https://framapiaf.org/tags/proxmox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>proxmox</span></a> et <a href="https://framapiaf.org/tags/ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ceph</span></a> question réseaux svp:<br>- 2 interfaces 10gb en bond0<br>- bond0 bridgé en vmbr0 (tolerance de panne + répartir 20gb sur vlans (vmbr0.x) (internes ceph et utilisés par VMs) </p><p>q1: MTU, je veux jumbo sur le vlan ceph et normal sur le reste. Je mets:<br>- 9000 sur le bond<br>- un peu moins sur vmbr0<br>- ce que je veux sur les differents vmbr0.x mais max &lt;vmbr0</p><p>q2: comment gérer les limites de bande passante par vlan?<br><span class="h-card" translate="no"><a href="https://friend.geoffray-levasseur.org/profile/fatalerrors" class="u-url mention" rel="nofollow noopener" target="_blank">@<span>fatalerrors</span></a></span> obviously ;)</p>
Matthew Vernon<p>Anyone got good suggestions for <a href="https://wikis.world/tags/Backups" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Backups</span></a> of <a href="https://wikis.world/tags/S3" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>S3</span></a> / <a href="https://wikis.world/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a> buckets?<br>I'd like not to have to download everything every time, and also have multiple backups but not growing without bound (and be able to say "restore object X from date Y").<br>I'd like to avoid "fuse-mount the bucket and then backup as if it were a fs".</p><p>Something like rsbackup (which uses rsync --link-dest meaning you only store changed objects; and that you can safely delete old backups) would be nice (rclone lacks this sort of thing)</p>
Rachel<p><span>Cluster rebuild project:<br><br>This isn't yak shaving, this is something else entirely <br><br>The goal: move all bulk storage off of the old NAS and onto a 3.5in HDD Ceph pool in the Minilab (plan is 4 nodes 8 disks 4+2 EC pool with osd failure domain). Most of the bulk data will be in CephFS. I plan to add a samba container for access by windows for misc uses as needed.<br><br>What needs to be accomplished to get there:<br><br>* Data needs to transfer and performance tested from each application it will be utilizer<br>* The HDD ool needs to be installed and built<br>* Before that I need to be confident of backups, Velero is working but it is failing on CephFS specifically and I'm not sure why. I'm able to manually take CephFS volume snapshots? Uggggggggh.<br>* Great now the Ceph mgrs are xrashing. The only errors I see are OOMKills but the nodes aren't close to our of memory any I haven't found anything else in the logs <br><br>So now I'm troubleshooting a pop-up event before the unclear backup issue before I can even get started on migrating the data.<br><br>With trip planning happening I probably won't have much progress until late Aug at best. The goal is to have this done and tested including backups before mid Nov.<br><br></span><a href="https://transitory.social/tags/Homelab" rel="nofollow noopener" target="_blank">#Homelab</a> <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Minilab" rel="nofollow noopener" target="_blank">#Minilab</a></p>
ij<p>My 3-node <a href="https://nerdculture.de/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a> cluster is showing nice performance when running "bin/tootctl search deploy --only=accounts" during the Mastodon upgrade...</p>
Rachel<p><span>Overview of cluster specs:<br><br>Networking:<br>* Mikrotik RB5009Upr+S+in<br>* Mikrotik CRS310-5s-4s+in<br>* Mikrotik CRS310-8g+2s+in (rear mounted)<br>* Cable modem (pending move into rack)<br>* 1x raspberry pi 4 running dnsmasq for DHCP/DNS, with a second acting as a coldish spare<br><br>Compute:<br>* 1x Intel core ultra 235 system with Nvidia p4 and 32gb ram, general compute<br>* 4x odroid H4s ultra with 1x NVMe boot, 2x 800gb, ssd 2x 22tb HDD (pending)<br><br>Each node has 1x 2.5gb link to the rear CRS310 with room to LACP the odroids if I need to upgrade networking.<br><br>Software: The cluster nodes are all running bare metal Talos. Three odroids act as control plane+storage, the fourth acts as a storage+compute, and the Intel core ultra is pure compute. There are/will be multiple Ceph storage pools for different use cases. This was the smallest Ceph that I felt ok with. Ceph people really suggest larger clusters than this so we'll see how it goes. That is also why I'm stalled until I get backups working fully.<br><br><br></span><a href="https://transitory.social/tags/Homelab" rel="nofollow noopener" target="_blank">#Homelab</a> <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Minilab" rel="nofollow noopener" target="_blank">#Minilab</a></p>
Rachel<p><span>Physical test fit is good!<br><br>(The 3.5in drives are not populated yet) </span><a href="https://transitory.social/tags/Homelab" rel="nofollow noopener" target="_blank">#Homelab</a> <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Minilab" rel="nofollow noopener" target="_blank">#Minilab</a></p>
Michael<p>Well of course. My Ceph cluster would produce a scrub inconsistency error while I'm 400 km away from the Homelab. Let's see what this is about.</p><p><a href="https://social.mei-home.net/tags/HomeLab" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HomeLab</span></a> <a href="https://social.mei-home.net/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a></p>
Marianne Spiller<p>Upgrade des Prod-Clusters von <a href="https://konfigurationsmanufaktur.de/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a> Reef zu Squid ✅ Und das sogar noch **vor** dem 1. August! 🎉 Lief unproblematisch.</p><p><a href="https://konfigurationsmanufaktur.de/tags/Proxmox" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Proxmox</span></a> <a href="https://konfigurationsmanufaktur.de/tags/Ceph" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Ceph</span></a></p>
Rachel<p>The Velero backups are working except for the CephFS volumes ​:neocat_sob:​<span><br><br>I have no idea why it just throws timeout errors but I can create a manual snapshot which doesn't take that long? <br><br>I'm so close to getting past this step but I'm not exactly where to look next<br></span><a href="https://transitory.social/tags/Kubernetes" rel="nofollow noopener" target="_blank">#Kubernetes</a> <a href="https://transitory.social/tags/Velero" rel="nofollow noopener" target="_blank">#Velero</a> <a href="https://transitory.social/tags/Ceph" rel="nofollow noopener" target="_blank">#Ceph</a> <a href="https://transitory.social/tags/Backups" rel="nofollow noopener" target="_blank">#Backups</a></p>

Seems some of my disks have seen a few writes! `smartctl -a ${DEVICE}` stats for the #Ceph and virtual machine host cluster here.

The 2TB SSDs are OSDs in Ceph, the others are OS disks or local VM cache disks. Ceph is moaning about the disks in "oxygen" and "fluorine".

Numbers here are Total Drive Writes in GB.

```
$ for h in *; do for d in ${h}/tmp/*.txt; do lbas=$( grep -F Total_LBAs_Written ${d} | cut -c 88- ); if [ -n "${lbas}" ]; then dev=$( basename "${d}" .txt ); dev=/dev/${dev##*-}; sect=$( grep '^Sector Size' ${d} | cut -c 18-21 ); printf "%-20s %8s %6d %s\n" ${h} ${dev} $(( ( ${lbas} * ${sect} ) / 1024 / 1024 / 1024 )) "$( grep '^Device Model' ${d} )"; fi; done; done
beryllium.chost.lan /dev/sda 0 Device Model: WD Green 2.5 240GB
boron.chost.lan /dev/sdb 76370 Device Model: Samsung SSD 870 EVO 2TB
carbon.chost.lan /dev/sda 113892 Device Model: Samsung SSD 870 EVO 2TB
carbon.chost.lan /dev/sdb 157993 Device Model: Samsung SSD 860 EVO 2TB
fluorine.chost.lan /dev/sda 111939 Device Model: Samsung SSD 870 QVO 2TB
helium.chost.lan /dev/sda 100476 Device Model: Samsung SSD 870 QVO 2TB
hydrogen.chost.lan /dev/sda 184564 Device Model: Samsung SSD 860 EVO 2TB
hydrogen.chost.lan /dev/sdb 58602 Device Model: Samsung SSD 870 EVO 2TB
lithium.chost.lan /dev/sda 0 Device Model: WD Green 2.5 240GB
magnesium.chost.lan /dev/sda 0 Device Model: WD Green 2.5 240GB
neon.chost.lan /dev/sdb 146926 Device Model: Samsung SSD 860 EVO 2TB
nitrogen.chost.lan /dev/sda 99473 Device Model: Samsung SSD 870 EVO 2TB
oxygen.chost.lan /dev/sdb 108748 Device Model: Samsung SSD 870 QVO 2TB
sodium.chost.lan /dev/sda 0 Device Model: WD Green 2.5 240GB
```

Replied in thread

I'm a glutton for punishment this week… I'm poking the octopus again, moving to Ceph 18.

So far, we're doing the monitor daemons. A fun fact, they switched from LevelDB to RocksDB some time back and in Ceph 17, they dropped support for LevelDB.

I found this out when updating the first monitor node, it crashed. Refused to start. All I had was a cryptic:

ceph_abort_msg("MonitorDBStore: error initializing keyvaluedb back storage")

Not very helpful. The docs for Ceph 17 (which I didn't read as I was going from 16→18 direct, which the Ceph 18 docs says you _can_ do), merely state:

> LevelDB support has been removed. WITH_LEVELDB is no longer a supported build option. Users should migrate their monitors and OSDs to RocksDB before upgrading to Quincy.
(docs.ceph.com/en/latest/releas)

How? 0 suggestions as to the procedure.

What worked here, was to manually remove the monitor node, rename the /var/lib/ceph/mon/* directory, then re-add the monitor using the manual instructions.

docs.ceph.com/en/latest/rados/

docs.ceph.comQuincy — Ceph Documentation