- Who we are
- What we know
- What we've created
- Hints and Kinks
- Checking Corosync cluster membership
- Configuring radosgw to behave like Amazon S3
- Downgrading to DRBD 8.3
- Fencing in Libvirt/KVM virtualized cluster nodes
- Fencing in VMware virtualized Pacemaker nodes
- GFS2 in Pacemaker (Debian/Ubuntu)
- Interleaving in Pacemaker clones
- Maintenance in active Pacemaker clusters
- Managing cron jobs with Pacemaker
- Mandatory and advisory ordering in Pacemaker
- Migrating virtual machines from block-based storage to RADOS/Ceph
- Network connectivity check in Pacemaker
- OCFS2 in Pacemaker (Debian/Ubuntu)
- Solid-state drives and Ceph OSD journals
- Solve a DRBD split-brain in 4 steps
- Testing Pacemaker clusters
- Totem "Retransmit List" in Corosync
- Turning Ceph RBD Images into SAN Storage Devices
- Which OSD stores a specific RADOS object?
- Ceph Tutorial (LCA 2013)
- Ceph: The Storage Stack for OpenStack (OpenStack Israel 2013)
- Die eigene Cloud mit OpenStack Essex (German, LinuxTag 2012)
- Fencing (LCE 2011)
- GlusterFS in HA Clusters (LCEU 2012)
- GlusterFS und Ceph (German, CeBIT 2012)
- Hands-On With Ceph (LCEU 2012)
- High Availability Update (OpenStack Summit Fall 2012)
- High Availability in OpenStack (CloudOpen 2012)
- High Availability in OpenStack (OpenStack Conference Spring 2012)
- Highly Available Cloud: Pacemaker integration with OpenStack (OSCON 2012)
- Mit OpenStack zur eigenen Cloud (German, CLT 2012)
- Mit OpenStack zur eigenen Cloud (German, OSDC 2012)
- More Reliable, More Resilient, More Redundant (OpenStack Summit April 2013)
- MySQL HA Deep Dive (MySQL Conference 2012)
- MySQL High Availability Deep Dive (PLUK 2012)
- MySQL High Availability Sprint (PLUK 2011)
- OpenStack Essex im Praxistest (German, Linuxwochen Wien 2012)
- OpenStack High Availability Update (Grizzly and Havana)
- Roll Your Own Cloud (LCA 2011)
- Storage Replication in HPHA (LCA 2012)
- Zen of Pacemaker (LCA 2012)
- hastexo in 100 Seconds
- Technical documentation
- News releases
- Hints and Kinks
- What we do
- What we charge
- What others say
"DRBD" is a trademark or registered trademark of LINBIT Information Technologies GmbH. hastexo is not affiliated with the trademark owner.
"GlusterFS" is a trademark or registered trademark of Red Hat, Inc.
"Ceph" is a trademark or registered trademark of New Dream Network, LLC.
hastexo High Availability Expert Training
You know how to install DRBD, Corosync and Pacemaker and how to get them set up for everyday's business? That's great, because then you're ready take your Linux and High Availability knowledge to the next level.
hastexo's High Availability Expert training is your ideal opportunity to understand Linux high availability in a way you would never have considered possible. Additionally, we'll introduce you to cluster filesystems (GFS & OCFS2) and replicated filesystems (GlusterFS, Ceph) and explain what their pros and cons are.
Want to Host?
The attendance fee per seat is 30 hastexo Professional Service Credits. Early bird discounts are available!
Day 1 (Advanced Pacemaker Concepts)
After a brief refresher covering basic Pacemaker concepts, you will:
- Learn to optimize resource failover and system utilization with Migration Thresholds, Failure Timeouts, System Health Monitoring, Resource Sets, and other great cluster stack features
- Learn to leverage Pacemaker's most important maintenance feature for Heartbeat-to-Corosync migrations and rolling upgrades
- Learn about site-to-site failover with booth
Day 2 (Cluster Filesystems)
- Get fencing right: Learn about STONITH and node fencing in Pacemaker, and know about resource-level fencing in DRBD.
- Learn about the basics of Cluster Filesystems and how they relate to SAN storage and DRBD
- Understand how GFS2 and OCFS2 work in detail
- Learn how to install and set up GFS2 and OCFS2
- Run DRBD in Dual-Primary configurations with OCFS2, GFS2 and Pacemaker
Day 3 (Multi-Node Replicated Storage)
- Understand the fundamental concepts of Replicated Filesystems
- Understand how GlusterFS and Ceph work in detail
- Learn how to install and set up GlusterFS and Ceph
- How to deploy GlusterFS and Ceph as alternatives to DRBD or a conventional SAN
Frequently Asked Questions about this Class
- What do I get out of this class?
- At the end of this class, you will have a deep understanding of Pacemaker and the existing storage strategies. You will be able to name their differences and you will know what storage solution works best for a certain setup - be it a storage cloud, a virtualization cluster, a scaled-out database service, you name it.
- I am new to High Availability. Is this class for me?
- If you are a complete novice, then likely, it's not. You should have a good understanding of Linux high availability and clustering concepts, else you won't be able to make the most of this class. There are a number of other training offerings on the market for high-availability novices.
- I don't run DRBD. Is this class for me?
- By all means, yes. We do cover DRBD, but this class is by no means DRBD centric.
- I run servers on Red Hat Cluster. Is this class for me?
- If you are planning to migrate to Pacemaker in the future (and you should), and you have a good level of hands-on expertise with Red Hat Cluster, then definitely so.