Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Ceph Cookbook

You're reading from   Ceph Cookbook Practical recipes to design, implement, operate, and manage Ceph storage systems

Arrow left icon
Product type Paperback
Published in Nov 2017
Publisher Packt
ISBN-13 9781788391061
Length 466 pages
Edition 2nd Edition
Languages
Tools
Arrow right icon
Authors (3):
Arrow left icon
Vikhyat Umrao Vikhyat Umrao
Author Profile Icon Vikhyat Umrao
Vikhyat Umrao
Michael Hackett Michael Hackett
Author Profile Icon Michael Hackett
Michael Hackett
Karan Singh Karan Singh
Author Profile Icon Karan Singh
Karan Singh
Arrow right icon
View More author details
Toc

Table of Contents (15) Chapters Close

Preface 1. Ceph – Introduction and Beyond 2. Working with Ceph Block Device FREE CHAPTER 3. Working with Ceph and OpenStack 4. Working with Ceph Object Storage 5. Working with Ceph Object Storage Multi-Site v2 6. Working with the Ceph Filesystem 7. Monitoring Ceph Clusters 8. Operating and Managing a Ceph Cluster 9. Ceph under the Hood 10. Production Planning and Performance Tuning for Ceph 11. The Virtual Storage Manager for Ceph 12. More on Ceph 13. An Introduction to Troubleshooting Ceph 14. Upgrading Your Ceph Cluster from Hammer to Jewel

Scaling up your Ceph cluster

At this point, we have a running Ceph cluster with one MON and three OSDs configured on ceph-node1. Now we will scale up the cluster by adding ceph-node2 and ceph-node3 as MON and OSD nodes.

How to do it…

A Ceph storage cluster requires at least one monitor to run. For high availability, a Ceph storage cluster relies on an odd number of monitors and more than one, for example, 3 or 5, to form a quorum. It uses the Paxos algorithm to maintain quorum majority. You will notice that your Ceph cluster is currently showing HEALTH_WARN; this is because we have not configured any OSDs other than ceph-node1. By default, the data in a Ceph cluster is replicated three times, that too on three different OSDs hosted on three different nodes.

Since we already have one monitor running on ceph-node1, let's create two more monitors for our Ceph cluster and configure OSDs on ceph-node2 and ceph-node3:

  1. Update the Ceph hosts ceph-node2 and ceph-node3 to /etc/ansible/hosts:
  1. Verify that Ansible can reach the Ceph hosts mentioned in /etc/ansible/hosts:
  1. Run Ansible playbook in order to scale up the Ceph cluster on ceph-node2 and ceph-node3:

Once playbook completes the ceph cluster scaleout job and plays the recap with failed=0, it means that the Ceph ansible has deployed more Ceph daemons in the cluster, as shown in the following screenshot.

You have three more OSD daemons and one more monitor daemon running in ceph-node2 and three more OSD daemons and one more monitor daemon running in ceph-node3. Now you have total nine OSD daemons and three monitor daemons running on three nodes:

  1. We were getting a too few PGs per OSD warning and because of that, we increased the default RBD pool PGs from 64 to 128. Check the status of your Ceph cluster; at this stage, your cluster is healthy. PGs - placement groups are covered in detail in Chapter 9, Ceph Under the Hood.
lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime