Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Learning Apache Cassandra

You're reading from   Learning Apache Cassandra Managing fault-tolerant, scalable data with high performance

Arrow left icon
Product type Paperback
Published in Apr 2017
Publisher
ISBN-13 9781787127296
Length 360 pages
Edition 2nd Edition
Languages
Concepts
Arrow right icon
Author (1):
Arrow left icon
Sandeep Yarabarla Sandeep Yarabarla
Author Profile Icon Sandeep Yarabarla
Sandeep Yarabarla
Arrow right icon
View More author details
Toc

Table of Contents (15) Chapters Close

Preface 1. Getting Up and Running with Cassandra FREE CHAPTER 2. The First Table 3. Organizing Related Data 4. Beyond Key-Value Lookup 5. Establishing Relationships 6. Denormalizing Data for Maximum Performance 7. Expanding Your Data Model 8. Collections, Tuples, and User-Defined Types 9. Aggregating Time-Series Data 10. How Cassandra Distributes Data 11. Cassandra Multi-Node Cluster 12. Application Development Using the Java Driver 13. Peeking under the Hood 14. Authentication and Authorization

How to handle big data

Now that we are convinced the relational model is not a good fit for big data, let's try to figure out ways to handle big data. These are the solutions that paved the way for various NoSQL databases:

  • Clustering: The data should be spread across different nodes in a cluster. The data should be replicated across multiple nodes in order to sustain node failures. This helps spread the data across the cluster, and different nodes contain different subsets of data. This improves performance and provides fault tolerance.
A node is an instance of database software running on a server. Multiple instances of the same database could be running on the same server.
  • Flexible schema: Schemas should be flexible unlike the relational model and should evolve with the data.
  • Relax consistency: We should embrace the concept of eventual consistency, which means data will eventually be propagated to all the nodes in the cluster (in case of replication). Eventual consistency allows data replication across nodes with minimum overhead. This allows for fast writes with the need for distributed locking.
  • Denormalization of data: Denormalize data to optimize queries. This has to be done at the cost of writing and maintaining multiple copies of the same data.
You have been reading a chapter from
Learning Apache Cassandra - Second Edition
Published in: Apr 2017
Publisher:
ISBN-13: 9781787127296
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at AU $24.99/month. Cancel anytime