Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases now! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
In-Memory Analytics with Apache Arrow

You're reading from   In-Memory Analytics with Apache Arrow Accelerate data analytics for efficient processing of flat and hierarchical data structures

Arrow left icon
Product type Paperback
Published in Sep 2024
Publisher Packt
ISBN-13 9781835461228
Length 406 pages
Edition 2nd Edition
Languages
Arrow right icon
Author (1):
Arrow left icon
Matthew Topol Matthew Topol
Author Profile Icon Matthew Topol
Matthew Topol
Arrow right icon
View More author details
Toc

Table of Contents (18) Chapters Close

Preface 1. Part 1: Overview of What Arrow is, Its Capabilities, Benefits, and Goals FREE CHAPTER
2. Chapter 1: Getting Started with Apache Arrow 3. Chapter 2: Working with Key Arrow Specifications 4. Chapter 3: Format and Memory Handling 5. Part 2: Interoperability with Arrow: The Power of Open Standards
6. Chapter 4: Crossing the Language Barrier with the Arrow C Data API 7. Chapter 5: Acero: A Streaming Arrow Execution Engine 8. Chapter 6: Using the Arrow Datasets API 9. Chapter 7: Exploring Apache Arrow Flight RPC 10. Chapter 8: Understanding Arrow Database Connectivity (ADBC) 11. Chapter 9: Using Arrow with Machine Learning Workflows 12. Part 3: Real-World Examples, Use Cases, and Future Development
13. Chapter 10: Powered by Apache Arrow 14. Chapter 11: How to Leave Your Mark on Arrow 15. Chapter 12: Future Development and Plans 16. Index 17. Other Books You May Enjoy

Learning about memory cartography

One draw of distributed systems such as Apache Spark is the ability to process very large datasets quickly. Sometimes, the dataset is so large that it can’t even fit entirely in memory on a single machine! Having a distributed system that can break up the data into chunks and process them in parallel is then necessary since no individual machine would be able to load the whole dataset in memory at one time to operate on it. But what if you could process a huge, multiple-GB file while using almost no RAM at all? That’s where memory mapping comes in.

Let’s look to our NYC Taxi dataset once again for help with demonstrating this concept. Let’s download the yellow taxi data for January 2015. The file we get is named yellow_tripdata_2015-01.parquet and is approximately 168 MB in size. If we convert it to a CSV file, it is around 1.3 GB and perfect to use as an example. For brevity, we’ll use the Python Arrow library...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime