Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Salesforce Data Architect Certification Guide

You're reading from   Salesforce Data Architect Certification Guide Comprehensive coverage of the Salesforce Data Architect exam content to help you pass on the first attempt

Arrow left icon
Product type Paperback
Published in Nov 2022
Publisher Packt
ISBN-13 9781801813556
Length 254 pages
Edition 1st Edition
Arrow right icon
Author (1):
Arrow left icon
Aaron Allport Aaron Allport
Author Profile Icon Aaron Allport
Aaron Allport
Arrow right icon
View More author details
Toc

Table of Contents (23) Chapters Close

Preface 1. Section 1: Salesforce Data Architect Theory
2. Chapter 1: Introducing the Salesforce Data Architect Journey FREE CHAPTER 3. Chapter 2: Data Modeling and Database Design 4. Chapter 3: Master Data Management 5. Chapter 4: Salesforce Data Management 6. Chapter 5: Data Governance 7. Chapter 6: Understanding Large Data Volumes 8. Chapter 7: Data Migration 9. Section 2: Salesforce Data Architect Design
10. Chapter 8: Accounts and Contacts 11. Chapter 9: Data APIs and Apex 12. Chapter 10: Tuning Performance 13. Chapter 11: Backup and Restore 14. Chapter 12: Territory Management 15. Section 3: Applying What We've Learned – Practice Questions and Revision Aids
16. Chapter 13: Practice Exam Questions 17. Chapter 14: Cheat Sheets 18. Chapter 15: Further Resources 19. Chapter 16: How to Take the Exam 20. Chapter 17: Answers to Practice Questions 21. Index 22. Other Books You May Enjoy

PK chunking to improve performance

PK chunking is designed as a mechanism to allow entire Salesforce table data to be extracted—for example, as part of a backup routine. PK chunking effectively adds record IDs as a WHERE clause parameter to query data from a Salesforce entity in batches.

In general, if an object in Salesforce has more than 10 million rows, you should use PK chunking when exporting its data. If you are finding that querying for data times out regularly, use PK chunking.

Given that PK chunking effectively separates one big query into separate queries by adding a WHERE clause and using a range of ordered IDs, the batch size can be set. This is defaulted to 100,000 (as in, 100,000 records will be returned by default for each batch) but can be as high as 250,000. Therefore, for a 10 million-row entity, a batch size of 250,000 would result in 40 data batches being returned.

In Chapter 7, Data Migration, we walked through a practical example of how PK chunking...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at €18.99/month. Cancel anytime