Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases now! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Delphi High Performance

You're reading from   Delphi High Performance Master the art of concurrency, parallel programming, and memory management to build fast Delphi apps

Arrow left icon
Product type Paperback
Published in Jun 2023
Publisher Packt
ISBN-13 9781805125877
Length 452 pages
Edition 2nd Edition
Languages
Arrow right icon
Author (1):
Arrow left icon
Primož Gabrijelčič Primož Gabrijelčič
Author Profile Icon Primož Gabrijelčič
Primož Gabrijelčič
Arrow right icon
View More author details
Toc

Table of Contents (15) Chapters Close

Preface 1. Chapter 1: About Performance 2. Chapter 2: Profiling the Code FREE CHAPTER 3. Chapter 3: Fixing the Algorithm 4. Chapter 4: Don’t Reinvent, Reuse 5. Chapter 5: Fine-Tuning the Code 6. Chapter 6: Memory Management 7. Chapter 7: Getting Started with the Parallel World 8. Chapter 8: Working with Parallel Tools 9. Chapter 9: Exploring Parallel Practices 10. Chapter 10: More Parallel Patterns 11. Chapter 11: Using External Libraries 12. Chapter 12: Best Practices 13. Index 14. Other Books You May Enjoy

Memory allocation in a parallel world

We’ve seen how FastMM4 boosts reallocation speed. Let’s take a look at another optimization that helps a lot when you write multithreaded code—as we will in the next three chapters.

The life of a memory manager is simple when there is only one thread of execution inside a program. When the memory manager is dealing out the memory, it can be perfectly safe in the knowledge that nothing can interrupt it in this work.

When we deal with parallel processing, however, multiple paths of execution simultaneously execute the same program and work on the same data. (We call them threads, and I’ll explain them in the next chapter.) Because of that, life from the memory manager’s perspective suddenly becomes very dangerous.

For example, let’s assume that one thread wants some memory. The memory manager finds a free memory block on a free list and prepares to return it. At that moment, however, another thread...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime