Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Essential Guide to LLMOps

You're reading from   Essential Guide to LLMOps Implementing effective strategies for Large Language Models in deployment and continuous improvement

Arrow left icon
Product type Paperback
Published in Jul 2024
Publisher Packt
ISBN-13 9781835887509
Length 190 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
Ryan Doan Ryan Doan
Author Profile Icon Ryan Doan
Ryan Doan
Arrow right icon
View More author details
Toc

Table of Contents (14) Chapters Close

Preface 1. Part 1: Foundations of LLMOps
2. Chapter 1: Introduction to LLMs and LLMOps FREE CHAPTER 3. Chapter 2: Reviewing LLMOps Components 4. Part 2: Tools and Strategies in LLMOps
5. Chapter 3: Processing Data in LLMOps Tools 6. Chapter 4: Developing Models via LLMOps 7. Chapter 5: LLMOps Review and Compliance 8. Part 3: Advanced LLMOps Applications and Future Outlook
9. Chapter 6: LLMOps Strategies for Inference, Serving, and Scalability 10. Chapter 7: LLMOps Monitoring and Continuous Improvement 11. Chapter 8: The Future of LLMOps and Emerging Technologies 12. Index 13. Other Books You May Enjoy

Optimizing model serving for performance

The effective deployment of LLMs in production environments demands meticulous attention to the architecture, performance tuning, and emergency procedures of the serving infrastructure. This section covers the nuances of optimizing model serving for performance, ensuring that LLM applications are fit to serve inferences in a performant way.

Let’s review the types of model deployment to understand their serving performance implications better.

Comparing serverless, containerized, and microservices architectures

Serverless architecture, by design, removes the need for developers to manage server infrastructure, focusing instead on code development. This model, which adjusts computing resources based on incoming request volume, is particularly cost-effective for applications with variable demand, aligning well with the sporadic usage patterns often seen with LLMs.

Before deployment, LLMs often undergo a process of model optimization...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image