Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Azure for Architects

You're reading from   Azure for Architects Create secure, scalable, high-availability applications on the cloud

Arrow left icon
Product type Paperback
Published in Jul 2020
Publisher Packt
ISBN-13 9781839215865
Length 698 pages
Edition 3rd Edition
Languages
Tools
Arrow right icon
Authors (3):
Arrow left icon
Jack Lee Jack Lee
Author Profile Icon Jack Lee
Jack Lee
Ritesh Modi Ritesh Modi
Author Profile Icon Ritesh Modi
Ritesh Modi
Rithin Skaria Rithin Skaria
Author Profile Icon Rithin Skaria
Rithin Skaria
Arrow right icon
View More author details
Toc

Table of Contents (21) Chapters Close

Preface 1. Getting started with Azure 2. Azure solution availability, scalability, and monitoring FREE CHAPTER 3. Design pattern – Networks, storage, messaging, and events 4. Automating architecture on Azure 5. Designing policies, locks, and tags for Azure deployments 6. Cost management for Azure solutions 7. Azure OLTP solutions 8. Architecting secure applications on Azure 9. Azure Big Data solutions 10. Serverless in Azure – Working with Azure Functions 11. Azure solutions using Azure Logic Apps, Event Grid, and Functions 12. Azure Big Data eventing solutions 13. Integrating Azure DevOps 14. Architecting Azure Kubernetes solutions 15. Cross-subscription deployments using ARM templates 16. ARM template modular design and implementation 17. Designing IoT solutions 18. Azure Synapse Analytics for architects 19. Architecting intelligent solutions Index

Migrating data from Azure Storage to Data Lake Storage Gen2

In this section, we will be migrating data from Azure Blob storage to another Azure container of the same Azure Blob storage instance, and we will also migrate data to an Azure Data Lake Gen2 instance using an Azure Data Factory pipeline. The following sections outline the steps that need to be taken to create such an end-to-end solution.

Preparing the source storage account

Before we can create Azure Data Factory pipelines and use them for migration, we need to create a new storage account, consisting of a number of containers, and upload the data files. In the real world, these files and the storage connection would already be prepared. The first step for creating a new Azure storage account is to create a new resource group or choose an existing resource group within an Azure subscription.

Provisioning a new resource group

Every resource in Azure is associated with a resource...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image