Call Us: +32 2 466 00 16
Email: info@u2u.be
Follow Us:

Implementing and Designing an Azure Data Solution

4 days
dp-200-201
4 days

Upcoming Sessions

Date: currently not scheduled

Format: Classroom

Price: 0€

Subscribe to waiting list

Date: currently not scheduled

Format: Classroom

Price: 0€

Subscribe to waiting list

Interested in a private company training? Request it here.

DP-200: Implementing an Azure Data Solution

Module 1: Azure for the Data Engineer

This module explores how the world of data has evolved and how cloud data platform technologies are providing new opportunities for business to explore their data in different ways. The student will gain an overview of the various data platform technologies that are available, and how a Data Engineers role and responsibilities has evolved to work in this new world to an organization benefit

  • Explain the evolving world of data
  • Survey the services in the Azure Data Platform
  • Identify the tasks that are performed by a Data Engineer
  • Describe the use cases for the cloud in a Case Study

Module 2: Working with Data Storage

This module teaches the variety of ways to store data in Azure. The Student will learn the basics of storage management in Azure, how to create a Storage Account, and how to choose the right model for the data you want to store in the cloud. They will also understand how data lake storage can be created to support a wide variety of big data analytics solutions with minimal effort.

  • Choose a data storage approach in Azure
  • Create an Azure Storage Account
  • Explain Azure Data Lake storage
  • Upload data into Azure Data Lake

Module 3: Enabling Team Based Data Science with Azure Databricks

This module introduces students to Azure Databricks and how a Data Engineer works with it to enable an organization to perform Team Data Science projects. They will learn the fundamentals of Azure Databricks and Apache Spark notebooks; how to provision the service and workspaces and learn how to perform data preparation task that can contribute to the data science project.

  • Explain Azure Databricks and Machine Learning Platforms
  • Describe the Team Data Science Process
  • Provision Azure Databricks and workspaces
  • Perform data preparation tasks

Module 4: Building Globally Distributed Databases with Cosmos DB

In this module, students will learn how to work with NoSQL data using Azure Cosmos DB. They will learn how to provision the service, and how they can load and interrogate data in the service using Visual Studio Code extensions, and the Azure Cosmos DB .NET Core SDK. They will also learn how to configure the availability options so that users are able to access the data from anywhere in the world.

  • Create an Azure Cosmos DB database built to scale
  • Insert and query data in your Azure Cosmos DB database
  • Provision a .NET Core app for Cosmos DB in Visual Studio Code
  • Distribute your data globally with Azure Cosmos DB

Module 5: Working with Relational Data Stores in the Cloud

In this module, students will explore the Azure relational data platform options including SQL Database and SQL Data Warehouse. The student will be able explain why they would choose one service over another, and how to provision, connect and manage each of the services.

  • SQL Database and SQL Data Warehouse
  • Provision an Azure SQL database to store data
  • Provision and load data into Azure SQL Data Warehouse

Module 6: Performing Real-Time Analytics with Stream Analytics

In this module, students will learn the concepts of event processing and streaming data and how this applies to Events Hubs and Azure Stream Analytics. The students will then set up a stream analytics job to stream data and learn how to query the incoming data to perform analysis of the data. Finally, you will learn how to manage and monitor running jobs.

  • Explain data streams and event processing
  • Querying streaming data using Stream Analytics
  • How to process data with Azure Blob and Stream Analytics
  • How to process data with Event Hubs and Stream Analytics

Module 7: Orchestrating Data Movement with Azure Data Factory

In this module, students will learn how Azure Data factory can be used to orchestrate the data movement and transformation from a wide range of data platform technologies. They will be able to explain the capabilities of the technology and be able to set up an end to end data pipeline that ingests and transforms data.

  • Explain how Azure Data Factory works
  • Create Linked Services and datasets
  • Create pipelines and activities
  • Azure Data Factory pipeline execution and triggers

Module 8: Securing Azure Data Platforms

In this module, students will learn how Azure Storage provides a multi-layered security model to protect your data. The students will explore how security can range from setting up secure networks and access keys, to defining permission through to monitoring with Advanced Threat Detection.

  • Configuring Network Security
  • Configuring Authentication
  • Configuring Authorization
  • Auditing Security

Module 9: Monitoring and Troubleshooting Data Storage and Processing

In this module, the student will look at the wide range of monitoring capabilities that are available to provide operational support should there be issue with a data platform architecture. They will explore the data engineering troubleshooting approach and be able to apply this to common data storage and data processing issues.

  • Data Engineering troubleshooting approach
  • Azure Monitoring Capabilities
  • Troubleshoot common data issues
  • Troubleshoot common data processing issues

Module 10: Integrating and Optimizing Data Platforms

In this module, the student will explore the various ways in which data platforms can be integrated based upon different business requirements. They will also explore the various ways in which data platforms can be optimized from a storage and data processing perspective to improve data loads. Finally, disaster recovery options are revealed to ensure business continuity.

  • Integrating data platforms
  • Optimizing data stores
  • Optimize streaming data
  • Manage disaster recovery

DP-201: Designing an Azure Data Solution

Module 1: Data Platform Architecture Considerations

In this module, the students will learn how to design and build secure, scalable, and performant solutions in Azure by examining the core principles found in every good architecture. They will learn how using key principles throughout architecture, regardless of technology choice, can help you design, build, and continuously improve the architecture for an organization's benefit.

  • Core Principles of Creating Architectures
  • Design with Security in Mind
  • Performance and Scalability
  • Design for availability and recoverability
  • Design for efficiency and operations
  • Case Study

Module 2: Azure Batch Processing Reference Architectures

In this module, students will learn the reference design and architecture patterns for dealing with the batch processing of data. The student will be exposed to dealing with the movement of data from on-premises systems into a cloud data warehouse and how it can be automated. The students will also be exposed to an AI architecture and how the data platform can integrate with an AI solution.

  • Lambda architectures from a Batch Mode Perspective
  • Design an Enterprise BI solution in Azure
  • Automate enterprise BI solutions in Azure
  • Architect an Enterprise-grade Conversational Bot in Azure

Module 3: Azure Real-Time Reference Architectures

In this module, the students will learn the reference design and architecture patterns for dealing with streaming data. They will learn how streaming data can be ingested by Event Hubs and Stream Analytics to deliver real-time analysis of data. They will also explore a data science architecture that streams data into Azure Databricks to perform trend analysis. They will finally learn how an Internet of Things (IoT) architecture will require data platform technologies to store data.

  • Describe Lambda architectures for a Real-Time Perspective
  • Architect a stream processing pipeline with Azure Stream Analytics
  • Design a stream processing pipeline with Azure Databricks
  • Create an Azure IoT reference architecture

 

In this course, the students will implement various data platform technologies into solutions that are in line with business and technical requirements including on-premises, cloud, and hybrid data scenarios incorporating both relational and No-SQL data. They will also learn how to process data using a range of technologies and languages for both streaming and batch data.

The primary audience for this course is data professionals, data architects, and business intelligence professionals who want to learn about the data platform technologies that exist on Microsoft Azure.

© 2020 U2U All rights reserved.