course
Implementing a Data Analytics Solution with Azure Databricks
Get started with data engineering on Azure Databricks
Description
Learn how to harness the power of Apache Spark and powerful clusters running on the Azure Databricks platform to run large data engineering workloads in the cloud.
Prior Knowledge
Knowledge of data engineering and Microsoft Azure
Subjects
Module 1: Explore Azure Databricks
Azure Databricks is a cloud service that provides a scalable platform for data analytics using Apache Spark.
After completing this module, you will be able to:
- Provision an Azure Databricks workspace.
- Identify core workloads and personas for Azure Databricks.
- Describe key concepts of an Azure Databricks solution.
Module 2: Use Apache Spark in Azure Databricks
Azure Databricks is built on Apache Spark and enables data engineers and analysts to run Spark jobs to transform, analyze and visualize data at scale.
After completing this module, you will be able to:
- Describe key elements of the Apache Spark architecture.
- Create and configure a Spark cluster.
- Describe use cases for Spark.
- Use Spark to process and analyze data stored in files.
- Use Spark to visualize data.
Module 3: Use Delta Lake in Azure Databricks
Delta Lake is an open source relational storage area for Spark that you can use to implement a data lakehouse architecture in Azure Databricks.
After completing this module, you will be able to:
- Describe core features and capabilities of Delta Lake.
- Create and use Delta Lake tables in Azure Databricks.
- Create Spark catalog tables for Delta Lake data.
- Use Delta Lake tables for streaming data.
Module 4: Use SQL Warehouses in Azure Databricks
Azure Databricks provides SQL Warehouses that enable data analysts to work with data using familiar relational SQL queries.
After completing this module, you will be able to:
- Create and configure SQL Warehouses in Azure Databricks.
- Create databases and tables.
- Create queries and dashboards.
Module 5: Run Azure Databricks Notebooks with Azure Data Factory
Using pipelines in Azure Data Factory to run notebooks in Azure Databricks enables you to automate data engineering processes at cloud scale.
After completing this module, you will be able to:
- Describe how Azure Databricks notebooks can be run in a pipeline.
- Create an Azure Data Factory linked service for Azure Databricks.
- Use a Notebook activity in a pipeline.
- Pass parameters to a notebook.
Schedule
All courses can also be conducted within your organization as customized or incompany training.
Our training advisors are happy to help you provide personal advice or find Incompany training within your organization.
Prior knowledge courses
Microsoft Azure Data Fundamentals (DP-900)
Learn foundational knowledge of cloud data services within Microsoft Azure
- Cloud
Introduction to Data Analytics and Business Intelligence
Principles of Data Analytics and Business Intelligence explained
- Business Intelligence
Essentials of Python Development
Attain a solid foundation of Python for developing software solutions
- Python
"Trainer who knows his profession!"Marc
-
Hoge waardering
-
Praktijkgerichte trainingen
-
Gecertificeerde trainers
-
Eigen docenten