course

NL/EN
This training is available in Dutch and English. More information

Implementing a Data Analytics Solution with Azure Databricks

Get started with data engineering on Azure Databricks

Not yet scheduled
- No location
-
Keep me posted
1 day
740 (excl. VAT)
Keep me posted

Description

Learn how to harness the power of Apache Spark and powerful clusters running on the Azure Databricks platform to run large data engineering workloads in the cloud.

Prior Knowledge

Knowledge of data engineering and Microsoft Azure

Subjects

Module 1: Explore Azure Databricks

Azure Databricks is a cloud service that provides a scalable platform for data analytics using Apache Spark.

After completing this module, you will be able to:

  • Provision an Azure Databricks workspace.
  • Identify core workloads and personas for Azure Databricks.
  • Describe key concepts of an Azure Databricks solution.

Module 2: Use Apache Spark in Azure Databricks

Azure Databricks is built on Apache Spark and enables data engineers and analysts to run Spark jobs to transform, analyze and visualize data at scale.

After completing this module, you will be able to:

  • Describe key elements of the Apache Spark architecture.
  • Create and configure a Spark cluster.
  • Describe use cases for Spark.
  • Use Spark to process and analyze data stored in files.
  • Use Spark to visualize data.

Module 3: Use Delta Lake in Azure Databricks

Delta Lake is an open source relational storage area for Spark that you can use to implement a data lakehouse architecture in Azure Databricks.

After completing this module, you will be able to:

  • Describe core features and capabilities of Delta Lake.
  • Create and use Delta Lake tables in Azure Databricks.
  • Create Spark catalog tables for Delta Lake data.
  • Use Delta Lake tables for streaming data.

Module 4: Use SQL Warehouses in Azure Databricks

Azure Databricks provides SQL Warehouses that enable data analysts to work with data using familiar relational SQL queries.

After completing this module, you will be able to:

  • Create and configure SQL Warehouses in Azure Databricks.
  • Create databases and tables.
  • Create queries and dashboards.

Module 5: Run Azure Databricks Notebooks with Azure Data Factory

Using pipelines in Azure Data Factory to run notebooks in Azure Databricks enables you to automate data engineering processes at cloud scale.

After completing this module, you will be able to:

  • Describe how Azure Databricks notebooks can be run in a pipeline.
  • Create an Azure Data Factory linked service for Azure Databricks.
  • Use a Notebook activity in a pipeline.
  • Pass parameters to a notebook.

Schedule

Keep me posted

All courses can also be conducted within your organization as customized or incompany training.

Our training advisors are happy to help you provide personal advice or find Incompany training within your organization.

Prior knowledge courses

"Trainer who knows his profession!"
Marc
  • icon

    Hoge waardering

  • icon

    Praktijkgerichte trainingen

  • icon

    Gecertificeerde trainers

  • icon

    Eigen docenten