Skip to main content

Home Specialist skills Artificial Intelligence Data Engineering with Databricks (Associate Level)

Data Engineering with Databricks (Associate Level)

  • bullet point
     Ingest, transform, and manage data using Delta Lake
  • bullet point
    Deploy and monitor data workloads with Databricks Workflows
  • bullet point
    Build scalable data pipelines using Delta Live Tables and the Medallion Architecture
  • bullet point
    Apply data governance principles and manage permissions using Unity Catalog
  • bullet point
    Troubleshoot, optimise, and monitor data workflows in Databricks.

Overview

Off the shelf (OTS)

This course provides an introduction to data engineering with Databricks, covering key tools and frameworks such as Delta Lake, Databricks Workflows, Delta Live Tables, and Unity Catalog. Participants will learn how to ingest, transform, and manage data using Delta Lake, deploy workloads with Databricks Workflows, build efficient pipelines with Delta Live Tables, and apply data governance principles using Unity Catalog. The course includes hands-on labs and real-world applications to ensure learners develop practical skills for working with Databricks effectively.
This course prepares learners for the Associate Data Engineering certification exam and provides the foundational knowledge required to advance to the Advanced Data Engineering with Databricks course.

Participants should have:
• Beginner familiarity with basic cloud concepts (virtual machines, object storage, identity management).
• Ability to perform basic code development tasks (e.g., creating compute instances, running code in notebooks, using basic notebook operations, and importing repositories from Git).
• Intermediate familiarity with SQL, including commands such as CREATE, SELECT, INSERT, UPDATE, DELETE, GROUP BY, JOIN.
• Intermediate experience with SQL concepts such as aggregate functions, filters, sorting, indexes, tables, and views.
• Basic knowledge of Python programming, Jupyter Notebook interface, and PySpark fundamentals.

This course is designed for:
• Data Engineers who want to enhance their knowledge of Databricks and Delta Lake.
• Data Analysts looking to expand their expertise in data pipelines and transformation.
• Cloud Engineers and Developers working with big data frameworks.
• Professionals preparing for the Databricks Associate Data Engineering certification.

This course features:
• Interactive labs to apply concepts in a real-world Databricks environment.
• Guided exercises demonstrating how to configure and optimise Delta Lake, Workflows, and Unity Catalog.
• Real-world case studies showcasing best practices in data engineering with Databricks.
• Troubleshooting scenarios to develop problem-solving skills.

Delivery method
Face to face icon

Face to face

Virtual icon

Virtual

Course duration
Duration icon

14 hours

Competency level
Working icon

Working

Pink building representing strand 4 of the campus map
Delivery method
  • face to face icon

    Face to face

  • Virtual icon

    Virtual

Course duration
Duration icon

14 hours

Competency level
  • Working icon

    Working

chatbotSpark login – Alpha testing