This is a multi-part (free) workshop featuring Azure Databricks. It covers basics of working with Azure Data Services from Spark on Databricks with Chicago crimes public dataset, followed by an end-to-end data engineering workshop with the NYC Taxi public dataset, and finally an end-to-end machine learning workshop. The workshop is offered in Scala and Python.
The goal of this workshop is deliver a clear understanding of how to provision Azure data services, how the data services services integrate with Spark on Azure Databricks, to give you end to end experience with basic data engineering and basic data science on Azure Databricks, and to share some boilerplate code to use in your projects.
This is a community contribution, so we appreciate feedback and contribution.
- Architects
- Data Engineers
- Data Scientists
- Prior knowledge of Spark, is beneficial
- Familiarity/experience with Scala/Python
A subscription with at least $200 credit for a continuous 10-14 hours of usage.
This module covers basics of integrating with Azure Data Services from Spark on Azure Databricks in batch mode and with structured streaming.
At the end of this module, you will know how to provision, configure, and integrate from Spark with:
- Azure storage - blob storage, ADLS gen1 and ADLS gen2; Includes Databricks Delta as well
- Azure Event Hub - publish and subscribe in batch and with structured streaming; Includes Databricks Delta
- HDInsight Kafka - publish and subscribe in batch and with structured streaming; Includes Databricks Delta
- Azure SQL database - read/write primer in batch and structured streaming
- Azure SQL datawarehouse - read/write primer in batch and structured streaming
- Azure Cosmos DB (core API - SQL API/document oriented) - read/write primer in batch and structured streaming; Includes structured streaming aggregation computation
- Azure Data Factory - automating Spark notebooks in Azure Databricks with Azure Data Factory version 2
- Azure Key Vault for secrets management
The Chicago crimes dataset is leveraged in the lab.
This is a batch focused module and covers building blocks of standing up a data engineering pipeline. The NYC taxi dataset (yellow and green taxi trips) is leveraged in the labs.
There are two versions of the Data Science Workshop - the one using Scala will show Spark MLLib models. The PySpark version will show Spark ML and Azure Machine Learning services working together.
If you would like to run Module 3 as standalone, you'll need to:
- Provision:
- Azure Databricks
- Azure Storage account
- Azure Machine Learning services Workspace
- Import the DBC file into the Databricks workspace
- Set the
module_3_only
flag in 99-Shared-Functions-and-Settings toTrue
The following is a summary of content covered:
- Perform feature engineering and feature selection activities
- Create an Azure Machine Learning (AML) service workspace
- Connect to an AML workspace
- Create PySpark models and leverage AML Experiment tracking
- Leverage Automated ML capabilities in AML
- Deploy the best performing model as a REST API in a Docker continer
- Anagha Khanolkar (Chicago) - creator of workshop, primary author of workshop, content design, all development in Scala, primer module in Pyspark
- Ryan Murphy (St Louis) - contribution to the data engineering workshop transformation rules, schema and more
- Rajdeep Biswas (Houston) - writing the entire PySpark version of the data engineering lab
- Steve Howard (St Louis) - contributing to the PySpark version of the data engineering lab
- Erik Zwiefel (Minneapolis) - content design of data science lab, PySpark version, Azure Machine Learning service integration for operationalization as a REST service, AutoML
- Thomas Abraham (St Louis) - development of ADFv2 integration primer in Pyspark
- Matt Stenzel, Christopher House (Minneapolis) - testing