8 Steps for a Developer to Learn Apache Spark with Delta Lake – DNU

For data engineers, building fast, reliable pipelines is only the beginning. Today, you also need to deliver clean, high quality data ready for downstream users to do BI and ML.

Apache Spark™ and Delta Lake deliver fast, reliable data to your data teams for all your data engineering, data science, machine learning, and business analytics use cases. And these projects are open source and use open formats, so you can easily access your data using your tools of choice.

  • Why Apache Spark and Delta Lake
  • Apache Spark and Delta Lake concepts, key terms and keywords
  • Advanced Apache Spark internals and core
  • DataFrames, Datasets and Spark SQL essentials
  • Graph processing with GraphFrames
  • Continuous applications with structured streaming
  • Machine learning for humans
  • Data reliability challenges for data lakes
  • Delta Lake for ACID transactions, schema enforcement and more
  • Unifying batch and streaming data pipelines
Read now

Resource Details

Databricks logo
Provided by: