Data Lakehouse with PySpark — High Level Architecture & DW Model

Representation Image

Today we are going to discuss over the High Level System Architecture and Data Warehousing model. We are going to use AWS Cloud services to design the Data Lakehouse and Processing power of Spark to load & manipulate the data.

This series requires basic understanding on Python, Apache Spark, Delta Lake and AWS.

By the end of the series, you would be confident enough to design any Data Lakehouse on cloud architectures.

Checkout out the complete discussion on YouTube:

Make sure to Like and Subscribe.

Follow us on YouTube:

If you are new to Data Warehousing checkout —

Buy me a Coffee

If you like my content and wish to buy me a COFFEE. Click the link below or Scan the QR.
Buy Subham a Coffee
*All Payments are secured through Stripe.

Scan the QR to Pay Securely

About the Author

Subham is working as Senior Data Engineer at a Data Analytics and Artificial Intelligence multinational organization.
Checkout portfolio: Subham Khandelwal