Spark, Hadoop, and Snowflake for Data Engineering
Completed by Mario Jiménez Gutiérrez
November 10, 2024
29 hours (approximately)
Mario Jiménez Gutiérrez's account is verified. Coursera certifies their successful completion of Spark, Hadoop, and Snowflake for Data Engineering
What you will learn
Create scalable data pipelines (Hadoop, Spark, Snowflake, Databricks) for efficient data handling.
Optimize data engineering with clustering and scaling to boost performance and resource use.
Build ML solutions (PySpark, MLFlow) on Databricks for seamless model development and deployment.
Implement DataOps and DevOps practices for continuous integration and deployment (CI/CD) of data-driven applications, including automating processes.
Skills you will gain
- Category: SQL
- Category: Data Warehousing
- Category: Big Data
- Category: Data Pipelines
- Category: MLOps (Machine Learning Operations)
- Category: Databricks
- Category: Data Integration
- Category: Database Architecture and Administration
- Category: Apache Hadoop
- Category: Python Programming
- Category: Data Quality
- Category: DevOps

