Data Scientists must be able to acquire data and to organize that data wherever it exists on online.
We'll learn Python and SQL, as well as the fundamental ideas underlying the web.
This is the entire 9-week Programming Course, and the first section of the Career Course.
The data pipeline is the step by step process for obtaining data, organizing data, and deploying data services.
We'll learn the skills to put a website or machine learning model into production with testing, a data pipeline, and deployment.
Here, we'll learn about big data and distributed computing with Pyspark. We'll learn how to process streaming data with Kafka.
We'll finish up our data pipeline skills with airflow.
We'll review and improve our coding skills as we go on data engineering interviews.
We'll review data structures and algorithms, improve our final projects, and review and improve our SQL skills.