At 5-10 hours/week
Get access to the classroom immediately upon enrollment
Hide details
To be successful in this program, you should have intermediate Python and SQL skills, as well as experience with ETL.See detailed requirements.
Learn the fundamentals of stream processing, including how to work with the Apache Kafka ecosystem, data schemas, ApacheAvro, Kafka Connect and REST proxy, KSQL, and Faust Stream Processing.
Grow expertise in the components of streaming data systems, and learn how to build a real-time analytics application. Understand the dimensions of Spark streaming, ingest streaming data to Apache Spark Structured Streaming, and integrate Spark streams into Kafka.
Real-world projects from industry experts
1-on-1 technical mentor
Personal career coach and career services
Flexible learning program

Staff Engineer at SpotHero
In his career as an engineer, Ben Goldberg has worked in fields ranging from Computer Vision to Natural Language Processing. At SpotHero, he founded and built out their Data Engineering team, using Airflow as one of the key technologies.

Senior Data Engineer at Netflix
Currently, Judit is a Senior Data Engineer at Netflix. Formerly a Data Engineer at Split, where she worked on the statistical engine of their full-stack experimentation platform, she has also been an instructor at Insight Data Science, helping software engineers and academic coders transition to DE roles.

VP of Engineering at Insight
David is VP of Engineering at Insight where he enjoys breaking down difficult concepts and helping others learn data engineering. David has a PhD in Physics from UC Riverside.

Senior Data Engineer at Change Healthcare
Jillian has worked in roles from building data analytics platforms to machine learning pipelines. Previously, she was a research engineer at Samsung focused on data analytics and ML, and now leads building pipelines at scale as a Senior Data Engineer at Change Healthcare.

Numbers don't lie. See what difference it makes in career searches.*
Career-seeking and job-ready graduates found a new, better job within six months of graduation.
Average salary increase for graduates who found a new, better job within six months of graduation.
As businesses increasingly rely on applications that produce and process data in real-time, data streaming is an increasingly in-demand skill for data engineers. The Data Streaming Nanodegree program will prepare you for the cutting edge of data engineering as more and more companies look to derive live insights from data at scale.
Students will learn how to process data in real-time by building fluency in modern data engineering tools, such as Apache Spark, Kafka, Spark Streaming, and Kafka Streaming.
You’ll start by understanding the components of data streaming systems. You’ll then build a real-time analytics application. You will also compile data and run analytics, as well as draw insights from reports generated by the streaming console.
This program is designed to upskill experienced Software Engineers and Data Engineers to learn the latest advancements in data processing, sending data records continuously to support live updating.
The projects in the Data Streaming Nanodegree program will prepare you to develop systems and applications capable of interpreting data in real-time, and position you for roles in all industries that require live data processing for functions including big data, cloud computing, web personalization, fraud detection, sensor monitoring, anomaly detection, supply chain maintenance, location-based services, and much more.
This program is intended for software engineers looking to build real-time data processing proficiency, as well as data engineers looking to enhance their existing skill set with the next advancement in data engineering.
There is no application. This Nanodegree program accepts everyone, regardless of experience and specific background.
The Data Streaming Nanodegree program is designed for students with intermediate Python and SQL skills, as well as experience with ETL. Basic familiarity with traditional batch processing and basic conceptual familiarity with traditional service architectures is desired, but not required.
Intermediate Python programming knowledge, of the sort gained through the Programming for Data Science Nanodegree program, other introductory programming courses or programs, or additional real-world software development experience. Including:
Intermediate SQL knowledge and linear algebra mastery, addressed in the Programming for Data Science Nanodegree program, including:
Udacity’s Programming for Data Science with Python Nanodegree program is great preparation for the Data Engineer Nanodegree program. You’ll learn to code with Python and SQL.
Similarly, the Data Engineering Nanodegree program is great preparation for the Data Streaming Nanodegree program.
The Data Streaming Nanodegree program is comprised of content and curriculum to support two projects. We estimate that students can complete the program in two months, working five to ten hours per week.
Each project will be reviewed by the Udacity reviewer network. Feedback will be provided, and if you do not pass the project, you will be asked to resubmit the project until it passes.
Access to this Nanodegree program runs for the length of time specified in the payment card above. If you do not graduate within that time period, you will continue learning with month to month payments. See the Terms of Use and FAQs for other policies regarding the terms of access to our Nanodegree programs.
Please see the Udacity Nanodegree program FAQs for policies on enrollment in our programs.
There are no software and version requirements to complete this Nanodegree program. All coursework and projects can be completed via Student Workspaces in the Udacity online classroom.