Data Literacy: Essentials of Azure Data Factory
Data Platform is one of the hottest topics in IT right now, as data has become a strategic component in today’s business environment. This course will teach you the fundamentals of Azure Data Factory terms, needed for the exam DP-900.
What you'll learn
Data Engineering is one of the hottest topics on the IT industry at the moment, as the velocity, volume, and variety of data nowadays require skills beyond just traditional ETL. The DP-900 exam, as one of the Microsoft certifications on the data platform, is a good entry point for the area. In this course, Data Literacy: Essentials of Azure Data Factory, you’ll learn the main concepts of Data Factory needed for the exam. First, you’ll explore important concepts such as pipelines and activities. Next, you’ll discover other key components of Azure Data Factory, such as integration runtimes and triggers. Finally, you’ll learn how data factory can be used for data ingestion and processing. When you’re finished with this course, you’ll have the skills and knowledge of Azure Data Factory needed to proceed with the DP-900 exam.
Table of contents
Course FAQ
In this course, you will learn what an Azure Data Factory is and how a data factory can be used. You will also learn essential concepts about the Azure Data Factory tool, inlcuding key Data Factory concepts that are related to the DP-900 exam.
Azure Data Factory is a fully managed, serverless data integration solution for ingesting, preparing, and transforming all your data at scale.
This course will give you the skills and knowledge to properly prepare for the DP-900 exam.
Before beginning this course, you should be familiar with azure basics and corresponding data concepts.
The Azure Data Factory is comprised of pipelines, activities, datasets, linked services, data flows, and integration runtimes.