The Data Engineering Practices training course covers the fundamental best practices in a professional data engineering team.
The course begins by focusing on the software development lifecycle and how it applies to Data Engineering. Next, students will learn about process tools including unit tests and integration testing. The course concludes by covering Continuous Integration/Continuous Delivery (CI/CD) and the best DevOps practices.
Purpose
|
Learn the foundational concepts of distributed computing, distributed data processing, data management and data pipelines. |
Audience
|
Developers, Systems Administrators, Data Scientists looking to learn the fundamental building blocks of big data engineering. |
Role
| Business Analyst - Data Engineer - Data Scientist - Software Developer - System Administrator |
Skill Level
| Intermediate |
Style
| Workshops |
Duration
| 3 Days |
Related Technologies
| CI/CD | Python |
Productivity Objectives
- Interpret the best practices in creating and maintaining data engineering pipelines in a professional setting.
- Explain the background of distributed systems, relational databases and key-value stores.
- Grasp the fundamentals of data stacks, their uses, advantages and limitations.
- Recognize the tools for data management, data access, governance and integration, operations and security.