-
Onboard
For Tech Teams
- Reduce initial time to productivity.
- Increase employee tenure.
- Plug-and-play into HR onboarding and career pathing programs.
- Customize for ad-hoc and cohort-based hiring approaches.
-
Upskill
For Tech Teams
- Upgrade and round out developer skills.
- Tailor to tech stack and specific project.
- Help teams, business units, centers of excellence and corporate tech universities.
-
Reskill
For Tech Teams
- Offer bootcamps to give employees a running start.
- Create immersive and cadenced learning journeys with guaranteed results.
- Supplement limited in-house L&D resources with all-inclusive programs to meet specific business goals.
-
Design
For Tech Teams
- Uplevel your existing tech learning framework.
- Extend HR efforts to provide growth opportunities within the organization.
- Prepare your team for an upcoming tech transformation.
Get your team started on a custom learning journey today!
Our Boulder, CO-based learning experts are ready to help!
Course Summary
The Google Cloud for Data Engineers training course teaches students the fundamentals of Google Cloud Platform (GCP) for building and running data pipelines that process batch or streaming data.
This course starts by learning about the GCP services most frequently used by data engineers. Next, students will advance their existing skills in SQL, Hadoop, and Python by understanding how to reuse existing applications to take advantage of managed MySQL and Hadoop/Spark infrastructure on GCP. Most of the course focuses on differentiating capabilities of GCP for data engineering. Next, students will learn how to process, analyze, and store petabytes of batch and streaming data with serverless capabilities like PubSub, BigQuery, and Dataflow. For example, students will work with Apache Beam code that enables going beyond the limitations of the original MapReduce framework. The course concludes with students being introduced to the Machine Learning capabilities of GCP that data engineers can start using without having prior data science experience.
The course will also provide architectural overviews of data processing pipelines enabled by GCP and how you choose the right GCP services for your project.
- Productivity Objectives:
- Describe the capabilities of Google Cloud for data engineering.
- Build and run data processing pipelines on GCP to ingest, analyze, and store data.
- Identify how to use managed Google Cloud infrastructure for MySQL and Hadoop/Spark.
- Discuss when and how to use PubSub, DataFlow, and BigQuery for serverless data pipelines.
- Integrate data pipelines with other GCP services.
- Identify what criteria to use for the design of data processing pipelines on GCP.
Request Information
Get your team upskilled or reskilled today. Chat with one of our experts to create a custom training proposal. Fully customized at no additional cost.

If you are not completely satisfied with your training class, we'll give you your money back.




about our training
-
Real-World Content
Project-focused demos and labs using your tool stack and environment, not some canned "training room" lab.
-
Expert Practitioners
Industry experts with 15+ years of industry experience that bring their battle scars into the classroom.
-
Experiential Learning
More coding than lecture, coupled with architectural and design discussions.
-
Fully Customized
One-size-fits-all doesn't apply to training teams. That's where we come in!
What You'll Learn
In the Google Cloud for Data Engineers training course, you'll learn:
- Google Cloud Basics
- Why Google Cloud
- Managed Virtual Infrastructure vs. Serverless
- Google Cloud for Data Engineers
- Compute Engine
- Virtualized Infrastructure
- Cloud Shell
- Persistent vs. Transient Storage
- Computer Engine User Interface
- Pre-emptible Instances
- Cloud Storage (GCS)
- Object Storage and Buckets
- Integration with GCP
- Web-based and Command Line Interfaces
- Cloud SQL
- Provisioning Managed Database Infrastructure
- Configuration of MySQL on GCP
- Batch Data Import/Export with Cloud SQL
- Web-based Interface
- Integration of MySQL with GCP Services and Applications
- Cloud Pub/Sub
- Distributed Messaging Basics
- Publish/Subscribe Messaging Model
- Topics and Subscriptions for Messaging
- Command Line and Python Interfaces
- Datastore
- Object-Relational Impedance Mismatch
- Datastore for Transactional Data
- Java APIs for Datastore
- Machine Learning APIs
- Colaboratory
- Vision, Natural Language, Translate APIs
- AutoML Vision
- Dataproc
- MapReduce Framework
- Provisioning Managed Apache Hadoop/Spark/YARN Infrastructure
- Customizing Apache Bigtop Distribution
- Pre-Emptible Instances for MapReduce
- Dataproc User and Command Line Interfaces
- Map vs. FlatMap for MapReduce
- Running Apache Hive, Apache Pig, and PySpark
- Running and Monitoring MapReduce Jobs
- Storage Migration from HDFS to GCS
- DataFlow
- Apache Beam Framework
- Batch and Streaming Data Processing Pipelines
- Run Apache Beam in Cloud Shell
- Apache Beam Combine vs. GroupBy
- Submitting Apache Beam Pipelines
- Running Batch and Streaming Dataflow Jobs
- Apache Beam Pipelines with Side-Inputs
- Autoscaling Streaming Apache Beam Jobs
- Apache Beam Windows and Triggers
- Web-based and Command Line Interface
- Monitoring Dataflow Jobs
- BigQuery
- Serverless data warehousing
- Columnar vs. Row-based Storage
- Normalization vs. Denormalization with Columnar Storage
- Projects, Datasets, Tables
- Batch Data Import/Export
- Semi-Structured Data Analysis with SQL Arrays and Structs
- Partitions and Performance Optimizations
- Data Engineering with GCP
- Architectures for sample batch and streaming pipelines
- GCP Storage Optimal Access Patterns
- GCP Storage Service Selection Decision Model
- Cost Estimation
Real-world content
Project-focused demos and labs using your tool stack and environment, not some canned "training room" lab.
Expert Practitioners
Industry experts that bring their battle scars into the classroom.
Experiential Learning
More coding than lecture, coupled with architectural and design discussions.
Fully Customized
One-size-fits-all doesn't apply to training teams. That's where we come in!

Elite Instructor Program
We recently launched our internal Elite Instructor Program. The community driven instructor program is designed to support instructors in transforming students’ lives by consistently showing a world-class level of engagement, ability, and teaching prowess. Reach out today to learn more about our instructors.
Customized Technical Learning Solutions to Help Attract and Retain Talented Developers
Let DI help you design solutions to onboard, upskill or reskill your software development organization. Fully customized. 100% guaranteed.
DevelopIntelligence leads technical and software development learning programs for Fortune 500 companies. We provide learning solutions for hundreds of thousands of engineers for over 250 global brands.



“I appreciated the instructor’s technique of writing live code examples rather than using fixed slide decks to present the material.”
VMwareAbout Us
LET’S DISCUSS
DevelopIntelligence has been in the technical/software development learning and training industry for nearly 20 years. We’ve provided learning solutions to more than 48,000 engineers, across 220 organizations worldwide.
Resources
Thank you for everyone who joined us this past year to hear about our proven methods of attracting and retaining tech talent.

- Boulder, Colorado Headquarters: 980 W. Dillon Road, Louisville, CO 80027
© 2013 - 2022 DevelopIntelligence LLC - Privacy Policy