GCP100DE

Introduction to Data Engineering on Google Cloud

In this course, you learn about data engineering on Google Cloud, the roles and responsibilities of data engineers, and how those map to offerings provided by Google Cloud. You also learn about ways to address data engineering challenges.

Google Cloud
✓ Official training Google CloudLevel Fundamentals⏱️ 1 day (7h)

What you will learn

  • Understand the role of a data engineer.
  • Identify data engineering tasks and core components used on Google Cloud.
  • Understand how to create and deploy data pipelines of varying patterns on Google Cloud.
  • Identify and utilize various automation techniques on Google Cloud.

Prerequisites

  • Prior Google Cloud experience at the fundamental level using Cloud Shell and accessing products from the Google Cloud console.
  • Basic proficiency with a common query language such as SQL.
  • Experience with data modeling and ETL (extract, transform, load) activities.
  • Experience developing applications using a common programming language such as Python.

Target audience

  • Data engineers, Database administrators, System administrators

Training Program

6 modules to master the fundamentals

Objectives
  • Explain the role of a data engineer.
  • Understand the differences between a data source and a data sink.
  • Explain the different types of data formats.
  • Explain the storage solution options on Google Cloud.
  • Learn about the metadata management options on Google Cloud.
  • Understand how to share datasets with ease using Analytics Hub.
  • Understand how to load data into BigQuery using the Google Cloud console or the gcloud CLI.
Topics covered
  • →The role of a data engineer
  • →Data sources versus data sinks
  • →Data formats
  • →Storage solution options on Google Cloud
  • →Metadata management options on Google Cloud
  • →Sharing datasets using Analytics Hub
Activities

Lab: Loading Data into BigQuery

Quiz

Objectives
  • Explain the baseline Google Cloud data replication and migration architecture.
  • Understand the options and use cases for the gcloud command-line tool.
  • Explain the functionality and use cases for Storage Transfer Service.
  • Explain the functionality and use cases for Transfer Appliance.
  • Understand the features and deployment of Datastream.
Topics covered
  • →Replication and migration architecture
  • →The gcloud command-line tool
  • →Moving datasets
  • →Datastream
Activities

Lab: Datastream: PostgreSQL Replication to BigQuery (optional for ILT)

Quiz

Objectives
  • Explain the baseline extract and load architecture diagram.
  • Understand the options of the bq command-line tool.
  • Explain the functionality and use cases for BigQuery Data Transfer Service.
  • Explain the functionality and use cases for BigLake as a non-extract-load pattern.
Topics covered
  • →Extract and load architecture
  • →The bq command-line tool
  • →BigQuery Data Transfer Service
  • →BigLake
Activities

Lab: BigLake: Qwik Start

Quiz

Objectives
  • Explain the baseline extract, load, and transform architecture diagram.
  • Understand a common ELT pipeline on Google Cloud.
  • Learn about BigQuery's SQL scripting and scheduling capabilities.
  • Explain the functionality and use cases for Dataform.
Topics covered
  • →Extract, load, and transform (ELT) architecture
  • →SQL scripting and scheduling with BigQuery
  • →Dataform
Activities

Lab: Create and Execute a SQL Workflow in Dataform

Quiz

Objectives
  • Explain the baseline extract, transform, and load architecture diagram.
  • Learn about the GUI tools on Google Cloud used for ETL data pipelines.
  • Explain batch data processing using Dataproc.
  • Learn how to use Dataproc Serverless for Spark for ETL.
  • Explain streaming data processing options.
  • Explain the role Bigtable plays in data pipelines.
Topics covered
  • →Extract, transform, and load (ETL) architecture
  • →Google Cloud GUI tools for ETL data pipelines
  • →Batch data processing using Dataproc
  • →Streaming data processing options
  • →Bigtable and data pipelines
Activities

Lab: Use Dataproc Serverless for Spark to Load BigQuery (optional for ILT)

Lab: Creating a Streaming Data Pipeline for a Real-Time Dashboard with Dataflow

Quiz

Objectives
  • Explain the automation patterns and options available for pipelines.
  • Learn about Cloud Scheduler and Workflows.
  • Learn about Cloud Composer.
  • Learn about Cloud Run functions.
  • Explain the functionality and automation use cases for Eventarc.
Topics covered
  • →Automation patterns and options for pipelines
  • →Cloud Scheduler and Workflows
  • →Cloud Composer
  • →Cloud Run Functions
  • →Eventarc
Activities

Lab: Use Cloud Run Functions to Load BigQuery (optional for ILT)

Quiz

Related Trainings

SFEIR Institute
Best

dbt

Learn to transform your data with dbt, the leading tool in the Modern Data Stack. You'll start by understanding the evolution of data architectures and the difference between ETL and ELT. You'll install dbt, create your first project and connect it to your data sources. Then you'll learn to build structured data models, choose the right materialization options (table, view, incremental) and organize your metadata with tags. You'll discover how to reference your sources and manage dependencies between models. You'll explore advanced features: seeds to initialize reference data, snapshots to track history and manage slowly changing dimensions, Jinja macros and variables to automate your transformations. Finally, you'll implement automated tests to ensure data quality, document your models with lineage, and discover packages from the dbt community. Hands-on training with 60% labs.

2 d
Fundamental
Google Cloud

Introduction to Data Analytics on Google Cloud

This course is an introduction to data analytics on Google Cloud. It is designed for learners who have no prior experience with data analytics or Google Cloud. The course covers the basics of data analysis, including collection, storage, exploration, visualization, and sharing. It also introduces learners to Google Cloud's data analytics tools and services. Through video lectures, demos, quizzes, and hands-on labs, the course demonstrates how to go from raw data to impactful visualizations and dashboards.

1 d
Fundamental
Google Cloud

Analyzing and Visualizing Data in Looker

In this course, you learn how to do the kind of data exploration and analysis in Looker that would formerly be done primarily by SQL developers or analysts. You learn to leverage Looker's modern analytics platform to find and explore relevant content in your organization's Looker instance. You also discover how to ask questions of your data, create new metrics as needed, and build and share visualizations and dashboards to facilitate data-driven decision-making.

1 d
Fundamental

Upcoming sessions

March 31, 2026
Distanciel • Français
Register
April 21, 2026
Distanciel • Français
Register
May 29, 2026
Distanciel • Français
Register
June 25, 2026
Distanciel • Français
Register
July 31, 2026
Distanciel • Français
Register
August 27, 2026
Distanciel • Français
Register
September 30, 2026
Distanciel • Français
Register
October 22, 2026
Distanciel • Français
Register
November 27, 2026
Distanciel • Français
Register
December 10, 2026
Distanciel • Français
Register

Quality Process

SFEIR Institute's commitment: an excellence approach to ensure the quality and success of all our training programs. Learn more about our quality approach

Teaching Methods Used
  • Lectures / Theoretical Slides — Presentation of concepts using visual aids (PowerPoint, PDF).
  • Technical Demonstration (Demos) — The instructor performs a task or procedure while students observe.
  • Guided Labs — Guided practical exercises on software, hardware, or technical environments.
  • Quiz / MCQ — Quick knowledge check (paper-based or digital via tools like Kahoot/Klaxoon).
Evaluation and Monitoring System

The achievement of training objectives is evaluated at multiple levels to ensure quality:

  • Continuous Knowledge Assessment : Verification of knowledge throughout the training via participatory methods (quizzes, practical exercises, case studies) under instructor supervision.
  • Progress Measurement : Comparative self-assessment system including an initial diagnostic to determine the starting level, followed by a final evaluation to validate skills development.
  • Quality Evaluation : End-of-session satisfaction questionnaire to measure the relevance and effectiveness of the training as perceived by participants.

790€ excl. VAT

per learner