> Formations > Introduction to Data Engineering on Google Cloud

Course : Introduction to Data Engineering on Google Cloud

Official course, preparation for Google Cloud certification exams

Practical course - 1d - 7h00 - Ref. GDE
Price : 850 € E.T.

Introduction to Data Engineering on Google Cloud

Official course, preparation for Google Cloud certification exams


New course

With this training, you'll gain the fundamental knowledge and skills in data engineering on Google Cloud Platform (GCP) to understand, design and implement efficient data pipelines. You'll learn how to use GCP's core services for data ingestion, storage, transformation and analysis, while adopting best practices for data performance, security and governance.


INTER
IN-HOUSE
CUSTOM

Practical course in person or remote class
Available in English on request

Ref. GDE
  1d - 7h00
850 € E.T.




With this training, you'll gain the fundamental knowledge and skills in data engineering on Google Cloud Platform (GCP) to understand, design and implement efficient data pipelines. You'll learn how to use GCP's core services for data ingestion, storage, transformation and analysis, while adopting best practices for data performance, security and governance.


Teaching objectives
At the end of the training, the participant will be able to:
Understanding the role of a data engineer
Identify data engineering tasks and the main components used on Google Cloud
Understand how to create and deploy data pipelines of different models on Google Cloud
Identify and use various automation techniques on Google Cloud

Intended audience
Data engineers, database administrators, system administrators.

Prerequisites
Basic knowledge of the Google Cloud environment. Basic command of a common query language. Experience in data modeling, ETL activities and application development.

Certification
We recommend you take this course if you want to prepare for certification as a "Google Cloud Professional Data Engineer".
Comment passer votre examen ?

Practical details
Teaching methods
Training in French. Official course material in digital format and in English. Good understanding of written English.

Course schedule

1
Data engineering tasks and components

  • Understand the role of a data engineer.
  • Understand the differences between a data source and a data sink.
  • Discover the different types of data formats.
  • Explain Google Cloud storage solution options.
  • Discover metadata management options on Google Cloud.
  • Discover how to easily share data sets using Analytics Hub.
  • Discover how to load data into BigQuery using the Google Cloud Console or the gcloud CLI.
Hands-on work
Lab: loading data into BigQuery.

2
Data replication and migration

  • Discover Google Cloud's basic replication and data migration architecture.
  • Understand the options and use cases of the gcloud command line tool.
  • Discover the features and use cases of Storage Transfer Service.
  • Discover Transfer Appliance features and use cases.
  • Discover Datastream functionalities and deployment.
Hands-on work
Lab: PostgreSQL to BigQuery replication.

3
The Extract and Load pipeline model

  • Extract and load architecture.
  • Understand the bq command line tool options.
  • Discover the features and use cases of the BigQuery data transfer service.
  • Discover the features and use cases of BigLake as a non-extraction loading model.
Hands-on work
Lab: Introduction to BigLake.

4
The Extract, Load and Transform pipeline model

  • Explain the basic extraction, loading and transformation architecture diagram.
  • Understand a common ELT pipeline on Google Cloud.
  • Discover BigQuery's programming and SQL scripting features.
  • Explain Dataform functionalities and use cases.
Hands-on work
Lab: Create and run a SQL workflow in Dataform.

5
Extract, Transform and Load pipeline model

  • Discover the basic extraction, transformation and loading architecture diagram.
  • Discover the graphical user interface tools on Google Cloud used for ETL data pipelines.
  • Explain batch data processing with Dataproc.
  • Using Dataproc Serverless for Spark for ETL...
  • Explore streaming data processing options.
  • Understand Bigtable's role in data pipelines.
Hands-on work
Lab: Using Dataproc Serverless for Spark to load BigQuery. Lab: Create a continuous data pipeline for a real-time dashboard with Dataflow.

6
Automation techniques

  • Discover the automation models and options available for pipelines.
  • Discover Cloud Scheduler and Workflows.
  • Discover Cloud Composer.
  • Discover Cloud Run functions.
  • Discover use cases for Eventarc functionalities and automation.
Hands-on work
Lab : Utiliser les fonctions Cloud Run pour charger BigQuery.


Dates and locations
Select your location or opt for the remote class then choose your date.
Remote class

Dernières places
Date garantie en présentiel ou à distance
Session garantie

REMOTE CLASS
2026 : 26 Mar., 18 June, 1 Oct., 10 Dec.

PARIS LA DÉFENSE
2026 : 26 Mar., 18 June, 1 Oct., 10 Dec.