Rejoindre Se connecter

Mettez en pratique vos compétences dans la console Google Cloud

Manuel Alejandro Vargas Carrillo

Date d'abonnement : 2024

Ligue d'Or

9185 points
Build a Data Warehouse with BigQuery Earned juil. 7, 2024 EDT
Get Started with Pub/Sub Earned juil. 4, 2024 EDT
Building Batch Data Pipelines on Google Cloud Earned juil. 2, 2024 EDT
Serverless Data Processing with Dataflow: Foundations Earned mai 16, 2024 EDT

Complete the intermediate Build a Data Warehouse with BigQuery skill badge to demonstrate skills in the following: joining data to create new tables, troubleshooting joins, appending data with unions, creating date-partitioned tables, and working with JSON, arrays, and structs in BigQuery. A skill badge is an exclusive digital badge issued by Google Cloud in recognition of your proficiency with Google Cloud products and services and tests your ability to apply your knowledge in an interactive hands-on environment. Complete the skill badge course, and final assessment challenge lab, to receive a digital badge that you can share with your network.

En savoir plus

Earn a skill badge by completing the Get Started with Pub/Sub skill badge course, where you learn how to use Pub/Sub through the Cloud console, how Cloud Scheduler jobs can save you effort, and when Pub/Sub Lite can save you money on high-volume event ingestion.

En savoir plus

Data pipelines typically fall under one of the Extract and Load (EL), Extract, Load and Transform (ELT) or Extract, Transform and Load (ETL) paradigms. This course describes which paradigm should be used and when for batch data. Furthermore, this course covers several technologies on Google Cloud for data transformation including BigQuery, executing Spark on Dataproc, pipeline graphs in Cloud Data Fusion and serverless data processing with Dataflow. Learners get hands-on experience building data pipeline components on Google Cloud using Qwiklabs.

En savoir plus

This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.

En savoir plus