Oussama Ben Nasr
成为会员时间:2022
成为会员时间:2022
Learn how to write and test pipelines with Dataflow and Apache Beam
本课程将向您介绍注意力机制,这是一种强大的技术,可令神经网络专注于输入序列的特定部分。您将了解注意力的工作原理,以及如何使用它来提高各种机器学习任务的性能,包括机器翻译、文本摘要和问题解答。
本课程向您介绍扩散模型。这类机器学习模型最近在图像生成领域展现出了巨大潜力。扩散模型的灵感来源于物理学,特别是热力学。过去几年内,扩散模型成为热门研究主题并在整个行业开始流行。Google Cloud 上许多先进的图像生成模型和工具都是以扩散模型为基础构建的。本课程向您介绍扩散模型背后的理论,以及如何在 Vertex AI 上训练和部署此类模型。
完成在 Vertex AI 上构建和部署机器学习解决方案课程,赢取中级技能徽章。 在此课程中,您将了解如何使用 Google Cloud 的 Vertex AI Platform、AutoML 以及自定义训练服务来 训练、评估、调优、解释和部署机器学习模型。 此技能徽章课程的目标受众是专业的数据科学家和机器学习 工程师。 技能徽章是由 Google Cloud 颁发的专属数字徽章,旨在认可 您对 Google Cloud 产品与服务的熟练度;您需要在 交互式实操环境中参加考核,证明自己运用所学知识的能力后才能获得此徽章。完成此技能徽章课程 和作为最终评估的实验室挑战赛,即可获得数字徽章, 在您的人际圈中炫出自己的技能。
探索生成式 AI - Vertex AI 课程汇集了多组实验, 指导用户在 Google Cloud 平台上运用生成式 AI。参与实验,您将了解 如何使用 Vertex AI PaLM API 系列模型,包括 text-bison、chat-bison 和 textembedding-gecko。您还将了解提示设计、最佳实践, 以及如何使用生成式 AI 进行构思、文本分类、文本提取、文本 总结等任务。您还将学习如何通过 Vertex AI 自定义训练对基础模型进行调优, 并将模型部署到 Vertex AI 端点。
本课程介绍 Vertex AI Studio,这是一种用于与生成式 AI 模型交互、围绕业务创意进行原型设计并在生产环境中落地的工具。通过沉浸式应用场景、富有吸引力的课程和实操实验,您将探索从提示到产品的整个生命周期,了解如何将 Vertex AI Studio 用于多模态 Gemini 应用、提示设计、提示工程和模型调优。本课程的目的在于帮助您利用 Vertex AI Studio,在自己的项目中充分发掘生成式 AI 的潜力。
完成 Introduction to Generative AI、Introduction to Large Language Models 和 Introduction to Responsible AI 三门课程,赢取技能徽章。通过最终测验,即表明您理解了生成式 AI 的基本概念。 技能徽章是由 Google Cloud 颁发的数字徽章,旨在认可您对 Google Cloud 产品与服务的了解程度。公开您的个人资料并将技能徽章添加到您的社交媒体个人资料中,以此来分享您获得的成就。
这是一节入门级微课程,旨在解释什么是负责任的 AI、它的重要性,以及 Google 如何在自己的产品中实现负责任的 AI。此外,本课程还介绍了 Google 的 7 个 AI 开发原则。
这是一节入门级微课程,旨在解释什么是生成式 AI、它的用途以及与传统机器学习方法的区别。该课程还介绍了可以帮助您开发自己的生成式 AI 应用的各种 Google 工具。
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Machine Learning Engineering professionals use tools for continuous improvement and evaluation of deployed models. They work with (or can be) Data Scientists, who develop models, to enable velocity and rigor in deploying the best performing models.
Looking to build or optimize your data warehouse? Learn best practices to Extract, Transform, and Load your data into Google Cloud with BigQuery. In this series of interactive labs you will create and optimize your own data warehouse using a variety of large-scale BigQuery public datasets. BigQuery is Google's fully managed, NoOps, low cost analytics database. With BigQuery you can query terabytes and terabytes of data without having any infrastructure to manage or needing a database administrator. BigQuery uses SQL and can take advantage of the pay-as-you-go model. BigQuery allows you to focus on analyzing data to find meaningful insights. Looking for a hands on challenge lab to demonstrate your skills and validate your knowledge? On completing this quest, enroll in and finish the additional challenge lab at the end of this quest to receive an exclusive Google Cloud digital badge.
In this course, you learn how to do the kind of data exploration and analysis in Looker that would formerly be done primarily by SQL developers or analysts. Upon completion of this course, you will be able to leverage Looker's modern analytics platform to find and explore relevant content in your organization’s Looker instance, ask questions of your data, create new metrics as needed, and build and share visualizations and dashboards to facilitate data-driven decision making.
Data Catalog is deprecated and will be discontinued on January 30, 2026. You can still complete this course if you want to. For steps to transition your Data Catalog users, workloads, and content to Dataplex Catalog, see Transition from Data Catalog to Dataplex Catalog (https://cloud.google.com/dataplex/docs/transition-to-dataplex-catalog). Data Catalog is a fully managed and scalable metadata management service that empowers organizations to quickly discover, understand, and manage all of their data. In this quest you will start small by learning how to search and tag data assets and metadata with Data Catalog. After learning how to build your own tag templates that map to BigQuery table data, you will learn how to build MySQL, PostgreSQL, and SQLServer to Data Catalog Connectors.
完成入门技能徽章课程使用 Dataplex 构建数据网格,展示以下方面的技能:使用 Dataplex 构建数据网格, 以在 Google Cloud 上实现数据安全、治理和发现。您将在 Dataplex 中练习和测试自己在标记资产、分配 IAM 角色和评估数据质量方面的技能。 技能徽章是由 Google Cloud 颁发的专属数字徽章,旨在认可 您在 Google Cloud 产品与服务方面的熟练度;您需要在 交互式实操环境中参加考核,证明自己运用所学知识的能力后才能获得此徽章。完成此技能 徽章课程和作为最终评估的实验室挑战赛,即可获得数字徽章, 在您的人际圈中炫出自己的技能。
Moving to the cloud creates numerous opportunities to start working in a new way and it empowers the workforce to better collaborate and innovate. But it’s also a big change. Sometimes the success of the change hinges not on the change itself, but on how it’s managed. This course will help people managers to understand some of the key challenges associated with cloud adoption, and provide them with a verified in-the-field framework that will assist them in supporting their teams on the change journey. By addressing the human factor of moving to the cloud, organizations increase their chances of realizing business objectives and investing in their future talent.
Cloud technology can bring great value to an organization, and combining the power of cloud technology with data has the potential to unlock even more value and create new customer experiences. “Exploring Data Transformation with Google Cloud” explores the value data can bring to an organization and ways Google Cloud can make data useful and accessible. Part of the Cloud Digital Leader learning path, this course aims to help individuals grow in their role and build the future of their business.
Welcome to Data Governance, where we discuss how to implement data governance on Google Cloud.
完成 Pub/Sub 使用入门挑战任务,赢取技能徽章。 您将学习如何通过 Cloud 控制台使用 Pub/Sub,如何使用 Cloud Scheduler 安排 作业以节省工作量,以及 Pub/Sub Lite 如何在有大量 事件注入的场景中节省费用。 技能徽章是一种专属数字徽章, 由 Google Cloud 颁发,旨在认可您已熟悉特定的 Google Cloud 产品与服务, 并在交互式实操环境中证明了运用所学知识 的能力。完成此技能徽章课程和作为最终评估的实验室挑战赛, 获得相应的数字徽章,在您的人际圈中秀出自己的技能。
Welcome to Cloud Composer, where we discuss how to orchestrate data lake workflows with Cloud Composer.
In this course, we see what the common challenges faced by data analysts are and how to solve them with the big data tools on Google Cloud. You’ll pick up some SQL along the way and become very familiar with using BigQuery and Dataprep to analyze and transform your datasets. This is the first course of the From Data to Insights with Google Cloud series. After completing this course, enroll in the Creating New BigQuery Datasets and Visualizing Insights course.
With Google Slides, you can create and present professional presentations for sales, projects, training modules, and much more. Google Slides presentations are stored safely in the cloud. You build presentations right in your web browser—no special software is required. Even better, multiple people can work on your slides at the same time, you can see people’s changes as they make them, and every change is automatically saved. You will learn how to open Google Slides, create a blank presentation, and create a presentation from a template. You will explore themes, layout options, and how to add and format content, and speaker notes in your presentations. You will learn how to enhance your slides by adding tables, images, charts, and more. You will also learn how to use slide transitions and object animations in your presentation for visual effects. We will discuss how to organize slides and explore some of the options, including duplicating and ordering your slides, importi…
这是一节入门级微学习课程,探讨什么是大型语言模型 (LLM)、适合的应用场景以及如何使用提示调整来提升 LLM 性能,还介绍了可以帮助您开发自己的 Gen AI 应用的各种 Google 工具。
In this course we will introduce you to Google Sheets, Google’s cloud-based spreadsheet software, included with Google Workspace. With Google Sheets, you can create and edit spreadsheets directly in your web browser—no special software is required. Multiple people can work simultaneously, you can see people’s changes as they make them, and every change is saved automatically. You will learn how to open Google Sheets, create a blank spreadsheet, and create a spreadsheet from a template. You will add, import, sort, filter and format your data using Google Sheets and learn how to work across different file types. Formulas and functions allow you to make quick calculations and better use your data. We will look at creating a basic formula, using functions, and referencing data. You will also learn how to add a chart to your spreadsheet. Google Sheets spreadsheets are easy to share. We will look at the different ways you can share with others. We will also discuss how to track changes…
Discover the ease of an integrated multi-cloud database with data services that simplify database deployment and management. Get practical experience using MongoDB Atlas on Google Cloud and unlock the power of a flexible document data model that makes data easier to work with in a global, scalable, fully-managed, and elastic service. Come expecting to learn, Building a Cloud Run application that leverages Atlas App Services to call Cloud Natural Language Processing and carry out sentiment analysis, Running a MongoDB Database in Kubernetes with StatefulSets and Ingesting New Datasets into BigQuery
The Google Cloud Computing Foundations courses are for individuals with little to no background or experience in cloud computing. They provide an overview of concepts central to cloud basics, big data, and machine learning, and where and how Google Cloud fits in. By the end of the series of courses, learners will be able to articulate these concepts and demonstrate some hands-on skills. The courses should be completed in the following order: 1. Google Cloud Computing Foundations: Cloud Computing Fundamentals 2. Google Cloud Computing Foundations: Infrastructure in Google Cloud 3. Google Cloud Computing Foundations: Networking and Security in Google Cloud 4. Google Cloud Computing Foundations: Data, ML, and AI in Google Cloud This first course provides an overview of cloud computing, ways to use Google Cloud, and different compute options.
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
Complete the introductory Create and Manage Cloud Spanner Instances skill badge to demonstrate skills in the following: creating and interacting with Cloud Spanner instances and databases; loading Cloud Spanner databases using various techniques; backing up Cloud Spanner databases; defining schemas and understanding query plans; and deploying a Modern Web App connected to a Cloud Spanner instance.
In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.
This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
This course explores what ML is and what problems it can solve. The course also discusses best practices for implementing machine learning. You’re introduced to Vertex AI, a unified platform to quickly build, train, and deploy AutoML machine learning models. The course discusses the five phases of converting a candidate use case to be driven by machine learning, and why it’s important to not skip them. The course ends with recognizing the biases that ML can amplify and how to recognize them.
完成中级技能徽章课程利用 BigQuery ML 构建预测模型时的数据工程处理, 展示自己在以下方面的技能:利用 Dataprep by Trifacta 构建 BigQuery 数据转换流水线; 利用 Cloud Storage、Dataflow 和 BigQuery 构建提取、转换和加载 (ETL) 工作流; 以及利用 BigQuery ML 构建机器学习模型。 技能徽章是由 Google Cloud 颁发的专属数字徽章,旨在认可 您对 Google Cloud 产品与服务的熟练度;您需要在 交互式实操环境中参加考核,证明自己运用所学知识的能力后才能获得此徽章。完成技能徽章课程和 作为最终评估的实验室挑战赛,即可获得数字徽章, 在您的人际圈中炫出自己的技能。
完成入门级技能徽章课程在 Google Cloud 上为机器学习 API 准备数据,展示以下技能: 使用 Dataprep by Trifacta 清理数据、在 Dataflow 中运行数据流水线、在 Dataproc 中创建集群和运行 Apache Spark 作业,以及调用机器学习 API,包括 Cloud Natural Language API、Google Cloud Speech-to-Text API 和 Video Intelligence API。 技能徽章是由 Google Cloud 颁发的专属数字徽章,旨在认可您在 Google Cloud 产品与服务方面的熟练度; 您需要在交互式实操环境中参加考核,证明自己运用所学知识的能力后才能获得。完成此技能徽章课程和作为最终评估的实验室挑战赛, 获得技能徽章,在您的人际圈中炫出自己的技能。
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.
完成中级技能徽章课程使用 BigQuery 构建数据仓库,展示以下技能: 联接数据以创建新表、排查联接故障、使用并集附加数据、创建日期分区表, 以及在 BigQuery 中使用 JSON、数组和结构体。 技能徽章是 Google Cloud 颁发的专属数字徽章, 旨在认可您在 Google Cloud 产品与服务方面的熟练度; 您需要在交互式实操环境中参加考核,证明自己运用所学知识的能力后 才能获得。完成此技能徽章课程和作为最终评估的实验室挑战赛, 获得数字徽章,在您的人际圈中炫出自己的技能。
Processing streaming data is becoming increasingly popular as streaming enables businesses to get real-time metrics on business operations. This course covers how to build streaming data pipelines on Google Cloud. Pub/Sub is described for handling incoming streaming data. The course also covers how to apply aggregations and transformations to streaming data using Dataflow, and how to store processed records to BigQuery or Bigtable for analysis. Learners get hands-on experience building streaming data pipeline components on Google Cloud by using QwikLabs.
Data pipelines typically fall under one of the Extract and Load (EL), Extract, Load and Transform (ELT) or Extract, Transform and Load (ETL) paradigms. This course describes which paradigm should be used and when for batch data. Furthermore, this course covers several technologies on Google Cloud for data transformation including BigQuery, executing Spark on Dataproc, pipeline graphs in Cloud Data Fusion and serverless data processing with Dataflow. Learners get hands-on experience building data pipeline components on Google Cloud using Qwiklabs.
“Google Cloud 基础知识:核心基础设施”介绍在使用 Google Cloud 时会遇到的重要概念和术语。本课程通过视频和实操实验来介绍并比较 Google Cloud 的多种计算和存储服务,并提供重要的资源和政策管理工具。
The two key components of any data pipeline are data lakes and warehouses. This course highlights use-cases for each type of storage and dives into the available data lake and warehouse solutions on Google Cloud in technical detail. Also, this course describes the role of a data engineer, the benefits of a successful data pipeline to business operations, and examines why data engineering should be done in a cloud environment. This is the first course of the Data Engineering on Google Cloud series. After completing this course, enroll in the Building Batch Data Pipelines on Google Cloud course.