Guillermo Hamity
Учасник із 2024
Учасник із 2024
This course helps learners create a study plan for the PMLE (Professional Machine Learning Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Machine Learning Engineering professionals use tools for continuous improvement and evaluation of deployed models. They work with (or can be) Data Scientists, who develop models, to enable velocity and rigor in deploying the best performing models.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
Processing streaming data is becoming increasingly popular as streaming enables businesses to get real-time metrics on business operations. This course covers how to build streaming data pipelines on Google Cloud. Pub/Sub is described for handling incoming streaming data. The course also covers how to apply aggregations and transformations to streaming data using Dataflow, and how to store processed records to BigQuery or Bigtable for analysis. Learners get hands-on experience building streaming data pipeline components on Google Cloud by using QwikLabs.
Пройдіть вступний кваліфікаційний курс Підготовка даних для інтерфейсів API машинного навчання в Google Cloud, щоб продемонструвати свої навички щодо очистки даних за допомогою сервісу Dataprep by Trifacta, запуску конвеєрів даних у Dataflow, створення кластерів і запуску завдань Apache Spark у Dataproc, а також виклику API машинного навчання, зокрема Cloud Natural Language API, Google Cloud Speech-to-Text API і Video Intelligence API. Кваліфікаційний значок – це ексклюзивна цифрова відзнака, яка підтверджує, що ви вмієте працювати з продуктами й сервісами Google Cloud і можете застосовувати ці знання в інтерактивному практичному середовищі. Щоб отримати кваліфікаційний значок і показати його колегам, пройдіть цей курс і підсумковий тест.
Data pipelines typically fall under one of the Extract and Load (EL), Extract, Load and Transform (ELT) or Extract, Transform and Load (ETL) paradigms. This course describes which paradigm should be used and when for batch data. Furthermore, this course covers several technologies on Google Cloud for data transformation including BigQuery, executing Spark on Dataproc, pipeline graphs in Cloud Data Fusion and serverless data processing with Dataflow. Learners get hands-on experience building data pipeline components on Google Cloud using Qwiklabs.
The two key components of any data pipeline are data lakes and warehouses. This course highlights use-cases for each type of storage and dives into the available data lake and warehouse solutions on Google Cloud in technical detail. Also, this course describes the role of a data engineer, the benefits of a successful data pipeline to business operations, and examines why data engineering should be done in a cloud environment. This is the first course of the Data Engineering on Google Cloud series. After completing this course, enroll in the Building Batch Data Pipelines on Google Cloud course.
Під час курсу ви зможете ознайомитися з продуктами й сервісами Google Cloud для роботи з масивами даних і машинним навчанням, які підтримують життєвий цикл роботи з даними для тренування моделей штучного інтелекту. У курсі розглядаються процеси, проблеми й переваги створення конвеєру масиву даних і моделей машинного навчання з Vertex AI у Google Cloud.