Gagan Ichake
Учасник із 2022
Діамантова ліга
Кількість балів: 22975
Учасник із 2022
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
In this course you will get hands-on in order to work through real-world challenges faced when building streaming data pipelines. The primary focus is on managing continuous, unbounded data with Google Cloud products.
This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
In this intermediate course, you will learn to design, build, and optimize robust batch data pipelines on Google Cloud. Moving beyond fundamental data handling, you will explore large-scale data transformations and efficient workflow orchestration, essential for timely business intelligence and critical reporting. Get hands-on practice using Dataflow for Apache Beam and Serverless for Apache Spark (Dataproc Serverless) for implementation, and tackle crucial considerations for data quality, monitoring, and alerting to ensure pipeline reliability and operational excellence. A basic knowledge of data warehousing, ETL/ELT, SQL, Python, and Google Cloud concepts is recommended.
Complete the introductory Derive Insights from BigQuery Data skill badge course to demonstrate skills in the following: Write SQL queries.Query public tables.Load sample data into BigQuery.Troubleshoot common syntax errors with the query validator in BigQuery.Create reports in Looker Studio by connecting to BigQuery data.
While the traditional approaches of using data lakes and data warehouses can be effective, they have shortcomings, particularly in large enterprise environments. This course introduces the concept of a data lakehouse and the Google Cloud products used to create one. A lakehouse architecture uses open-standard data sources and combines the best features of data lakes and data warehouses, which addresses many of their shortcomings.
Під час курсу ви зможете ознайомитися з продуктами й сервісами Google Cloud для роботи з масивами даних і машинним навчанням, які підтримують життєвий цикл роботи з даними для тренування моделей штучного інтелекту. У курсі розглядаються процеси, проблеми й переваги створення конвеєру масиву даних і моделей машинного навчання з Vertex AI у Google Cloud.
This course helps learners create a study plan for the PCA (Professional Cloud Architect) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
Welcome to the Getting Started with Google Kubernetes Engine course. If you're interested in Kubernetes, a software layer that sits between your applications and your hardware infrastructure, then you’re in the right place! Google Kubernetes Engine brings you Kubernetes as a managed service on Google Cloud. The goal of this course is to introduce the basics of Google Kubernetes Engine, or GKE, as it’s commonly referred to, and how to get applications containerized and running in Google Cloud. The course starts with a basic introduction to Google Cloud, and is then followed by an overview of containers and Kubernetes, Kubernetes architecture, and Kubernetes operations.
This course equips students to build highly reliable and efficient solutions on Google Cloud using proven design patterns. It is a continuation of the Architecting with Google Compute Engine or Architecting with Google Kubernetes Engine courses and assumes hands-on experience with the technologies covered in either of those courses. Through a combination of presentations, design activities, and hands-on labs, participants learn to define and balance business and technical requirements to design Google Cloud deployments that are highly reliable, highly available, secure, and cost-effective.
This course introduces participants to the strategies to migrate from a source environment to Google Cloud. Participants are introduced to Google Cloud's fundamental concepts and more in depth topics, like creating virtual machines, configuring networks and managing access and identities. The course then covers the installation and migration process of Migrate for Compute Engine, including special features like test clones and wave migrations.
Complete the intermediate Build Infrastructure with Terraform on Google Cloud skill badge to demonstrate skills in the following: Infrastructure as Code (IaC) principles using Terraform, provisioning and managing Google Cloud resources with Terraform configurations, effective state management (local and remote), and modularizing Terraform code for reusability and organization.
Пройдіть вступний кваліфікаційний курс Налаштування розподілу навантаження в Compute Engine, щоб продемонструвати свої навички написання команд gcloud і використання Cloud Shell, створення й розгортання віртуальних машин у Compute Engine, а також налаштування мережі й розподілювачів навантаження HTTP. Кваліфікаційний значок – це ексклюзивний цифровий значок від Google Cloud, який засвідчує, що ви знаєтеся на продуктах і сервісах цієї платформи й можете застосовувати ці знання в інтерактивному практичному середовищі. Щоб отримати кваліфікаційний значок і показати його колегам, пройдіть цей курс і підсумковий тест.
Щоб отримати кваліфікаційний значок, пройдіть курс Налаштування середовища для розробки додатка в Google Cloud. У ньому ви навчитеся створювати й підключати хмарну інфраструктуру, спрямовану на зберігання даних, за допомогою базових можливостей таких технологій, як Cloud Storage, система керування ідентифікацією і доступом, Cloud Functions та Pub/Sub.
Complete these 6 hands-on Google Cloud skills challenges by October 13th to earn a special digital badge, plus a no-cost e-copy of Priyanka Vergadia’s best selling Visualizing Google Cloud book!
This accelerated on-demand course introduces participants to the comprehensive and flexible infrastructure and platform services provided by Google Cloud. Through a combination of video lectures, demos, and hands-on labs, participants explore and deploy solution elements, including securely interconnecting networks, load balancing, autoscaling, infrastructure automation and managed services.
This accelerated on-demand course introduces participants to the comprehensive and flexible infrastructure and platform services provided by Google Cloud with a focus on Compute Engine. Through a combination of video lectures, demos, and hands-on labs, participants explore and deploy solution elements, including infrastructure components such as networks, systems and applications services. This course also covers deploying practical solutions including customer-supplied encryption keys, security and access management, quotas and billing, and resource monitoring.
This accelerated on-demand course introduces participants to the comprehensive and flexible infrastructure and platform services provided by Google Cloud with a focus on Compute Engine. Through a combination of video lectures, demos, and hands-on labs, participants explore and deploy solution elements, including infrastructure components such as networks, virtual machines and applications services. You will learn how to use the Google Cloud through the console and Cloud Shell. You'll also learn about the role of a cloud architect, approaches to infrastructure design, and virtual networking configuration with Virtual Private Cloud (VPC), Projects, Networks, Subnetworks, IP addresses, Routes, and Firewall rules.
Курс "Знайомство з Google Cloud: основна інфраструктура" охоплює важливі поняття й терміни щодо використання Google Cloud. Переглядаючи відео й виконуючи практичні завдання, слухачі ознайомляться з різними сервісами Google Cloud для обчислень і зберігання даних, а також важливими ресурсами й інструментами для керування правилами. Крім того, вони зможуть їх порівнювати.