BARNABAS NGUNYI
成为会员时间:2022
白银联赛
9734 积分
成为会员时间:2022
瞭解如何使用 Agent Development Kit (ADK),建構複雜的 AI 代理並用於正式環境。本課程介紹 ADK 的開放原始碼框架,包括簡單的提示工程,以及程式碼優先的結構化軟體開發做法 (適用於企業級多代理系統)。
建立您的第一個 Gemini Enterprise 應用程式,獲得技能徽章!將各種資料來源連接到您的應用程式,建立強大的統合式搜尋和分析引擎。掌握 Deep Research 代理、多代理構思和 NotebookLM 等進階功能,進行重點分析。
瞭解 AI 代理如何發揮更高的業務影響力,包括根據您的 KPI 規劃要使用的代理類型,以及探索能解決實際瓶頸的用途。您也將認識各種無程式碼到高程式碼解決方案,瞭解 Gemini Enterprise 如何協助建構和自動調度合適的代理。
AI 代理是超越傳統大型語言模型 (LLM) 的重大轉變,這類服務不僅可以生成以文字為基礎的解決方案,還能自主加以執行。本課程介紹 AI 代理的基礎知識、與 LLM API 的差異,以及在實際應用上帶來的價值。本課程內容奠基於 Google 的代理白皮書,介紹實際編寫代理程式碼之前所需的理論基礎,非常適合有意透過目標導向自主行為 (不僅限於文字生成) 來理解 AI 系統的開發人員、架構師和技術決策者。加入社群論壇,提出問題並參與討論。
瞭解 AI 代理的概念,探索代理如何藉由自主行動及推論解決複雜問題。您將瞭解代理如何透過模型、工具和調度管理程序等技術架構,助您學習、規劃和實現目標。
這是一堂入門級的微學習課程,旨在探討大型語言模型 (LLM) 的定義和用途,並說明如何調整提示來提高 LLM 成效。此外,也會介紹多項 Google 工具,協助您自行開發生成式 AI 應用程式。
這個入門微學習課程主要說明生成式 AI 的定義和使用方式,以及此 AI 與傳統機器學習方法的差異。本課程也會介紹各項 Google 工具,協助您開發自己的生成式 AI 應用程式。
In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
In this course you will get hands-on in order to work through real-world challenges faced when building streaming data pipelines. The primary focus is on managing continuous, unbounded data with Google Cloud products.
In this intermediate course, you will learn to design, build, and optimize robust batch data pipelines on Google Cloud. Moving beyond fundamental data handling, you will explore large-scale data transformations and efficient workflow orchestration, essential for timely business intelligence and critical reporting. Get hands-on practice using Dataflow for Apache Beam and Serverless for Apache Spark (Dataproc Serverless) for implementation, and tackle crucial considerations for data quality, monitoring, and alerting to ensure pipeline reliability and operational excellence. A basic knowledge of data warehousing, ETL/ELT, SQL, Python, and Google Cloud concepts is recommended.
While the traditional approaches of using data lakes and data warehouses can be effective, they have shortcomings, particularly in large enterprise environments. This course introduces the concept of a data lakehouse and the Google Cloud products used to create one. A lakehouse architecture uses open-standard data sources and combines the best features of data lakes and data warehouses, which addresses many of their shortcomings.
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.
Gmail is Google’s cloud based email service that allows you to access your messages from any computer or device with just a web browser. In this course, you’ll learn how to compose, send and reply to messages. You will also explore some of the common actions that can be applied to a Gmail message, and learn how to organize your mail using Gmail labels. You will explore some common Gmail settings and features. For example, you will learn how to manage your own personal contacts and groups, customize your Gmail Inbox to suit your way of working, and create your own email signatures and templates. Google is famous for search. Gmail also includes powerful search and filtering. You will explore Gmail’s advanced search and learn how to filter messages automatically.
完成 建立 Google Cloud 網路 課程即可獲得技能徽章。這個課程將說明 部署及監控應用程式的多種方法,包括查看 IAM 角色及新增/移除 專案存取權、建立虛擬私有雲網路、部署及監控 Compute Engine VM、編寫 SQL 查詢、在 Compute Engine 部署及監控 VM,以及 使用 Kubernetes 透過多種方法部署應用程式。
完成「在 Google Cloud 使用 Terraform 建構基礎架構」技能徽章中階課程, 即可證明自己具備下列知識與技能:使用 Terraform 的基礎架構即程式碼 (IaC) 原則、運用 Terraform 設定佈建及管理 Google Cloud 資源、有效管理狀態 (本機和遠端),以及將 Terraform 程式碼模組化,以利重複使用和管理。
歡迎參加「開始使用 Google Kubernetes Engine」課程。Kubernetes 是位於應用程式和硬體基礎架構之間的軟體層。如果您對這項技術感興趣,這堂課程可以滿足您的需求。有了 Google Kubernetes Engine,您就能在 Google Cloud 中以代管服務的形式使用 Kubernetes。 本課程的目標在於介紹 Google Kubernetes Engine (常簡稱為 GKE) 的基本概念,以及如何將應用程式容器化,以便在 Google Cloud 中執行。課程首先會初步介紹 Google Cloud,隨後簡介容器、Kubernetes、Kubernetes 架構和 Kubernetes 作業。
這堂課程可讓參加人員瞭解如何使用確實有效的設計模式,在 Google Cloud 中打造相當可靠且效率卓越的解決方案。這堂課程接續了「設定 Google Compute Engine 架構」或「設定 Google Kubernetes Engine 架構」課程的內容,並假設參加人員曾實際運用上述任一課程涵蓋的技術。這堂課程結合了簡報、設計活動和實作研究室,可讓參加人員瞭解如何定義業務和技術需求,並在兩者之間取得平衡,設計出相當可靠、可用性高、安全又符合成本效益的 Google Cloud 部署項目。
只要修完「在 Google Cloud 設定應用程式開發環境」課程,就能獲得技能徽章。 在本課程中,您將學會如何使用以下技術的基本功能,建構和連結以儲存空間為中心的雲端基礎架構:Cloud Storage、Identity and Access Management、Cloud Functions 和 Pub/Sub。
This course introduces participants to the strategies to migrate from a source environment to Google Cloud. Participants are introduced to Google Cloud's fundamental concepts and more in depth topics, like creating virtual machines, configuring networks and managing access and identities. The course then covers the installation and migration process of Migrate for Compute Engine, including special features like test clones and wave migrations.
完成「在 Compute Engine 導入 Cloud Load Balancing」技能徽章入門課程,即可證明您具備下列技能: 在 Compute Engine 建立及部署虛擬機器, 以及設定網路和應用程式負載平衡器。
這堂隨選密集課程會向參加人員說明 Google Cloud 提供的全方位彈性基礎架構和平台服務。這堂課結合了視訊講座、示範和實作研究室,可讓參加人員探索及部署解決方案元素,包括安全地建立互連網路、負載平衡、自動調度資源、基礎架構自動化,以及代管服務。
這堂隨選密集課程會向參加人員說明 Google Cloud 提供的全方位彈性基礎架構和平台服務,並將重點放在 Compute Engine。這堂課程結合了視訊講座、示範和實作研究室,可讓參加人員探索及部署解決方案元素,例如網路、系統和應用程式服務等基礎架構元件。另外,這堂課也會介紹如何部署實用的解決方案,包括客戶提供的加密金鑰、安全性和存取權管理機制、配額與帳單,以及資源監控功能。
這堂隨選密集課程會向參加人員說明 Google Cloud 提供的全方位彈性基礎架構和平台服務,尤其側重於 Compute Engine。這堂課程結合了視訊講座、示範和實作研究室,可讓參加人員探索及部署解決方案元素,例如網路、虛擬機器和應用程式服務等基礎架構元件。您會瞭解如何透過控制台和 Cloud Shell 使用 Google Cloud。另外,您也能瞭解雲端架構師的職責、基礎架構設計方法,以及具備虛擬私有雲 (VPC)、專案、網路、子網路、IP 位址、路徑和防火牆規則的虛擬網路設定。
「Google Cloud 基礎知識:核心基礎架構」介紹了在使用 Google Cloud 時會遇到的重要概念和術語。本課程會透過影片和實作實驗室,介紹並比較 Google Cloud 的多種運算和儲存服務,同時提供重要的資源和政策管理工具。