Dominique Lambert
成为会员时间:2021
白银联赛
50328 积分
成为会员时间:2021
在本课程中,您将学习如何使用 Google 智能体开发套件构建复杂的多智能体系统。您将构建搭载工具的智能体,利用父子层级关系和工作流进行连接,以此定义它们的交互方式。您将在本地运行智能体,将其部署到 Vertex AI Agent Engine 并作为托管式智能体流运行,基础设施决策和资源扩缩则由 Agent Engine 处理。请注意,这些实验基于此产品的预发布版本。在进行维护更新时,这些实验可能会出现一些延迟。
NotebookLM is an AI-powered collaborator that helps you do your best thinking. After uploading your documents, NotebookLM becomes an instant expert in those sources so you can read, take notes, and collaborate with it to refine and organize your ideas. NotebookLM Pro gives you everything already included with NotebookLM, as well as higher utilization limits, access to premium features, and additional sharing options and analytics.
Gemini Enterprise 结合了 Google 在搜索和 AI 领域的专长。它是一款强大的工具,让员工只需通过一个搜索栏,就能从文档库、邮件、聊天消息、工单系统及其他数据源中查找具体信息。Gemini Enterprise 助理还能帮助进行头脑风暴、开展研究、生成文档大纲并执行其他操作,比如邀请同事参加某个日历活动。因此它能加快知识型工作的进度并提升协作效率。(请注意,Gemini Enterprise 以前称为 Google Agentspace,本课程中可能会提及以前的产品名称。)
完成中级技能徽章课程使用 Vertex AI 中的 Gemini API 探索生成式 AI,展示自己在以下方面的技能: 文本生成技能、用于增强内容创作能力的图像和视频分析技能,以及在 Gemini API 中应用函数调用技术的技能。 了解如何运用先进的 Gemini 技术、探索多模态内容生成方法,并扩展 AI 赋能项目的功能。
(This course was previously named Multimodal Prompt Engineering with Gemini and PaLM) This course teaches how to use Vertex AI Studio, a Google Cloud console tool for rapidly prototyping and testing generative AI models. You learn to test sample prompts, design your own prompts, and customize foundation models to handle tasks that meet your application's needs. Whether you are looking for text, chat, code, image or speech generative experiences Vertex AI Studio offers you an interface to work with and APIs to integrate your production application.
This course on Integrate Vertex AI Search and Conversation into Voice and Chat Apps is composed of a set of labs to give you a hands on experience to interacting with new Generative AI technologies. You will learn how to create end-to-end search and conversational experiences by following examples. These technologies complement predefined intent-based chat experiences created in Dialogflow with LLM-based, generative answers that can be based on your own data. Also, they allow you to porvide enterprise-grade search experiences for internal and external websites to search documents, structure data and public websites.
In this course, you'll use text embeddings for tasks like classification, outlier detection, text clustering and semantic search. You'll combine semantic search with the text generation capabilities of an LLM to build Retrieval Augmented Generation (RAG) solutions, such as for question-answering systems, using Google Cloud's Vertex AI and Google Cloud databases.
This course explores Google Cloud technologies to create and generate embeddings. Embeddings are numerical representations of text, images, video and audio, and play a pivotal role in many tasks that involve the identification of similar items, like Google searches, online shopping recommendations, and personalized music suggestions. Specifically, you’ll use embeddings for tasks like classification, outlier detection, clustering and semantic search. You’ll combine semantic search with the text generation capabilities of an LLM to build Retrieval Augmented Generation (RAG) systems and question-answering solutions, on your own proprietary data using Google Cloud’s Vertex AI.
在本次课程中,探索 AI 赋能的搜索技术、工具和应用。学习利用向量嵌入的语义搜索、融合语义和关键字的混合搜索方法,以及检索增强生成 (RAG) 技术,以打造基于事实的 AI 智能体,尽可能减少 AI 幻觉。获取 Vertex AI Vector Search 实战经验,打造您自己的智能搜索引擎。
This course will help ML Engineers, Developers, and Data Scientists implement Large Language Models for Generative AI use cases with Vertex AI. The first two modules of this course contain links to videos and prerequisite course materials that will build your knowledge foundation in Generative AI. Please do not skip these modules. The advanced modules in this course assume you have completed these earlier modules.
本课程教您如何使用深度学习来创建图片标注模型。您将了解图片标注模型的不同组成部分,例如编码器和解码器,以及如何训练和评估模型。学完本课程,您将能够自行创建图片标注模型并用来生成图片说明。
本课程向您介绍扩散模型。这类机器学习模型最近在图像生成领域展现出了巨大潜力。扩散模型的灵感来源于物理学,特别是热力学。过去几年内,扩散模型成为热门研究主题并在整个行业开始流行。Google Cloud 上许多先进的图像生成模型和工具都是以扩散模型为基础构建的。本课程向您介绍扩散模型背后的理论,以及如何在 Vertex AI 上训练和部署此类模型。
本课程向您介绍 Transformer 架构和 Bidirectional Encoder Representations from Transformers (BERT) 模型。您将了解 Transformer 架构的主要组成部分,例如自注意力机制,以及该架构如何用于构建 BERT 模型。您还将了解可以使用 BERT 的不同任务,例如文本分类、问答和自然语言推理。完成本课程估计需要大约 45 分钟。
本课程简要介绍了编码器-解码器架构,这是一种功能强大且常见的机器学习架构,适用于机器翻译、文本摘要和问答等 sequence-to-sequence 任务。您将了解编码器-解码器架构的主要组成部分,以及如何训练和部署这些模型。在相应的实验演示中,您将在 TensorFlow 中从头编写简单的编码器-解码器架构实现代码,以用于诗歌生成。
本课程将向您介绍注意力机制,这是一种强大的技术,可令神经网络专注于输入序列的特定部分。您将了解注意力的工作原理,以及如何使用它来提高各种机器学习任务的性能,包括机器翻译、文本摘要和问题解答。
Earn a skill badge by passing the final quiz, you'll demonstrate your understanding of foundational concepts in generative AI. A skill badge is a digital badge issued by Google Cloud in recognition of your knowledge of Google Cloud products and services. Share your skill badge by making your profile public and adding it to your social media profile.
此课程将探索如何使用 AI 功能套件 Gemini in BigQuery 为“数据到 AI”工作流提供助力。其中涉及到的功能包括数据探索和准备、代码生成和问题排查,以及工作流发现和可视化。此课程包含概念解释、真实使用场景以及实操实验等内容,可帮助数据从业者提升效率并加快流水线开发速度。
Text Prompt Engineering Techniques introduces you to consider different strategic approaches & techniques to deploy when writing prompts for text-based generative AI tasks.
本课程介绍 Vertex AI Studio,这是一种用于与生成式 AI 模型交互、围绕业务创意进行原型设计并在生产环境中落地的工具。通过沉浸式应用场景、富有吸引力的课程和实操实验,您将探索从提示到产品的整个生命周期,了解如何将 Vertex AI Studio 用于多模态 Gemini 应用、提示设计、提示工程和模型调优。本课程的目的在于帮助您利用 Vertex AI Studio,在自己的项目中充分发掘生成式 AI 的潜力。
随着企业对人工智能和机器学习的应用越来越广泛,以负责任的方式构建这些技术也变得更加重要。但对很多企业而言,真正践行 Responsible AI 并非易事。如果您有意了解如何在组织内践行 Responsible AI,本课程正适合您。 本课程将介绍 Google Cloud 目前如何践行 Responsible AI,以及从中总结的最佳实践和经验教训,便于您以此为框架构建自己的 Responsible AI 方法。
这是一节入门级微课程,旨在解释什么是负责任的 AI、它的重要性,以及 Google 如何在自己的产品中实现负责任的 AI。此外,本课程还介绍了 Google 的 7 个 AI 开发原则。
A Business Leader in Generative AI can articulate the capabilities of core cloud Generative AI products and services and understand how they benefit organizations. This course provides an overview of the types of opportunities and challenges that companies often encounter in their digital transformation journey and how they can leverage Google Cloud's generative AI products to overcome these challenges.
In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.
完成入门技能徽章课程使用 Dataplex 构建数据网格,展示以下方面的技能:使用 Dataplex 构建数据网格, 以在 Google Cloud 上实现数据安全、治理和发现。您将在 Dataplex 中练习和测试自己在标记资产、分配 IAM 角色和评估数据质量方面的技能。
In this intermediate course, you will learn to design, build, and optimize robust batch data pipelines on Google Cloud. Moving beyond fundamental data handling, you will explore large-scale data transformations and efficient workflow orchestration, essential for timely business intelligence and critical reporting. Get hands-on practice using Dataflow for Apache Beam and Serverless for Apache Spark (Dataproc Serverless) for implementation, and tackle crucial considerations for data quality, monitoring, and alerting to ensure pipeline reliability and operational excellence. A basic knowledge of data warehousing, ETL/ELT, SQL, Python, and Google Cloud concepts is recommended.
This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
This workload aims to upskill Google Cloud Partners about what Google Cloud VMware Engine is, design considerations, and basic implementation. It will also cover interoperability with on-premises vSphere and real world use cases for deployment. Finally, it will cover how VMware Engine compares to other cloud providers' offerings of VMware as a Service and future plans for VMware Engine.
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Learners will get hands-on practice using Vertex AI Feature Store's streaming ingestion at the SDK layer.
This course introduces participants to MLOps tools and best practices for deploying, evaluating, monitoring and operating production ML systems on Google Cloud. MLOps is a discipline focused on the deployment, testing, monitoring, and automation of ML systems in production. Machine Learning Engineering professionals use tools for continuous improvement and evaluation of deployed models. They work with (or can be) Data Scientists, who develop models, to enable velocity and rigor in deploying the best performing models.
This course helps developers customize Chronicle and augment its abilities with third party integrations.
This course helps you understand how to use Chronicle to properly handle security incidents.
This course will familiarize you with the core functionality of Chronicle, including the user interface, connections, and settings.
Get hands-on experience applying and building rules for Chronicle. You learn what YARA-L is and how to customize & create event rules.
Learn the technical aspects you need to know about Chronicle and how it can help you detect and action threats.
Earn a DRI badge by completing the Infra Foundations - Implementing Least Privilege for Service Accounts quest, where you demonstrate your capabilities to manage service accounts, assign IAM roles, setting up and using impersonation and implementing logging sinks that target GCS buckets. When you complete this activity, you can earn the badge displayed above! View all the badges you have earned by visiting your profile page.
While the traditional approaches of using data lakes and data warehouses can be effective, they have shortcomings, particularly in large enterprise environments. This course introduces the concept of a data lakehouse and the Google Cloud products used to create one. A lakehouse architecture uses open-standard data sources and combines the best features of data lakes and data warehouses, which addresses many of their shortcomings.
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.
完成在 Google Cloud 上使用 Terraform 构建基础设施技能徽章中级课程, 展示您在以下方面的技能:在使用 Terraform 时遵循基础设施即代码 (IaC) 原则;利用 Terraform 配置 来预配和管理 Google Cloud 资源;管理有效状态(本地和远程);以及将 Terraform 代码模块化,以方便重复使用和整理。
本课程指导学员运用久经考验的设计模式在 Google Cloud 上构建高度可靠且高效的解决方案。它是“Google Compute Engine 架构设计”或“Google Kubernetes Engine 架构设计”课程的延续,并假定您有使用其中任何一门课程所涵盖技术的实践经验。通过一系列演示、设计活动和动手实验,学员可以了解如何定义及平衡业务要求和技术要求,以便设计可靠性和可用性高、安全且经济实惠的 Google Cloud 部署。
完成开发 Google Cloud 网络课程,赢取技能徽章。在此课程中,您将学习 部署和监控应用的多种方法,包括执行以下任务的方法:探索 IAM 角色并添加/移除 项目访问权限、创建 VPC 网络、部署和监控 Compute Engine 虚拟机、 编写 SQL 查询、在 Compute Engine 中部署和监控虚拟机,以及使用 Kubernetes 通过多种部署方法部署应用。
欢迎学习“Google Kubernetes Engine 使用入门”课程。Kubernetes 是位于应用和硬件基础架构之间的软件层,如果您对 Kubernetes 感兴趣,那就来对地方了!Google Kubernetes Engine 将 Kubernetes 作为 Google Cloud 上的代管式服务提供给您使用。 本课程的目标是介绍 Google Kubernetes Engine(通常称为 GKE)的基础知识,以及将应用容器化并在 Google Cloud 中运行的方法。本课程首先介绍 Google Cloud 的基础知识,然后概述容器、Kubernetes、Kubernetes 架构以及 Kubernetes 操作。
完成入门级技能徽章课程为 Compute Engine 实现云负载均衡,展示以下方面的技能: 在 Compute Engine 中创建和部署虚拟机 以及配置网络和应用负载均衡器。
完成“在 Google Cloud 上设置应用开发环境”课程,赢取技能徽章;通过该课程,您将了解如何使用以下技术的基本功能来构建和连接以存储为中心的云基础设施: Cloud Storage、Identity and Access Management、Cloud Functions 和 Pub/Sub。
这是一套自助式速成课程,向学员介绍 Google Cloud 提供的灵活全面的基础架构和平台服务。学员将通过一系列视频讲座、演示和实操实验,探索和部署各种解决方案元素,包括安全互连网络、负载均衡、自动扩缩、基础架构自动化和代管式服务。
这门自助式速成课程向学员介绍 Google Cloud 提供的灵活全面的基础架构和平台服务,着重介绍了 Compute Engine。学员将通过一系列视频讲座、演示和动手实验,探索和部署各种解决方案元素,包括网络、系统和应用服务等基础架构组件。本课程的内容还包括如何部署实用的解决方案,包括客户提供的加密密钥、安全和访问权限管理、配额和结算,以及资源监控。
这门自助式速成课程向学员介绍 Google Cloud 提供的灵活全面的基础架构和平台服务,其中着重介绍了 Compute Engine。学员将通过一系列视频讲座、演示和动手实验,探索和部署各种解决方案元素,包括网络、虚拟机和应用服务等基础架构组件。您将学习如何通过控制台和 Cloud Shell 使用 Google Cloud。您还将了解云架构师角色、基础架构设计方法以及虚拟网络配置和虚拟私有云 (VPC)、项目、网络、子网、IP 地址、路由及防火墙规则。
“Google Cloud 基础知识:核心基础设施”介绍在使用 Google Cloud 时会遇到的重要概念和术语。本课程通过视频和实操实验来介绍并比较 Google Cloud 的多种计算和存储服务,并提供重要的资源和政策管理工具。
Welcome Gamers! Today's game is all about experimenting with Big Query for Machine Learning! Use real life case studies to learn various concepts of BQML and have fun. Take labs to earn points. The faster you complete the lab objectives, the higher your score.