Jordan Victor Scher
成为会员时间:2024
钻石联赛
41710 积分
成为会员时间:2024
生成式 AI 应用可以提供大语言模型 (LLM) 问世前几乎不可能实现的全新用户体验。作为应用开发者,您要如何利用生成式 AI 在 Google Cloud 上构建更具吸引力且功能强大的应用? 在本课程中,您将了解生成式 AI 应用,以及如何利用提示设计和检索增强生成 (RAG) 技术,构建使用 LLM 的强大应用。您将了解可用于生产用途且适合生成式 AI 应用的架构,并构建一个基于 LLM 和 RAG 的聊天应用。
本课程介绍 AI 隐私保护和安全方面的重要主题,还将探索使用 Google Cloud 产品和开源工具实施建议的 AI 隐私保护和安全实践的实用方法和工具。
本课程能让机器学习从业者掌握评估生成式和预测式 AI 模型的基本工具、方法和最佳实践。要确保机器学习系统在实际运用中提供可靠、准确、高效的结果,做好模型评估至关重要。 学员将深入了解各项评估指标、方法及如何在不同模型类型和任务中适当应用这些指标和方法。课程将着重介绍生成式 AI 模型带来的独特挑战,并提供有效解决这些挑战的策略。通过利用 Google Cloud 的 Vertex AI Platform,学员可学习如何在模型选择、优化和持续监控工作中实施卓有成效的评估流程。
本课程介绍了 AI 可解释性和透明度的相关概念,探讨了 AI 透明度对于开发者和工程师的重要性。同时探索了有助于在数据和 AI 模型中实现可解释性和透明度的实用方法及工具。
本课程介绍了 Responsible AI 的概念和 AI 原则,还介绍了在 AI/机器学习实践中识别公平性与偏见以及减少偏见的实用技巧,同时探索了使用 Google Cloud 产品和开源工具来实施 Responsible AI 最佳实践的实用方法和工具。
这是一节入门级微课程,旨在解释什么是生成式 AI、它的用途以及与传统机器学习方法的区别。该课程还介绍了可以帮助您开发自己的生成式 AI 应用的各种 Google 工具。
这是一节入门级微学习课程,探讨什么是大型语言模型 (LLM)、适合的应用场景以及如何使用提示调整来提升 LLM 性能,还介绍了可以帮助您开发自己的 Gen AI 应用的各种 Google 工具。
本课程致力于为您提供所需的知识和工具,让您能够了解 MLOps 团队在部署和管理生成式 AI 模型以及探索 Vertex AI 如何帮助 AI 团队简化 MLOps 流程时面临的独特挑战,并帮助您在生成式 AI 项目中取得成功。
This course is an introduction to Vertex AI Notebooks, which are Jupyter notebook-based environments that provide a unified platform for the entire machine learning workflow, from data preparation to model deployment and monitoring. The course covers the following topics: (1) The different types of Vertex AI Notebooks and their features and (2) How to create and manage Vertex AI Notebooks.
本课程介绍 Google Cloud 的 AI 和机器学习 (ML) 能力,重点讲解如何开发生成式和预测式 AI 项目。本课程将探讨“数据到 AI”全生命周期中的多种技术、产品和工具,并通过互动练习帮助数据科学家、AI 开发者和机器学习工程师提升专业能力。
Learn how to use NotebookLM to create a personalized study guide for the Professional Machine Learning Engineer certification exam (PMLE). You'll review NotebookLM features, create a notebook, and use the study guide to practice for a certification exam.
随着企业对人工智能和机器学习的应用越来越广泛,以负责任的方式构建这些技术也变得更加重要。但对很多企业而言,真正践行 Responsible AI 并非易事。如果您有意了解如何在组织内践行 Responsible AI,本课程正适合您。 本课程将介绍 Google Cloud 目前如何践行 Responsible AI,以及从中总结的最佳实践和经验教训,便于您以此为框架构建自己的 Responsible AI 方法。
完成入门级技能徽章课程在 Google Cloud 上为机器学习 API 准备数据,展示以下技能: 使用 Dataprep by Trifacta 清理数据、在 Dataflow 中运行数据流水线、在 Dataproc 中创建集群和运行 Apache Spark 作业,以及调用机器学习 API,包括 Cloud Natural Language API、Google Cloud Speech-to-Text API 和 Video Intelligence API。
完成中级技能徽章课程利用 BigQuery ML 构建预测模型时的数据工程处理, 展示自己在以下方面的技能:利用 Dataprep by Trifacta 构建 BigQuery 数据转换流水线; 利用 Cloud Storage、Dataflow 和 BigQuery 构建提取、转换和加载 (ETL) 工作流; 以及利用 BigQuery ML 构建机器学习模型。
完成中级技能徽章课程使用 BigQuery 构建数据仓库,展示以下技能: 联接数据以创建新表、排查联接故障、使用并集附加数据、创建日期分区表, 以及在 BigQuery 中使用 JSON、数组和结构体。
完成入门技能徽章课程使用 Dataplex 构建数据网格,展示以下方面的技能:使用 Dataplex 构建数据网格, 以在 Google Cloud 上实现数据安全、治理和发现。您将在 Dataplex 中练习和测试自己在标记资产、分配 IAM 角色和评估数据质量方面的技能。
This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
In the last installment of the Dataflow course series, we will introduce the components of the Dataflow operational model. We will examine tools and techniques for troubleshooting and optimizing pipeline performance. We will then review testing, deployment, and reliability best practices for Dataflow pipelines. We will conclude with a review of Templates, which makes it easy to scale Dataflow pipelines to organizations with hundreds of users. These lessons will help ensure that your data platform is stable and resilient to unanticipated circumstances.
In this second installment of the Dataflow course series, we are going to be diving deeper on developing pipelines using the Beam SDK. We start with a review of Apache Beam concepts. Next, we discuss processing streaming data using windows, watermarks and triggers. We then cover options for sources and sinks in your pipelines, schemas to express your structured data, and how to do stateful transformations using State and Timer APIs. We move onto reviewing best practices that help maximize your pipeline performance. Towards the end of the course, we introduce SQL and Dataframes to represent your business logic in Beam and how to iteratively develop pipelines using Beam notebooks.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
In this course you will get hands-on in order to work through real-world challenges faced when building streaming data pipelines. The primary focus is on managing continuous, unbounded data with Google Cloud products.
In this intermediate course, you will learn to design, build, and optimize robust batch data pipelines on Google Cloud. Moving beyond fundamental data handling, you will explore large-scale data transformations and efficient workflow orchestration, essential for timely business intelligence and critical reporting. Get hands-on practice using Dataflow for Apache Beam and Serverless for Apache Spark (Dataproc Serverless) for implementation, and tackle crucial considerations for data quality, monitoring, and alerting to ensure pipeline reliability and operational excellence. A basic knowledge of data warehousing, ETL/ELT, SQL, Python, and Google Cloud concepts is recommended.
While the traditional approaches of using data lakes and data warehouses can be effective, they have shortcomings, particularly in large enterprise environments. This course introduces the concept of a data lakehouse and the Google Cloud products used to create one. A lakehouse architecture uses open-standard data sources and combines the best features of data lakes and data warehouses, which addresses many of their shortcomings.