Nishant Raghav
成为会员时间:2023
黄金联赛
28955 积分
成为会员时间:2023
完成入门技能徽章课程使用 Dataplex 构建数据网格,展示以下方面的技能:使用 Dataplex 构建数据网格, 以在 Google Cloud 上实现数据安全、治理和发现。您将在 Dataplex 中练习和测试自己在标记资产、分配 IAM 角色和评估数据质量方面的技能。
This 1-week, accelerated on-demand course builds upon Google Cloud Platform Big Data and Machine Learning Fundamentals. Through a combination of video lectures, demonstrations, and hands-on labs, you'll learn to build streaming data pipelines using Google cloud Pub/Sub and Dataflow to enable real-time decision making. You will also learn how to build dashboards to render tailored output for various stakeholder audiences.
这是一节入门级微课程,旨在解释什么是生成式 AI、它的用途以及与传统机器学习方法的区别。该课程还介绍了可以帮助您开发自己的生成式 AI 应用的各种 Google 工具。
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
Incorporating machine learning into data pipelines increases the ability to extract insights from data. This course covers ways machine learning can be included in data pipelines on Google Cloud. For little to no customization, this course covers AutoML. For more tailored machine learning capabilities, this course introduces Notebooks and BigQuery machine learning (BigQuery ML). Also, this course covers how to productionalize machine learning solutions by using Vertex AI.
Explore how to use AI to automate document processing tasks, such as classifying documents, extracting data from documents, and summarizing documents. Learn how to use the Document AI Workbench to create custom document extractors and summarizers. Upload documents, define fields, create versions, and call endpoints to get structured data and summaries back. Discover a new service called Document AI Warehouse, which is a fully managed service to search, store, govern, and manage documents and their extracted metadata. You will also learn about how it integrates with other Google Cloud services like Document AI, BigQuery, and Cloud Storage.
This workload aims to upskill Google Cloud partners to perform specific tasks associated with building a Custom Doc Extractor using the Google Cloud AI solution. The following will be addressed: Service: Document AI Task: Extract fields Processors: Custom Document Extractor and Document Splitter Prediction: Using Endpoint to programmatically extract fields
In this intermediate course, you will learn to design, build, and optimize robust batch data pipelines on Google Cloud. Moving beyond fundamental data handling, you will explore large-scale data transformations and efficient workflow orchestration, essential for timely business intelligence and critical reporting. Get hands-on practice using Dataflow for Apache Beam and Serverless for Apache Spark (Dataproc Serverless) for implementation, and tackle crucial considerations for data quality, monitoring, and alerting to ensure pipeline reliability and operational excellence. A basic knowledge of data warehousing, ETL/ELT, SQL, Python, and Google Cloud concepts is recommended.
完成入门级技能徽章课程 Dataplex 使用入门, 展现您在以下方面的技能:创建 Dataplex 资产,创建切面类型, 以及将切面应用于 Dataplex 中的条目。
While the traditional approaches of using data lakes and data warehouses can be effective, they have shortcomings, particularly in large enterprise environments. This course introduces the concept of a data lakehouse and the Google Cloud products used to create one. A lakehouse architecture uses open-standard data sources and combines the best features of data lakes and data warehouses, which addresses many of their shortcomings.
This course helps learners create a study plan for the PDE (Professional Data Engineer) certification exam. Learners explore the breadth and scope of the domains covered in the exam. Learners assess their exam readiness and create their individual study plan.
完成入门级技能徽章课程为 Compute Engine 实现云负载均衡,展示以下方面的技能: 在 Compute Engine 中创建和部署虚拟机 以及配置网络和应用负载均衡器。
This course introduces the Google Cloud big data and machine learning products and services that support the data-to-AI lifecycle. It explores the processes, challenges, and benefits of building a big data pipeline and machine learning models with Vertex AI on Google Cloud.
完成中级技能徽章课程使用 BigQuery 构建数据仓库,展示以下技能: 联接数据以创建新表、排查联接故障、使用并集附加数据、创建日期分区表, 以及在 BigQuery 中使用 JSON、数组和结构体。
Looking to build or optimize your data warehouse? Learn best practices to Extract, Transform, and Load your data into Google Cloud with BigQuery. In this series of interactive labs you will create and optimize your own data warehouse using a variety of large-scale BigQuery public datasets. BigQuery is Google's fully managed, NoOps, low cost analytics database. With BigQuery you can query terabytes and terabytes of data without having any infrastructure to manage or needing a database administrator. BigQuery uses SQL and can take advantage of the pay-as-you-go model. BigQuery allows you to focus on analyzing data to find meaningful insights. Looking for a hands on challenge lab to demonstrate your skills and validate your knowledge? On completing this quest, enroll in and finish the additional challenge lab at the end of this quest to receive an exclusive Google Cloud digital badge.
完成入门级技能徽章课程“从 BigQuery 数据中挖掘数据洞见”,展示您在以下方面的技能: 编写 SQL 查询、查询公共表、将示例数据加载到 BigQuery 中、 在 BigQuery 中使用查询验证器排查常见的语法错误,以及通过连接到 BigQuery 数据在 Looker Studio 中 创建报告。