Complete the introductory Get Started with Dataplex skill badge to demonstrate skills in the following: creating Dataplex assets, creating aspect types, and applying aspects to entries in Dataplex.
This course is part 1 of a 3-course series on Serverless Data Processing with Dataflow. In this first course, we start with a refresher of what Apache Beam is and its relationship with Dataflow. Next, we talk about the Apache Beam vision and the benefits of the Beam Portability framework. The Beam Portability framework achieves the vision that a developer can use their favorite programming language with their preferred execution backend. We then show you how Dataflow allows you to separate compute and storage while saving money, and how identity, access, and management tools interact with your Dataflow pipelines. Lastly, we look at how to implement the right security model for your use case on Dataflow.
L'integrazione del machine learning nelle pipeline di dati aumenta la capacità di estrarre insight dai dati. Questo corso illustra i modi in cui il machine learning può essere incluso nelle pipeline di dati su Google Cloud. Per una personalizzazione minima o nulla, il corso tratta di AutoML. Per funzionalità di machine learning più personalizzate, il corso introduce Notebooks e BigQuery Machine Learning (BigQuery ML). Inoltre, il corso spiega come mettere in produzione soluzioni di machine learning utilizzando Vertex AI.
L'elaborazione dei flussi di dati sta diventando sempre più diffusa poiché la modalità flusso consente alle aziende di ottenere parametri in tempo reale sulle operazioni aziendali. Questo corso tratta la creazione di pipeline di dati in modalità flusso su Google Cloud. Pub/Sub viene presentato come strumento per la gestione dei flussi di dati in entrata. Il corso spiega anche come applicare aggregazioni e trasformazioni ai flussi di dati utilizzando Dataflow e come archiviare i record elaborati in BigQuery o Bigtable per l'analisi. Gli studenti acquisiranno esperienza pratica nella creazione di componenti della pipeline di dati in modalità flusso su Google Cloud utilizzando QwikLabs.
I due componenti chiave di qualsiasi pipeline di dati sono costituiti dai data lake e dai data warehouse. In questo corso evidenzieremo i casi d'uso per ogni tipo di spazio di archiviazione e approfondiremo i dettagli tecnici delle soluzioni di data lake e data warehouse disponibili su Google Cloud. Inoltre, descriveremo il ruolo di un data engineer, illustreremo i vantaggi di una pipeline di dati di successo per le operazioni aziendali ed esamineremo i motivi per cui il data engineering dovrebbe essere eseguito in un ambiente cloud. Questo è il primo corso della serie Data engineering su Google Cloud. Dopo il completamento di questo corso, iscriviti al corso Creazione di pipeline di dati in batch su Google Cloud.
Questo corso presenta i prodotti e i servizi per big data e di machine learning di Google Cloud che supportano il ciclo di vita dai dati all'IA. Esplora i processi, le sfide e i vantaggi della creazione di una pipeline di big data e di modelli di machine learning con Vertex AI su Google Cloud.
This course reviews the processes for creating table calculations, pivots and visualizations
This course is designed for Looker users who want to create their own ad-hoc reports. It assumes experience of everything covered in our Get Started with Looker course (logging in, finding Looks & dashboards, adjusting filters, and sending data)
In this course you will discover Liquid, the templating language invented by Shopify and explore how it can be used in Looker to create dynamic links, content, formatting, and more.
Hands on course covering the main uses of extends and the three primary LookML objects extends are used on as well as some advanced usage of extends.
This course is designed to teach you about roles, permission sets and model sets. These are areas that are used together to manage what users can do and what they can see in Looker.
This course aims to introduce you to the basic concepts of Git: what it is and how it's used in Looker. You will also develop an in-depth knowledge of the caching process on the Looker platform, such as why they are used and why they work
This course provides an introduction to databases and summarized the differences in the main database technologies. This course will also introduce you to Looker and how Looker scales as a modern data platform. In the lessons, you will build and maintain standard Looker data models and establish the foundation necessary to learn Looker's more advanced features.
This course provides an iterative approach to plan, build, launch, and grow a modern, scalable, mature analytics ecosystem and data culture in an organization that consistently achieves established business outcomes. Users will also learn how to design and build a useful, easy-to-use dashboard in Looker. It assumes experience with everything covered in our Getting Started with Looker and Building Reports in Looker courses.
In this course, we’ll show you how organizations are aligning their BI strategy to most effectively achieve business outcomes with Looker. We'll follow four iterative steps: Plan, Build, Launch, Grow, and provide resources to take into your own services delivery to build Looker with the goal of achieving business outcomes.
By the end of this course, you should be able to articulate Looker's value propositions and what makes it different from other analytics tools in the market. You should also be able to explain how Looker works, and explain the standard components of successful service delivery.