Engineer Data in Google Cloud


Engineer Data in Google Cloud

8 godz. Advanced universal_currency_alt Punkty: 51
Earn a skill badge by completing the Engineer Data in Google Cloud quest, where you will learn how to: 1. Build data pipelines using Cloud Dataprep by Trifacta, Pub/Sub, and Dataflow. 2. Use Cloud IoT Core to collect and manage MQTT-based devices. 3. Use BigQuery to analyze IoT data. 4. Use Cloud Storage, Dataflow, and BigQuery to perform ETL. 5. Build a machine learning model using BigQuery ML. 6. Build a machine learning model using TensorFlow 1.x and AI Platform.

This quest is a great resource for understanding topics that will appear in the Google Cloud Certified Professional Data Engineer Certification.

A skill badge is an exclusive digital badge issued by Google Cloud in recognition of your proficiency with Google Cloud products and services and tests your ability to apply your knowledge in an interactive hands-on environment. Complete the skill badge quest, and final assessment challenge lab, to receive a digital badge that you can share with your network.

Jeśli wykonasz wszystkie zadania w tym szkoleniu, otrzymasz odznakę. Zbieraj odznaki w Qwiklabs za ukończone kursy, gry i szkolenia. Zdobądź je wszystkie, aby pochwalić się swoimi umiejętnościami.

  • Moduł

    Creating a Data Transformation Pipeline with Cloud Dataprep

    Cloud Dataprep by Trifacta is an intelligent data service for visually exploring, cleaning, and preparing structured and unstructured data for analysis. In this lab you will explore the Cloud Dataprep UI to build a data transformation pipeline.

  • Moduł

    Building an IoT Analytics Pipeline on Google Cloud

    This lab shows you how to connect and manage devices using Cloud IoT Core; ingest the stream of information using Cloud Pub/Sub; process the IoT data using Cloud Dataflow; use BigQuery to analyze the IoT data. Watch this short video, Easily Build an IoT Analytics Pipeline.

  • Moduł

    ETL Processing on Google Cloud Using Dataflow and BigQuery

    In this lab you will build several Data Pipelines that will ingest data from a publicly available dataset into BigQuery.

  • Moduł

    Predict Visitor Purchases with a Classification Model in BQML

    In this lab you will use a newly available ecommerce dataset to run some typical queries that businesses would want to know about their customers’ purchasing habits.

  • Moduł

    Predict Housing Prices with Tensorflow and AI Platform

    In this lab you will build an end to end machine learning solution using Tensorflow + AI Platform and leverage the cloud for distributed training and online prediction.

  • Moduł

    Cloud Composer: Copying BigQuery Tables Across Different Locations

    In this advanced lab you will create and run an Apache Airflow workflow in Cloud Composer that exports tables from a BigQuery dataset located in Cloud Storage bucktes in the US to buckets in Europe, then import th0se tables to a BigQuery dataset in Europe.

  • Moduł

    Engineer Data in Google Cloud: Challenge Lab

    This challenge lab tests your skills and knowledge from the labs in the Engineer Data in Google Cloud quest. You should be familiar with the content of labs before attempting this lab.

  • info
    Quest Info
    Prior to enrolling in this skill badge quest, it is recommended that you complete the the following quests: