menu
arrow_back

Cloud Composer: Copying BigQuery Tables Across Different Locations

—/100

Checkpoints

arrow_forward

Create Cloud Composer environment.

Create two Cloud Storage buckets.

Create a dataset.

Uploading the DAG and dependencies to Cloud Storage

Cloud Composer: Copying BigQuery Tables Across Different Locations

1 jam 7 Kredit

GSP283

Google Cloud Self-Paced Labs

Overview

In this advanced lab, you will learn how to create and run an Apache Airflow workflow in Cloud Composer that completes the following tasks:

  • Reads from a config file the list of tables to copy
  • Exports the list of tables from a BigQuery dataset located in US to Cloud Storage
  • Copies the exported tables from US to EU Cloud Storage buckets
  • Imports the list of tables into the target BigQuery Dataset in EU

cce6bf21555543ce.png

Bergabunglah dengan Qwiklabs untuk membaca tentang lab ini selengkapnya... beserta informasi lainnya!

  • Dapatkan akses sementara ke Google Cloud Console.
  • Lebih dari 200 lab mulai dari tingkat pemula hingga lanjutan.
  • Berdurasi singkat, jadi Anda dapat belajar dengan santai.
Bergabung untuk Memulai Lab Ini