This directory contains samples of Batch related GCP workflows
- You have a Google Cloud project.
- You have enabled the Batch API and understand the basic Batch concepts.
- You have the permissions required to create Batch jobs.
- You should know about how to create and deploy GCP Workflows, more details in Workflows Overview.
- To grant the necessary IAM roles to the Workflows runner by following Workflows IAM control guide.
- This workflow will try to create big query dataset and table, it then will export the jobs into the big query table and delete the exported jobs.
- To run this workflow, the proper IAM roles should be granted to work with Big Query, for more details, please refer to BigQuery permissions and predefined IAM roles.
- The arguments of the workflow are:
- project
- location
- job_filter: Used with list/delete the Batch jobs, default to list FAILED or SUCCEEDED Batch jobs which are created before 2023/05/01 for demonstration purpose
- page_size: default to 100
- dateset_id: default is default_dataset_id
- table_id: default is default_table_id
- The schema of the big query table is three string type columns: name, uid, job.