Connection configuration

Data Integrity Suite

Product
Spatial_Analytics
Data_Integration
Data_Enrichment
Data_Governance
Precisely_Data_Integrity_Suite
geo_addressing_1
Data_Observability
Data_Quality
dis_core_foundation
Services
Spatial Analytics
Data Integration
Data Enrichment
Data Governance
Geo Addressing
Data Observability
Data Quality
Core Foundation
ft:title
Data Integrity Suite
ft:locale
en-US
PublicationType
pt_product_guide
copyrightfirst
2000
copyrightlast
2026

This section provides you the details on various fields required to configure BigQuery in Data Integrity Suite.

Field Description
Name Specifies the name of the data connection. This is a mandatory field.
Description Allows you to describe the purpose of this data connection.
Use Connection As Select the type of connection to define its role:
  • Source: Choose this option if the connection is used to extract data from the data source.
  • Target: Choose this option if the connection is used to transfer data to the data source.
  • Source and Target: Choose this option if the connection will be used for both extracting data from and transferring data to the data source.
By default, the BigQuery connection is selected as the Source, and you are required to enter the following fields:
Credentials JSON Specifies the contents of the service account key file, which is a JSON object required for authentication with the data source. This is a mandatory field.
Default Dataset ID Specifies the name of the BigQuery dataset to connect to. A dataset in BigQuery acts as a container for tables and other objects.
Project ID Specifies the unique identifier of your BigQuery project, which can be located in the Google Cloud Console. This is a mandatory field.
If using the BigQuery connection as the Target, provide the following additional fields:
Credentials JSON Specifies the contents of the service account key file, which is a JSON object required for authentication with the data source.
Default Dataset ID Specifies the name of the BigQuery dataset to connect to. A dataset in BigQuery acts as a container for tables and other objects. This is a mandatory field.
Project ID Specifies the unique identifier of your BigQuery project, which can be located in the Google Cloud Console. This is a mandatory field.
Dataset Location Select the dataset location where the data will be loaded. Options include US, EU, or Asia. This is a mandatory field.
Loading Method Specifies the method for loading data. Choose between Standard Inserts or Google Cloud Storage (GCS) Staging.

If using Standard Inserts, configure the following additional fields:

  • Transformation Query Run Type: Select whether query jobs must run in interactive or batch mode.

  • Google BigQuery Client Chunk Size: Specifies the maximum number of rows to read from BigQuery per API request.

If using (GCS) Staging, configure the following additional fields:

  • Credential: Specifies the credential type for authentication and the HMAC Key is selected. This is a mandatory field
  • HMAC Key Access ID: Specifies the Access ID associated with the HMAC Key used for authentication. This is a mandatory field.
  • HMAC Key Secret: Specifies the secret associated with the HMAC Key. This is a mandatory field.
  • GCS Bucket Name: Specifies the name of the GCS bucket where data will be staged. This is a mandatory field.
  • GCS Bucket Path: Specifies the path within the GCS bucket where data will be stored. This is a mandatory field.
  • GCS Tmp Files Afterward Processing: Specifies the handling of temporary files in GCS after data loading. The two options are:
  • Delete All Tmp Files from GCS: Select this option to automatically remove all temporary files from the GCS bucket once the data has been loaded into BigQuery.
  • Keep All Tmp Files in GCS: Select this option to retain all temporary files in the GCS bucket even after the data has been loaded into BigQuery.
  • Transformation Query Run Type: Select whether query jobs must run in interactive or batch mode.
  • Google BigQuery Client Chunk Size: Specifies the maximum number of rows to read from BigQuery per API request.