Infoworks 6.1.3
Getting Started

Configuring Infoworks with BigQuery

Introduction

Infoworks’ BigQuery data environment enables onboarding data directly into BigQuery and transforming data in BigQuery. Workloads may be orchestrated using Infoworks’ orchestration capabilities and monitored using the operational dashboards. This section will help you configure an Infoworks data environment to work with BigQuery.

NOTE Infoworks also supports onboarding and processing data in data lake environments (for example, Databricks and Dataproc) and exporting the data from the data lake into BigQuery. If you wish to do this, you should set up a data lake environment instead or in addition to setting up a BigQuery data environment. Refer to the section on configuring Infoworks for your data lake environment for details.

Procedure

To configure and connect to the required BigQuery instance, navigate to Admin > Manage Data Environments, and then click Add button under the BigQuery option.

The following window appears:

There are three tabs to be configured as follows:

Data Environment

To onboard data directly to BigQuery, you should configure a BigQuery environment that includes cloud storage and one or more Spark computes. Cloud storage is used temporarily to stage data during ingestion and to store sample data. To configure the data environment details, enter values in the following fields. This defines the environmental parameters, to allow Infoworks to be configured to the required BigQuery instance:

FieldDescriptionDetails
Data Environment NameData environment defines where and how your data will be stored and accessed. Data environment name must help the user to identify the environment being configured.User-defined. Provide a meaningful name for the data environment being configured.
DescriptionDescription for the environment being configured.User-defined. Provide required description for the environment being configured.
Authentication Mechanism

Specifies the authentication mechanism using which the security information is stored. Includes the following authentication mechanisms:

  • Use system role credentials
  • Use service account credentials
The System role credentials option uses the default credentials of the instance to identify and authorize the application. The Service account credentials uses the IAM Service Account to identify and authorize the application.
Service CredentialsOption to select the service credential which authenticates calls to Google Cloud APIs or other non-Google APIs.Radio Button. Available options are Upload File or Enter File Location.
Section: Additional ParametersClick the Add button to provide the parameters in key-value pair.Provide additional parameters required to connect to BigQuery. It is non mandatory.

After entering all the required values, click Continue to move to the Compute tab.

Compute

A Compute template is the infrastructure used to execute a job. This compute infrastructure requires access to the metastore and storage that needs to be processed. To configure the compute details, enter values in the following fields. This defines the compute template parameters to allow Infoworks to be configured to the required BigQuery instance.

You can select one of the clusters as the default cluster for running the jobs. However, this can be overwritten at individual job level.

Infoworks supports creating multiple persistent clusters in a BigQuery environment, by clicking the Add button.

Enter the fields in the Compute section:

FieldDescriptionDetails
Compute Provider
Compute ProviderThe type of compute provider required for creating compute templatesChoose from the available options: Google Dataproc and Databricks on GCP
Workspace UrlURL of the workspace that Infoworks must be attached to.Provide the required workspace URL.
Databricks TokenDatabricks access token of the user who uses Infoworks.Provide the required Databricks token.
RegionGeographical location where you can host your resources.Select the required region from the drop-down.
Compute Clusters
Cluster TypeThe type of compute cluster that you want to launch.Choose from the available options: Persistent or Ephemeral. Jobs can be submitted on both Ephemeral as well as Persistent clusters.
NameName required for the compute template that you want to use for the jobs.User-defined. Provide a meaningful name for the compute template being configured.
DescriptionDescription required for the compute template.User-defined. Provide required description for the compute template being configured.
Custom TagsEnter the key-value pairs that help you identify resources based on settings that are relevant to your organization.Provide the required custom tags.
Runtime VersionSelect the Runtime version of the compute cluster that is being used.

Select the Runtime version.

If you are using Databricks on GCP, you can either select 7.3 or 9.1 from the drop-down as per your requirement.

If you are using Google Dataproc, select 2.0 from the dropdown.

RegionGeographical location where you can host your resources.Provide the required region. For example: East US.
Workspace URLURL of the workspace that Infoworks must be attached to.Provide the required workspace URL.
Databricks TokenDatabricks access token of the user who uses Infoworks.Provide the required Databricks token.
Allow single node instanceOption to run single node clusters.A single node cluster is a cluster consisting of an apache spark driver and no spark workers.
Use Instance PoolOption to use a set of idle instances which optimizes cluster start and auto-scaling times.If Use Instance pool check box is checked, provide the ID of the created instance pool in the additional field that appears.
Worker TypeWorker type configured in the edge node.This field appears only if Use Instance pool check box is unchecked. Provide the required worker type. For example: Standard_L4
Driver TypeDriver type configured in the edge node.This field appears only if Use Instance pool check box is unchecked. Provide the required worker type. For example: Standard_L4
Max Allowed Worker NodesMaximum number of worker instances allowed.Provide the maximum allowed limit of worker instances.
Enable AutoscaleOption for the instances in the pool to dynamically acquire additional disk space when they are running low on disk space.Select this option to enable autoscaling.
Min Workers NodeMinimum number of workers that workspace maintains.This field appears only if Enable Autoscale check box is checked.
Max Workers NodeMaximum number of workers that workspace maintains.This field appears only if Enable Autoscale check box is checked. This must be greater than or equal to Default Min Worker value.
Number of Worker NodesNumber of workers configured for availability.This field appears only if Enable Autoscale check box is unchecked.
Terminate after minutes of inactivityNumber of minutes after inactivity which the pool maintains before being terminated.Provide the minimum number of minutes to be maintained before termination.

NOTE If an admin edits the interactive cluster, the cluster restarts, and hence the job running on that cluster fails.

Storage

To configure the storage details, enter values in the following fields. This defines the storage parameters, to allow Infoworks to be configured to the required BigQuery instance. After configuring a storage, you can choose to make it default storage for all jobs. However, this can be overwritten at individual job level.

NOTE To configure a new storage after the first time configuration, click the Add button in the UI.

Enter the following fields under the Storage section:

FieldDescriptionDetails
NameStorage name must help the user to identify the storage credentials being configured.User-defined. Provide a meaningful name for the storage set up being configured.
DescriptionType of storage system where all the artifacts will be stored. This depends on the type of cloud/platform provider you choose in the Compute tab.Select the required storage type from the drop-down menu. The available option is GCS.
Access SchemeScheme used to access GCS.This field appears only when the storage type is selected as GCS.
Storage Authentication
Bucket NameBuckets are the basic containers that hold, organise, and control access to your data.Provide the storage bucket key. Do not use gs:// for storage bucket.
Authentication Mechanism

Specifies the authentication mechanism using which the security information is stored. Includes the following authentication mechanisms:

  • Use system role credentials
  • Use service account credentials
The System role credentials option uses the default credentials of the instance to identify and authorize the application. The Service account credentials uses the IAM Service Account to identify and authorize the application.
Service CredentialsOption to select the service credential which authenticate calls to Google Cloud APIs or other non-Google APIs.This field appears only when the Authentication mechanism is selected as Use service account credentials. Available options are Upload File or Enter File Location.

NOTES

  • For indirect method of ingestion in BigQuery environment, the GCS bucket used for staging and the BigQuery target should be in the same project. If the two are in different GCP projects, the ingestion will fail.
  • Different service accounts can be used for BigQuery and GCS bucket, but they should belong to the same GCP project.
  • Ingestion job works fine if BigQuery and GCS are in the same project and the cluster created is in some other GCP project.

After entering all the required values, click Save. Click Finish to view and access the list of all the environments configured. Edit, Clone, and Delete actions are available on the UI, corresponding to every configured environment.

  Last updated