Infoworks Release Notes
Release Notes

v6.0.0

Date of Release: July 2024

This section consists of the new features and enhancements introduced in this release.

  • Databricks Unity Catalog as a metastore:

    • Infoworks now supports the unity catalog metastore in azure databricks environments. So that user can administer data access policies that apply across workspaces.
    • Infoworks now supports new storage type universal format (uniForm) in azure databricks environments. So that iceberg clients can access the data written by Infoworks.
  • Infoworks now supports Ability to Re-Order Columns during the Ingestion on UI itself. User can specify the target column order in which he wants the columns to land in target.

  • Infoworks now supports the TLS version 1.3 for all it's internal communications, so that customer compliance is enforced.

  • Infoworks now supports the custom tags to be set on ingestion, which can be used to identify/segregate the resource based on tags.

  • Users can now specify lower and upper watermark for a particular run. Read more about this under Onboarding Data.

  • Restricted users' access to secrets by associating with domains. Refer here for details.

  • Restricted users' access to Data Environments to only their assigned domains. For details, please refer here and Adding Secret Store/Keyvault.

  • Users will now be able to submit jobs and workflows during upgrade.

  • Secrets from secret store can be accessed in Bash node of workflow. Refer to the Bash Script section for details.

  • Secrets from secret store can be accessed in job pre-post hooks and custom target for pipeline extension. Refer to the Managing Job Hooks section.

  • For prerequisites of Network Policy on AKS Cluster: please refer here.

  • Confluent Cloud now supports using AzureAD as Identity provider. You can view the details here.

Resolved Issues

This section consists of the resolved issues in this release:

JIRA IDIssue
IPD-25709Solution to reduce/eliminate Integer and Decimal datatype mismatches between Teradata source and target Snowflake
IPD-26474Regression: Redshift pipeline build failing on azure databricks 14.3
IPD-25719Data validation for failed failure
IPD-25837API to change table group scheduled user is not working
IPD-25839GET table group call gives a refresh token of the scheduled user in the response
IPD-25840Source_schema_name and Source_table_name are interchanged in the ingestion metrics response in v5.5.0.5
IPD-25868Incorrect success response of verify refresh token API
IPD-25869Add table action of the file mapping doesn't validate the uniqueness of the target DB, schema, and table name combination
IPD-26017Ingestion jobs fail when access control list parameter is configured with 'group'
IPD-25952Status.sh showing connection error warnings/exception stack trace when all services are in stopped state
IPD-26115CICD failing for Pipeline group migration
IPD-26110The upload schema option is erroring if the user tries to update the column name
IPD-26143Issue with executing snowflake stored procedure
IPD-25940Snowflake pipeline export job on the existing artifacts (created in 5.0) creating the DB/SCHEMA/TABLE names as case sensitive (in lower case)
IPD-26224Case of column names converted to lower case on pipeline build to cosmos db target
IPD-26178Merge to Snowflake table fails SQL compilation error
IPD-26285Ingestion failing with java.lang.NullPointerException in v5.5.1.2
IPD-26304Ingestion on user-managed tables failing with schema mismatch error post 5.5.1.2 upgrade
IPD-26278v5.5.1.3 - Add user-agent to databricks API calls

Enhancements

This section consists of the improvements in this release:

JIRA IDImprovements
IPD-25709Solution to reduce/eliminate Integer and Decimal datatype mismatches between Teradata source and target Snowflake
IPD-24208Optimize the incremental query to retrieve a maximum watermark value for segmented load tables

Known Issues

  • Streaming jobs that have been stopped may show running state for the cluster job - users can verify that the job is actually stopped by observing that the number of batches run for that job does not increase after stopping it; more details here.
  • Pipeline node preview data request times out for initial few tries.
  • Pipeline build failing with NPE when read from merged/deduplicated table is selected.
  • User is able to delete to the pipeline extension added to a domain.
  • Insert Overwrite/Merge Job failure for datalake environments for dt jobs.
  • Config migrating a table configuration is not updating the "table_type" to user managed table, when the imported json has the table as user managed table.
  • Sync to target job to Snowflake from a CSV source is having quotes added to the last varchar column.
  • If workflows rest-api calls are failing with 504 errors add environment PYTHON_REQUESTS_TIMEOUT with value “60” in orchestrator deployment.
  • Micro batches processing stop for streaming ingestion if the source stops streaming the data for Databricks runtime version 14.3.
  • Ingestion with data quality does not work for Databricks runtime version 14.3.

Limitations

  • Databricks job on single node ephemeral cluster with unity catalog fails with error INVALID_PARAMETER_VALUE : Spark Conf: 'spark.databricks.cluster.profile' is not allowed when choosing an access mode.
  • Case sensitivity of column names is not supported in direct onboarding to Snowflake (Ingestion on snowflake environment).
  • Pipeline targets are converting all target columns to lowercases.
  • The timestamp columns cannot be used as partition columns directly with uniForm storage type. The user can still use timestamp columns as partition columns if created with derived partitions out of it.
  • Pipeline previews are not working for Databricks runtime version 14.3.
  • Snowflake variables cannot be used as Table Identifiers/Table Names in the pipeline query. If the user tries to use, it will not fetch/validate the metadata for the table.
  • Target entity names on table configuration are accepting two double quotes as valid input.
  • Pipeline with update mode containing InNotIn or Exist Node is not updating the data.

For Kubernetes-based installation, refer to Infoworks Installation on Azure Kubernetes Service (AKS).

For more information, contact support@infoworks.io.

Upgrade

For upgrading from 5.5.1.x to 6.0.0 for Azure Kubernetes, refer to Upgrading Infoworks from 5.5.1.x to 6.0.0 for Azure Kubernetes.

PAM

The Product Availability Matrix (PAM) is available here.

On This Page
v6.0.0