v2.22.0

Release 2.22.0 of the TetraScience Data Platform includes many new enhancements. Enhancements, as well as bug fixes and minor functionality changes, are listed below.

Enhancements (New Features and Functionality)

There were several enhancements made in this version of the software.

Pipeline Self-Service Feature

The Pipeline Self-Service feature allows customers to program their own pipelines for file processing. Note that work continues for this feature and enhancements are planned for future versions of the software.

Enhancements made in this version of the software.

  • The TetraScience Python SDK is now available here: https://pypi.org/project/ts-sdk/.
  • The traceID has been added to all pipeline master and task script logs.
  • Logging is available on the pipeline workflow page.
  • The status of the pipeline processing (success or failure) is now sent as an email.

Data Hub Enhancements

Several data hub improvements were also made in this version of the software:

  • There are now no additional settings that you need to configure to set up a proxy; all of the settings are handled in the installation script.
  • If the disk space and memory requirements are not met for the Data Hub installer, the installer fails. If the recommended requirements are not met, a warning message is displayed.
  • The software now restarts the connector only the connector has changed. (There no more periodic restart.)
  • Data Hub now syncs periodically to ensure that it has the latest Data Hub and Data Connector upgrades.
  • In the Data Management page, there is now a new message that indicates that the online status might take a minute to update.

Documentation Enhancements

Several documentation improvements were also made to this version of the software. Some improvements include:

  • The File Info Service API has been documented.
  • The overview for the Empower Data Science Link has been documented.
  • The proxy setting documentation now has more implementation detail.

Bug Fixes

The following bug fixes were made for this version of the software

  • SQL queries now work for organizations that have a hyphen in the organizational slug (name).
  • You can now only set 0 to 5 standby instances in the pipeline details.
  • You can now delete files from the Elasticsearch results.
  • The stack overflow error that occurred during logging was fixed.
  • Validation checks are now made on the required fields for the pipeline edit.
  • All files that fall within the range of the dates indicated by the filter are now returned.
  • Updating an existing batch creates only one instance of each file in the data lake.
  • The filter date from the UI is now being passed as YYYY-MM-DD.
  • The discrepancy between the timestamp in the file list row and the workflow history has been resolved.
  • Data Hub now syncs periodically to ensure that it has the latest Data Hub and Data Connector upgrades.
  • Redirected the offline alarm output to stdout.
  • The file summary now loads for all files.
  • The status of the pipeline processing (success or failure) is now sent as an email.

Deprecated Features

The following features are deprecated for this release.

Usage of the API Key for Authentication

The API Key, which is used for authentication, is deprecated. While the API Key still works, we strongly recommend that you use the JSON Web Token (JWT) instead.

Functionality Changes

There are a few minor functionality changes in this release.

Pipeline Execution

When the pipeline runs for the first time, there is a 1 GB memory limit for all tasks. If a task fails with an Out Of Memory error, the software automatically reruns the task and increases the memory limit to 2 GB. If the task fails again, the pipeline fails.

Additional Notes

There are no additional notes for this release.