EDSL Frequently Asked Questions

The following lists frequently asked questions about EDSL.

What environments can EDSL be hosted on?

EDSL can be deployed inside of the Customer’s AWS, in accounts dedicated to TDP. Note that one instance of the TDP has three environments:

  • Development (Dev)
  • Test, which is used to run the GxP validation
  • Production (Prod)

Note that each environment should have a separate AWS account.

How Does EDSL Work with My Data Lake?

There are two options:

  • TDP can use a data pipline to push data to your data lake.
  • TDP can be deployed inside of your private cloud, which can be used as part of your data lake.

TDP stores all of the data in S3. Customers have direct access via AWS IAM. Note that both structured and unstructured data is can be stored in the data lake:

  • Data warehouse – holds structured data
  • Data lake – holds both structured and unstructured data

How Does EDSL Handle GxP (Good Practice Quality Guidelines and Regulations)?

TDP can be used in GxP compliant workflows. Here are a few features that help the TDP conform to those guidelines and regulations:

  • Thorough user action audit trail for 21 CFR part 11 use cases.
  • Flexible deployment in three environments: Development, Test and Production inside of the customer’s own AWS.
  • Example/template qualification/validation scripts.
  • Validation support to create and run qualification and validation.
  • Continuous qualification via automated integration test in the platform.
  • All system actions related to the data are logged (data agents, data connectors, DataHub, data pipeline).

Validation and Diagnostic Pipelines

The TDP has diagnostic pipelines that help handle validation. Diagnostic pipelines check for issues in pipelines, such as whether a pipeline has not been triggered since a certain date, if there are "stuck" or timed-out workflows, too many active workflows, diagnostic lambda errors, or workflow errors.

Diagnostic pipelines can be run manually or automatically and are available for Development, Test, and Production environments, as well as multi-tenant and single-tenant deployments.

Diagnostic pipelines are triggered every 3 hours and after platform updates. All alerts are sent to a special SNS alert topic. All alerts can be received by subscribing to the SNS topic and are available in Cloudwatch.

How is data archiving handled?

TDP saves all data in AWS S3. You can configure archive policies in your AWS account to archive the data to AWS Glacier if desired.

What are the advanced features provided in the full version of TDP?

EDSL provides core features needed to process Waters Empower-generated data, securely push and store data in the data lake, and transform the data into an IDS format that is vendor-agnostic, and that easily be used by data scientists and others.

There are some advanced features that are available in the full version of the TDP product. These include:

  • Broader Ecosystem Integrations
  • Additional Agents
  • Additional Connectors
  • Self-service data pipelines; you can build your own pipelines
  • 50+ production-ready IDS’s that are available to process data from various vendors, right out of the box

Did this page help you?