Skip to content
Last updated

Treasure Workflow Prerequisites and Limitations

Review these Treasure Workflow prerequisites and limitations to understand what you can and cannot accomplish with workflows.

Prerequisites

Before you start creating a Treasure Workflow, you must have a database and table in Treasure Data.

Limitations

Unsupported Digdag Features

Treasure Workflow is based on Digdag, but following features are not supported:

  • td> operator's download_file parameter for downloading query results locally. Instead, use the Treasure Data result export functionality

  • sh> for running shell scripts

  • rb> for running ruby scripts

  • embulk> for running arbitrary Embulk jobs (but you can use td_load> for importing bulk data into Treasure Data)

  • emr> for running Amazon EMR jobs

  • param_get> for getting persistent data from ParamServer and set it as a value of store parameters

  • param_set> for setting a value into a ParamServer as persistent data

py> is the only supported custom script.

Treasure Workflow Service Limits

TD Workflow is designed to provide a scalable and flexible solution for managing your data pipelines in a cloud-hosted environment.

Limits of Use

To ensure optimal and fair processing, the following limits exist:

  • A maximum of 30 tasks are running concurrently per account. All other tasks get queued and are issued on a first-come-first-served basis.

  • The maximum size of a project archive is 10 MB.

  • 12,000 Saved Workflows. Over 12,000, and not all workflows will load within our Workflow UI. They are still accessible via CLI.

  • The maximum total number of tasks in an attempt is 1,000.

  • The maximum length of a task's full name is 640 bytes.

  • Maximum attempts per account are:

    • US - 200

    • Tokyo - 200

    • EU - 300

    • All other regions - 100

  • The period of time until a running task is killed is 1 day (24 hours).

  • The period of time until a running attempt is killed is 7 days.

  • The maximum response size for a td> task is 4 MB.

  • The maximum response size for a http> task is 1 MB.

  • The maximum response size for a http_call> task is 2 MB.

  • The maximum output size for a py> task (Custom Scripts), including exported variables and any generated tasks, is 36 MB.

  • The total response size for a td_for_each> task is 16 MB.

Advised Limits of Use

For better performance of the TD Console Workflow, try to stay below the following limits:

  • 200 Saved Workflows

  • 400 Tasks in a Workflow

These limits are subject to change if there are changes to the configuration of Treasure Data capacity.

See also: List of Upper Limits (MAX) on Treasure Data