Visit our new documentation site! This documentation page is no longer updated.

Data Connector for Zuora

You can import Zuora Data Source objects into Treasure Data. There are 2 modes of import: base and full (see fetch_related config).

Table of Contents


  • Basic knowledge of Treasure Data
  • Basic knowledge of Zuora

Option 1: Use Web Console

Create a new connection

Go to Treasure Data Connections. Locate and select Zuora.

Complete the Zuora login information, click Next and provide a connection name:

Create a new transfer

After creating the connection, you are automatically taken to the My Connections tab. Look for the connection you created and click New Transfer.

Provide information details and click Next.

Fetch related objects: Most of Zuora’s data sources provide related business objects, improving performance and eliminating the need to construct nested queries. If you need this data, check the box and select the data source you want to query.

API Version (default v85): Zuora API Version, it’s best to keep default value v85 (v84 will be deprecated soon).

Duration to import: This option enables you to select the 'end_date' since the 'start_date' of import.

  • If Start date is set to = 02/07/2017, 12:00 AM AND Number of Unit = 1 DAY, your data transfer will include all data available up until 02/08/2017, 12:00 AM.

In this example, the 'end_date' is calculated as follows:

'start_date' + 'number_of_unit' = 'end_date'
02/07/2017, 12:00AM + 1 DAY = 02/08/2017, 12:00AM
  • If the transfer is scheduled, the 'start_date' will be constantly changing based on the new 'end_date' since last import (as calculated previously).

  • If incremental is not selected, import will be from 'start_date' to now.

  • User timezone is reserved.

** We highly recommend you re-create a transfer if you’d like to change the 'start_date'. Editing an existing transfer does not affect the 'start_date'.

Next, you see a Preview of your data. If you want to change anything, click Advanced Settings or else click Next.

Third step is to select the database and table where you want to transfer the data:

Finally, specify the schedule of the data transfer and click Start Transfer:

You will see the new data transfer in progress, listed under the My Input Transfers tab, and a corresponding job will be listed in the Jobs section.

Now, you are ready to start analyzing your data.

Option 2: Use Command Line

Step 0: Install ‘td’ command v0.11.9 or later

You can install the newest Treasure Data Toolbelt.

$ td --version

Step 1: Create Configuration File

Prepare configuration file (for eg: load.yml), as shown in the following example, with your Zuora account access information.

  type: zuora
  username: xxxxxxxx
  password: xxxxxxxx
  base_object: Account #(required if fetch_related = false, see Appendix B)
  data_source: Account #(required if fetch_related = true, see Appendix C)
  fetch_related: true #(optional, default: false)
  from_timestamp: 12-05-2016T11:00:00.000Z
  incremental: true #(optional, default: true)
  duration: { unit: DAY, num: 1 } #(required if `incremental: true`)
  mode: replace

This example dumps Zuora Account Data Source:

  • username and password: your login information.
  • describe_api_verion: Zuora Describe API version, See Zuora Describe API
  • base_object: name of Zuora SOAP (Base) Object you want to import, required if fetch_related: false
    • See Appendix B for list of available base_object.
  • data_source: name of Zuora Data Source you want to import, required if fetch_related: true
    • See Appendix C for list of available data_source.
  • fetch_related: you can choose to import base or full object (including all pre-joined objects), default as false
  • from_timestamp: from which moment data will be imported, format is: yyyy-MM-dd'T'HH:mm:ss.SSS'Z'
  • incremental: should data import be continuous or once, default as true. If true, duration is required
  • duration: for how long data will be imported, determined by UpdatedDate field of target object. unit has 3 options: DAY, HOUR, and MIN, default as DAY.

Some Limited Availibility objects from Zuora require additional configuration settings. The following is a list of additional configuration settings for Billing Preview Run

  • from_timestamp : When used with Billing Preview Run, this is the Target Date for the preview run.
  • include_evergreen_sub : Indicated whether to include the evergreen subscription. The default is ‘false’
  • batch : Customer batch to include in this preview run
  • assume_renewal : Whether to assume renewal for the accounts. Valid value is All, None or Autorenew. The default is ‘None’.
  • charge_type_to_exclude : The charge types to exclude from the forecast run. Possible values: OneTime, Recurring, Usage, and any comma-separated combination of these values.

For more details on available out modes, see Appendix A

Step 2 (optional): Preview data to import

You can preview data to be imported using the command td connector:preview.

$ td connector:preview load.yml
| id:long | title:string | body_html:string | ...
| 1       | "33"         | <p>body_html<p>  |
| 2       | "34"         | <p>body_html<p>  |
| 3       | "35"         | <p>body_html<p>  |
| 4       | "36"         | <p>body_html<p>  |
| 6       | "37"         | <p>body_html<p>  |

Step 3: Execute Load Job

Finally, submit the load job. It may take a couple of hours depending on the data size. Users need to specify the database and table where their data are stored.

It is recommended to specify --time-column option, because Treasure Data’s storage is partitioned by time (see also data partitioning) If the option is not given, the Data Connector will choose the first long or timestamp column as the partitioning time. The type of the column specified by --time-column must be either of long and timestamp type.

If your data doesn’t have a time column you can add it using add_time filter option. More details at add_time filter plugin

$ td connector:issue load.yml --database td_sample_db --table td_sample_table --time-column updated_date

The connector:issue command assumes you have already created database(td_sample_db) and table(td_sample_table). If the database or the table do not exist in TD, the command fails, so create the database and table manually or use --auto-create-table option with td connector:issue command to auto create the database and table:

$ td connector:issue load.yml --database td_sample_db --table td_sample_table --time-column updated_date --auto-create-table
You can assign Time Format column to the "Partitioning Key" by "--time-column" option.

Scheduled execution

You can schedule periodic Data Connector execution for periodic Zuora import. The load distribution and operation of Treasure Data’s scheduler is optimized to achieve high availability. By using Treasure Data’s scheduler, you no longer need a cron daemon on your local datacenter.

Create the schedule

A new schedule can be created using the td connector:create command. TSpecify the name of the schedule, the cron-style schedule, the database and table where the data will be stored, and the Data Connector configuration file. The values are required.

$ td connector:create \
    daily_zuora_import \
    "10 0 * * *" \
    td_sample_db \
    td_sample_table \
The `cron` parameter also accepts these three options: `@hourly`, `@daily` and `@monthly`.

By default, schedule is setup in UTC timezone. You can set the schedule in a timezone using -t or —timezone option. Note that --timezone option only supports extended timezone formats like ‘Asia/Tokyo’, ‘America/Los_Angeles’ etc. Timezone abbreviations like PST, CST are not supported and may lead to unexpected schedules.

List the Schedules

You can see the list of currently scheduled entries by td connector:list.

$ td connector:list
| Name                  | Cron         | Timezone | Delay | Database     | Table           | Config                     |
| daily_zuora_import    | 10 0 * * *   | UTC      | 0     | td_sample_db | td_sample_table | {"type"=>"zuora", ... }    |

Show the Setting and History of Schedules

td connector:show shows the execution setting of a schedule entry.

% td connector:show daily_zuora_import
Name     : daily_zuora_import
Cron     : 10 0 * * *
Timezone : UTC
Delay    : 0
Database : td_sample_db
Table    : td_sample_table

td connector:history shows the execution history of a schedule entry. To investigate the results of each individual execution, use td job <jobid>.

% td connector:history daily_zuora_import
| JobID  | Status  | Records | Database     | Table           | Priority | Started                   | Duration |
| 578066 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-18 00:10:05 +0000 | 160      |
| 577968 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-17 00:10:07 +0000 | 161      |
| 577914 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-16 00:10:03 +0000 | 152      |
| 577872 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-15 00:10:04 +0000 | 163      |
| 577810 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-14 00:10:04 +0000 | 164      |
| 577766 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-13 00:10:04 +0000 | 155      |
| 577710 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-12 00:10:05 +0000 | 156      |
| 577610 | success | 10000   | td_sample_db | td_sample_table | 0        | 2015-04-11 00:10:04 +0000 | 157      |
8 rows in set

Delete the Schedule

td connector:delete removes the schedule.

$ td connector:delete daily_zuora_import


A) Modes for out plugin

You can specify file import mode in out section of load.yml.

append (default)

This is the default mode and records are appended to the target table.

  mode: append

replace (In td 0.11.10 and later)

This mode replaces data in the target table. Any manual schema changes made to the target table remains intact with this mode.

  mode: replace

B) Available Base Object

Base Object Base Object Base Object
Account InvoiceItem. ProductRatePlanChargeTier
AccountingCode InvoiceItemAdjustment RatePlan
AccountingPeriod InvoicePayment RatePlanCharge
BillingRun Payment RatePlanChargeTier
BillingPreviewRun PaymentMethod Refund
ChargeMetrics PaymentMethodTransactionLog RefundInvoicePayment
Contact PaymentTransactionLog RefundTransactionLog
CreditBalanceAdjustment Product Subscription
Feature ProductFeature SubscriptionProductFeature
Invoice ProductRatePlan TaxationItem
InvoiceAdjustment ProductRatePlanCharge Usage

C) Available Data Source

Data Source Data Source Data Source
Account PaymentReconciliationLog RevenueEventItemInvoiceItem
AccountingCode PaymentTransactionLog RevenueEventItemInvoiceItemAdjustment
BillingRun ProcessedUsage RevenueScheduleItem
Contact Product RevenueScheduleItemInvoiceItem
ChargeMetrics ProductRatePlanChargeTier RevenueScheduleItemInvoiceItemAdjustment
CreditBalanceAdjustment RatePlan RevenueScheduleItemCreditMemoItem
Invoice RatePlanCharge RevenueScheduleItemDebitMemoItem
InvoiceAdjustment RatePlanChargeTier Subscription
InvoiceItemAdjustment Refund SubscriptionProductFeature
InvoicePayment RefundInvoicePayment UpdaterDetail (PaymentMethodUpdate)
JournalEntryItem RefundTransactionLog Usage
Payment RevenueChargeSummaryItem
PaymentGatewayReconciliationEventLog RevenueEventItem
PaymentMethod RevenueEventItemCreditMemoItem
PaymentMethodTransactionLog RevenueEventItemDebitMemoItem

Last modified: May 24 2018 16:57:09 UTC

If this article is incorrect or outdated, or omits critical information, let us know. For all other issues, access our support channels.