Use Treasure Data’s data.ai connector to import data.ai (formerly AppAnnie) data source objects into Treasure Data.
Basic knowledge of Treasure Data
Basic knowledge of data.ai (formerly AppAnnie)
There are 2 different rate limits in data.ai:
call per minute
call per user per day rate limits
The call-per-minute limit auto-refreshes after a certain amount of seconds, while the daily call limit is refreshed on a daily basis at 00:00 PST.
If you have multiple transfers under the same data.ai account, you can control the rate limit usage of each data.ai transfer via both calls_per_minute_limit and calls_per_day_limit settings as long as the total limits are lower than or equal to your account quota. For example, assume that your account has quotas as 100 calls/minute and 10000 calls/day, if you create 2 transfers, for example, product sales & product usage data, you can use 50 cpm & 5000 cpd for product sales transfer and the rest (50 cpm & 5000 cpd) for product usage transfer.
Use TD Console
- Create a New Connection
In Treasure Data, you must create and configure the data connection prior to running your query. As part of the data connection, you provide authentication to access the integration.
Open TD Console.
Navigate to Integrations Hub > Catalog.
- Click the search icon on the far right of the Catalog screen, and enter data.ai.
- Hover over the data.ai connector and select Create Authentication.
The following dialog opens.
The method you use to authenticate Treasure Data with data.ai affects the steps you take to enable the data connector to import from data.ai.
Treasure Data supports:
Using Credentials to Authenticate
Provide your data.ai API key information and select Continue.
Using OAuth to Authenticate
OAuth is only available in the US region.
- Select “OAuth” from the “Authentication Method” dropdown to connect your data.ai account using OAuth 2.
- When you select the OAuth authentication method, select Click here to connect a new account. Log into your data.ai account from the new window:
- And grant access to the Data Connector and Treasure Data app:
- You will be redirected back to the Catalog. Repeat the Create a new connection step and choose your new OAuth connection.
When you complete the connection form, select Continue and give your connection a name:
Create a New Transfer
After creating the connection, you are automatically taken to the Authentications tab. Look for the connection you created and select New Transfer.
The following dialog opens. Complete the details and select Next.
Next, you see a Preview of your data similar to the following dialog. If you want to change any settings, select Advanced Settings otherwise, select Next.
If you want to change some options, such as skipping on errors or rate limits, you do so in Advanced Settings:
Select the database and table where you want to transfer the data, as shown in the following dialog:
Specify the schedule of the data transfer using the following dialog and select Start Transfer:
You see the new data transfer in progress listed under the My Input Transfers tab and a corresponding job is listed in the Jobs section.
Use Command Line
Install ‘td’ command v0.11.9 or later
You can install the newest TD Toolbelt.
Create Configuration File
Prepare a configuration file (for eg: load.yml) with your data.ai account access information, as follows:
This example dumps data.ai Account Data Source:
apikey: data.ai apiKey.
target: data.ai entity object to be imported.
See Appendix: Appendix: Available targets for the list of available targets.
breakdown: Breakdown type for which product sale or usage data is fetched.
This field name is changed according to which target is selected, either breakdown_sales or breakdown_usage.
See Appendix: Available breakdowns for usage and the list of available breakdowns.
fetch_type: The source of products to pull (any products from connected accounts or via sharing or both) to be imported.
See Appendix: Available fetch-types for usage and the list of available fetch_type.
start_date: From which date (yyyy-MM-dd) product data is to be imported. This field is required if either fetching product usage (target is product_usage) or product sale (target is product_sales) with an in-app-purchase breakdown (breakdown has iap).
end_date: Until what date (yyyy-MM-dd) product data is to be imported? This field is optional and is automatically adjusted to a maximum of 60 days since start_date based on the current date.
currency: The monetary currency in which the data is presented.
See Appendix: Available currencies for the list of available currencies.
skip_on_invalid_records: Ignore errors (such as invalid JSON, unsupported data) and continue fetching record. (false by default)
calls_per_minute_limit / calls_per_day_limit: Limit number of API calls per minute / per day
See Appendix: Rate Limits for how to use these options
For more details on available out modes, see the Appendix: Modes for out Plugin
Optionally, Preview Data to Import
You can preview data to be imported using the command td connector:preview.
Execute Load Job
Submit the load job. It may take a couple of hours depending on the data size. Users need to specify the database and table where their data are stored.
It is recommended to specify --time-column option, since Treasure Data’s storage is partitioned by time (see also architecture) If the option is not given, the Data Connector will choose the first long or timestamp column as the partitioning time. The type of the column specified by --time-column must be either of long or timestamp type.
If your data doesn’t have a time column you can add it using the add_time filter option. More details at add_time Filter Plugin for Integrations.
The preceding command assumes that you have already created database(td_sample_db) and table(td_sample_table). If the database or the table do not exist in TD, this command will not succeed, so create the database and table manually or use --auto-create-table option with td connector:issue command to auto create the database and table:
You can assign the Time Format column to the "Partitioning Key" by the "--time-column" option.
You can schedule periodic Data Connector execution for periodic data.ai import. We configure our scheduler carefully to ensure high availability. By using this feature, you no longer need a
cron daemon on your local data center.
Create the Schedule
A new schedule can be created using the td connector:create command. The name of the schedule, cron-style schedule, the database and table where their data will be stored, and the Data Connector configuration file are required.
The `cron` parameter also accepts these three options: `@hourly`, `@daily`, and `@monthly`.
By default, the schedule is set up in the UTC timezone. You can set the schedule in a timezone using -t or --timezone option. The `--timezone` option supports only extended timezone formats like 'Asia/Tokyo', 'America/Los_Angeles', etc. Timezone abbreviations like PST and CST are *not* supported and may lead to unexpected schedules.
List the Schedules
You can see the list of scheduled entries by td connector:list.
Show the Setting and History of Schedules
td connector:show displays the execution setting of a schedule entry.
td connector:history shows the execution history of a schedule entry. To investigate the results of each individual execution, use td job <jobid>.
Delete the Schedule
td connector:delete removes the schedule.
Modes for Out Plugin
You can specify file import mode in the out section of load.yml.
The following examples show the default mode and records are appended to the target table.
replace (In td 0.11.10 and later)
This mode replaces data in the target table. Any manual schema changes made to the target table remain intact with this mode.
Products from connected accounts
Shared products from external accounts
Product sales data
Product usage data
This field is available for importing only product sales or product usages.
If the target is product_sales, the breakdown field name is breakdown_sales
If the target is product_usage, the breakdown field name is breakdown_usage
Available Fetch Types
This field is available for importing product sales, product usage and app details.
Import only data of products from connected accounts
Import only data of products from sharing list
Import both product sources
This field is available for importing only product sales. Contact data.ai support for more detail if needed.
Full Name of Currency
Hong Kong dollar
Israeli new shekel
South Korean won
New Zealand dollar
Romanian new leu
New Taiwan dollar
United States dollar
South African rand