On This Page
Amplitude Analytics helps generate thorough product analytics of web and mobile application usages to help you make data driven decisions. You can replicate data from your Amplitude account to a database, data warehouse, or file storage system using Hevo Pipelines.
For creating Pipelines using this Source, Hevo provides you a fully managed BigQuery data warehouse as a possible Destination. This option remains available till the time you set up your first BigQuery Destination irrespective of any other Destinations that you may have. With the managed warehouse, you are only charged the cost that Hevo incurs for your project in Google BigQuery. The invoice is generated at the end of each month and payment is recovered as per the payment instrument you have set up. You can now create your Pipeline and directly start analyzing your Source data. Read Hevo Managed Google BigQuery.
- An active account on Amplitude with access to at least one project.
Configuring Amplitude Analytics as a Source
Perform the following steps to configure Amplitude Analytics as the Source in your Pipeline:
Click PIPELINES in the Asset Palette.
Click + CREATE in the Pipelines List View.
In the Select Source Type page, select Amplitude Analytics.
In the Configure your Amplitude Analytics Source page, specify the following:
Pipeline Name: A unique name for your Pipeline.
API Key: The API key you retrieved from your Amplitude account.
Secret Key: The secret key you retrieved from your Amplitude account.
Historical Load: The date offset from when the data ingestion must start. By default, the date 12 months prior is displayed.
Click TEST & CONTINUE.
Proceed to configuring the data ingestion and setting up the Destination.
Retrieving the Amplitude API Key and Secret
Log in to your Amplitude account.
In the left navigation pane, scroll down and click Settings.
In the Org Settings page, click Projects in the left pane, and select a project whose data you would like to sync:
In the project details, copy the API Key and Secret Key shown on the screen, and save these securely:
Events Data: Data for all the Events or user actions that occur in the account are ingested every hour. This includes information relating to:
Event Properties: A full load is ingested every hour.
The following is the list of tables (objects) that are created at the Destination when you run the Pipeline.
|Cohort||A list of all unique behavioural cohorts created within Amplitude|
|Event||An action that a user takes in your product. This could be anything from pushing a button, completing a level, or making a payment|
|Event Category||All event data is mapped to an Event Category entity which helps to categorise and describe live events and properties.|
|Event Type||All events are mapped to an Event Type entity which is maintained in this table.|
|Group||Each grouping of users that is created in Amplitude along with their dedicated name and description.|
|User||Any person who has logged at least one event and to whom events are attributed.|
|User Cohort||A mapping between User and the User Cohort they belong in.|
|User Group||Groups of users defined by their actions within a specific time period.|
Schema and Primary Keys
Hevo uses the following schema to upload the records in the Destination:
User object defines each unique user through a combination of User ID, Amplitude ID, and Device ID. You can reference these three columns while making joins to the
There is a two hour delay in the data exported from Amplitude Analytics getting loaded into your data warehouse.
For example, data sent between 8-9 PM begins to load at 9 PM and becomes available in your Destination after 11 PM, depending on the load frequency you have set.
Refer to the following table for the list of key updates made to this page:
|Date||Release||Description of Change|
|Apr-06-2021||1.60||- Added a note to the section Schema and Primary Keys
- Updated the ERD. The