Amplitude Analytics

Amplitude Analytics helps generate thorough product analytics of web and mobile application usages to help you make data driven decisions. You can replicate data from your Amplitude account to a database, data warehouse, or file storage system using Hevo Pipelines.

For creating Pipelines using this Source, Hevo provides you a fully managed BigQuery data warehouse as a possible Destination. This option remains available till the time you set up your first BigQuery Destination irrespective of any other Destinations that you may have. With the managed warehouse, you are only charged the cost that Hevo incurs for your project in Google BigQuery. The invoice is generated at the end of each month and payment is recovered as per the payment instrument you have set up. You can now create your Pipeline and directly start analyzing your Source data. Read Hevo Managed Google BigQuery.


  • An active account on Amplitude with access to at least one project.

Configuring Amplitude Analytics as a Source

Perform the following steps to configure Amplitude Analytics as the Source in your Pipeline:

  1. Click PIPELINES in the Asset Palette.

  2. Click + CREATE in the Pipelines List View.

  3. In the Select Source Type page, select Amplitude Analytics.

  4. In the Configure your Amplitude Analytics Source page, specify the following:

    Amplitude settings in Hevo

  5. Click TEST & CONTINUE.

  6. Proceed to configuring the data ingestion and setting up the Destination.

Retrieving the Amplitude API Key and Secret

  1. Log in to your Amplitude account.

  2. In the left navigation pane, scroll down and click Settings.

    Amplitude settings

  3. In the Org Settings page, click Projects in the left pane, and select a project whose data you would like to sync:

    Select project

  4. In the project details, copy the API Key and Secret Key shown on the screen, and save these securely:

    API Key details

Data Replication

Default Pipeline Frequency Minimum Pipeline Frequency Maximum Pipeline Frequency Custom Frequency Range (Hrs)
1 Hr 15 Mins 24 Hrs 1-24

Note: The custom frequency must be set in hours, as an integer value. For example, 1, 2, 3 but not 1.5 or 1.75.

  • Events Data: Data for all the Events or user actions that occur in the account are ingested every hour. This includes information relating to:

    • Attributions.

    • Users.

  • Event Properties: A full load is ingested every hour.

Data Model

The following is the list of tables (objects) that are created at the Destination when you run the Pipeline.

Table Name Description
Cohort A list of all unique behavioural cohorts created within Amplitude
Event An action that a user takes in your product. This could be anything from pushing a button, completing a level, or making a payment
Event Category All event data is mapped to an Event Category entity which helps to categorise and describe live events and properties.
Event Type All events are mapped to an Event Type entity which is maintained in this table.
Group Each grouping of users that is created in Amplitude along with their dedicated name and description.
User Any person who has logged at least one event and to whom events are attributed.
User Cohort A mapping between User and the User Cohort they belong in.
User Group Groups of users defined by their actions within a specific time period.

Schema and Primary Keys

Hevo uses the following schema to upload the records in the Destination:

The User object defines each unique user through a combination of User ID, Amplitude ID, and Device ID. You can reference these three columns while making joins to the Event object.


  • There is a two hour delay in the data exported from Amplitude Analytics getting loaded into your data warehouse.

    For example, data sent between 8-9 PM begins to load at 9 PM and becomes available in your Destination after 11 PM, depending on the load frequency you have set.

Revision History

Refer to the following table for the list of key updates made to this page:

Date Release Description of Change
Oct-25-2021 NA Added the Pipeline frequency information in the Data Replication section.
Apr-06-2021 1.60 - Added a note to the section Schema and Primary Keys
- Updated the ERD. The User object now has three fields, user_id, amplitude_id and device_id as primary keys. The field uuid in the Event object is also a primary key now.
Last updated on 22 Oct 2021