Loading Events in Batches

For some Destinations, Events may be loaded in batches of files to improve the performance. This is specially applicable to data warehouse Destinations such as the following:

  • Amazon Redshift

  • Google BigQuery

  • Snowflake

The writes to the warehouses include scanning of the tables for deduplication of the Events, which incurs costs for users. Major cloud-based data warehouses, such as, Amazon Redshift and Google BigQuery recommend loading Events through files in batches. Batches provide much better performance at a much lower cost compared to direct and individual writes to the tables.

Advantages of loading Events in batches

  • Batches allow Hevo to load millions of Events in the warehouse without consuming a lot of resource bandwidth.

  • Loading in batches is faster at scale than direct inserts.

  • Deduplication needs to be done fewer times for batches as compared to individual records.

Disadvantages of loading Events in batches

The batching process understandably introduces some delay in loading the data. The delay usually varies between 5-15 minutes. This means that once an Event is ingested by the Pipeline, and provided it is mapped and does not encounter any other failure, the Event should be visible in the Destination within 5-15 minutes.

In case you have stricter SLAs in terms of data latency, contact Hevo Support.

Last updated on 19 Sep 2022

Tell us what went wrong