- Getting Started
- Using Hevo
-
Pipelines
- Familiarizing with the Pipelines UI
- Pipeline Concepts
- Working with Pipelines
- Transformations
-
Schema Mapper
- Using Schema Mapper
- Mapping Statuses
- Auto Mapping Event Types
- Mapping a Source Event Type with a Destination Table
- Mapping a Source Event Type Field with a Destination Table Column
- Schema Mapper Actions
- Fixing Unmapped Fields
- Resolving Incompatible Schema Mappings
- Resizing String Columns in the Destination
- Creating File Partitions for S3 Destination through Schema Mapper
- Schema Mapper Compatibility Table
- Sources
- Destinations
- Transform
- Activate
- Concepts and Reference
- Account Management
- About Hevo
- FAQs
-
Release Notes
- Release Version 1.60 (06-Apr-2021)
- Release Version 1.59 (23-Mar-2021)
- Release Version 1.58 (09-Mar-2021)
- Release Version 1.57 (22-Feb-2021)
- Release Version 1.56 (09-Feb-2021)
- Release Version 1.55 (25-Jan-2021)
- Release Version 1.54 (12-Jan-2021)
- Release Version 1.53 (22-Dec-2020)
- Release Version 1.52 (03-Dec-2020)
- Release Version 1.51 (10-Nov-2020)
- Release Version 1.50 (19-Oct-2020)
- Release Version 1.49 (28-Sep-2020)
- Release Version 1.48 (01-Sep-2020)
- Release Version 1.47 (06-Aug-2020)
- Release Version 1.46 (21-Jul-2020)
- Release Version 1.45 (02-Jul-2020)
- Release Version 1.44 (11-Jun-2020)
- Release Version 1.43 (15-May-2020)
- Release Version 1.42 (30-Apr-2020)
- Release Version 1.41 (Apr-2020)
- Release Version 1.40 (Mar-2020)
- Release Version 1.39 (Feb-2020)
- Release Version 1.38 (Jan-2020)
Creating a Pipeline
The settings that you must define to create a Pipeline are displayed based on multiple factors, such as:
-
The type of Source. For example, JDBC, SaaS, or File-based. Read each Source for the settings required for it.
-
The format of the Source data, and the format you require to maintain at the Destination. For example, the format in which JSON Events may be parsed, or the format the data is in (XML, CSV) in an Amazon S3 Source. Read Parsing Nested JSON Fields in Events.
-
The amount and type of data you want to ingest. For example, historical data vs incremental data. Read Types of Data Synchronization.
-
The way in which you want Hevo to read your data. Read Pipeline Modes and Query Modes for Ingesting Data from Relational Databases.
Note: You can load data from the same Source and Destination instance, but you cannot move data between the same database via the Pipeline.
See Also
Was this page helpful?
Thank you for helping improve Hevo's documentation. If you need help or have any questions, please consider contacting support.