Skip to main content
Version: 26.x

Import

The Import module provides a unified, visual interface for data integration. It allows you to seamlessly ingest data from external sources into your VeloDB Cloud target data warehouse. Supporting both batch data loading and continuous data ingestion with automatic table creation, this module significantly lowers the barrier to building data pipelines.

You can access this feature by navigating to Data -> Import in the left sidebar of VeloDB Studio, and click the Add Import Job button to create a new task.

Supported Data Sources

Depending on your business scenarios and data storage types, VeloDB Cloud currently supports importing data from three main categories:

  • Databases
    • MySQL
    • PostgreSQL
  • Object Storage
    • Object Storage S3
  • Event Streams
    • Confluent Cloud
    • Apache Kafka
    • Amazon MSK

Creating an Import Job

Although configuration details vary slightly by data source, VeloDB Cloud provides a consistent, 5-step wizard-driven experience. Here is a general guide using a database import as an example:

1. Select a Data Source

  1. Log in to the VeloDB Cloud console and click Import in the left navigation pane under the Data section,
  2. Click the + Add Import Job button on the top right.
  3. Select your desired data source card (e.g., PostgreSQL) from the corresponding category.

import-select-source

2. Configure Source Data Connection

Fill in the authentication details to establish a secure connection with your data source.

  • Job Name​: Specify a globally unique name for your import job for easy management and identification.
  • Connection Credentials​: Depending on the source, enter details such as Host, Port, Database Name, Username, Password, or AK/SK keys.
  • Click ​Next​. The system will automatically verify network connectivity and credential validity in the background.

import-configure-connection

3. Configure Sync Scope (Tables / Topics)

Define what data to pull and how it should be written.

  • Sync Type​: Choose between Full + Incremental sync or Incremental only.
  • Data Scope​:
    • For Databases​: Select the schemas and tables to sync. You can expand tables to apply column-level filtering under Column Settings.
    • For Event Streams​: Specify the topics to consume and the starting offset.
    • For Object Storage​: Specify the bucket path and file matching patterns.
  • Sync Interval​: Set the frequency for data fetching (the default is typically 60 seconds).

import-sync-scope

4. Verify and Confirm

To ensure the job runs successfully, the system automatically runs a pre-check Checklist.

  • Checks typically include: network connectivity tests, source-side permission validation (e.g., WAL/Binlog permissions required for CDC), and target-side table creation/write permissions.
  • Once all checklist items display a green ​Succeeded​, click Start to officially launch the data import job.

import-verify

5. Manage and Monitor

After starting the job, you will be redirected back to the Import jobs list.

  • View Status​: Monitor the real-time Status (e.g., *Running*, *Paused*, *Failed*) and creation time of your jobs.
  • Actions​: Click the ... (Actions) menu next to a job to ​Pause​, ​Resume​, or Delete the import task as your business needs change.

import-manage