Skip to main content
Version: 4.x

Connect to Confluent Cloud

This guide walks you through connecting VeloDB Cloud to Confluent Cloud Kafka using the visual interface.

warning

Prerequisites: Before proceeding, complete the Confluent Cloud Setup Guide to create your cluster, API keys, and sample data topic.

Step 1: Navigate to Import

In your VeloDB warehouse, go to the left sidebar and find the Data section, then click Import.

Click Create to start a new import job.

New Import Page

Step 2: Select Confluent Cloud

Under Event Streams, click Confluent Cloud to start the streaming import wizard.

Step 3: Configure Connection

Fill in the connection details from Confluent Cloud Setup:

Connection Configuration

FieldDescriptionExample
Task NameUnique name for this import jobconfluent_orders
Consumer GroupKafka consumer group IDvelodb-consumer
BrokersBootstrap server URLpkc-xxxxx.us-east-1.aws.confluent.cloud:9092
API KeyConfluent Cloud API KeyYour API Key
API SecretConfluent Cloud API SecretYour API Secret
SASL MechanismAuthentication methodPLAIN

Click Next after filling in all fields.

Step 4: Select Source Data

Configure the Kafka topic and data format:

Source Data Configuration

FieldDescription
TopicSelect your Kafka topic from the dropdown
Offset SelectionFrom beginning to load all data, or From latest for new data only
Data FormatSelect JSON
Data Format

VeloDB only supports JSON and CSV formats. If your Confluent topic uses AVRO, you'll need to create a new topic with JSON format in Confluent Cloud.

Click Next to proceed.

Step 5: Configure Destination Table

VeloDB automatically detects the schema from your Kafka messages:

Data Preview

The Data Preview section shows sample records from your topic.

Configure Table Settings

Destination Table

FieldDescription
Load Data toSelect New Table to create a new table
DatabaseSelect or create a database
TableEnter a table name

Column Settings

Review and adjust the column mappings:

Column Settings

SettingDescription
Source FieldJSON field from Kafka messages
Column NameVeloDB table column name
Data TypeVeloDB data type (auto-detected)

Advanced Settings

SettingDescription
Table ModelsDUPLICATE for append-only, UNIQUE for upserts
Sorting KeyColumn(s) for data ordering
Bucket KeyColumn(s) for data distribution
Bucket NumberAUTO recommended

Click Next to continue.

Step 6: Configure Settings

Adjust the import job settings:

Settings

SettingDefaultDescription
Concurrency256Number of parallel consumers
Max Batch Interval (s)60Maximum wait time before committing
Max Batch Rows20000000Maximum rows per batch
Max Batch Size (MB)1024Maximum batch size

The defaults work well for most use cases. Click Next to proceed.

Step 7: Verification

VeloDB validates your configuration:

Verification

The checklist verifies:

  • Kafka Validation - Connection to Confluent Cloud
  • Warehouse Connectivity Test - VeloDB can reach Kafka
  • User Permission Check - You have permission to create the import

If all checks show Succeeded, click Start to begin streaming data.


Verify Data Import

After starting the import, verify data is flowing:

Check Import Status

Go to Import in the sidebar to see your job status:

  • RUNNING - Job is actively consuming data
  • PAUSED - Job is paused (check for errors)

Query Your Data

Navigate to SQL Editor and run:

-- Check row count
SELECT COUNT(*) FROM your_database.your_table;

-- View sample data
SELECT * FROM your_database.your_table LIMIT 10;

Manage Import Jobs

ActionHow
PauseClick the job, then Pause
ResumeClick the job, then Resume
DeleteClick the job, then Delete

Or use SQL:

-- Pause job
PAUSE ROUTINE LOAD FOR database.job_name;

-- Resume job
RESUME ROUTINE LOAD FOR database.job_name;

-- Stop job
STOP ROUTINE LOAD FOR database.job_name;

-- View job status
SHOW ROUTINE LOAD FOR database.job_name;

Troubleshooting

IssueSolution
"Incorrect credentials"Verify API Key and Secret from Confluent Cloud
"Broker transport failure"Ensure SASL Mechanism is set to PLAIN
"Topic not found"Check topic name matches exactly (case-sensitive)
JSON parse errorEnsure your Confluent topic uses JSON format, not AVRO
Job paused with errorsCheck SHOW ROUTINE LOAD for error details

References