Rays

Kafka SQL Real-Time Queries

Real-time analytics over your Kafka topics with Tinybird's fully managed Kafka Connector. Instant, hassle-free real-time data analytics
Try for free
kafka_project
$ ❯ tb

Fast time-to-value, fully managed and auto-scaled

Connect a topic, ingest, aggregate, transform or query with SQL, publish APIs in minutes
Tinybird Kafka Connector Diagram
Rays

Real Kafka Use Cases

Your Kafka Pipeline, Without the Ops Pain

Connect, store, process and query your data with a platform built to serve APIs at petabyte scale.
Try for free

scaled to petabytes via Kafka with 7 use cases in 7 months.

6.45PBdata
ingested
15.27Trows
ingested
<0.2singest
latency

Our devs evaluated Pinot, Druid and Tinybird. They preferred Tinybird for performance, reliability, integrations, and developer experience.

Damian Grech

Damian Grech

Director of Engineering, Data Platform at FanDuel

Background

Enterprise features

Why Tinybird Kafka Connector

Tinybird Kafka ConnectorClickHouse® OSS Kafka Engine
Pricing & Support
Pricing ModelThroughput basedCluster resource pricing
Topic LimitUnlimitedLimited by infra
Enterprise support
Scaling & performance
Serverless
Consumer autoscaling
Compute-storage separation
Push/Pull ModelPushPull
Sub-second flushing
Operations & failure handling
Circuit breaker & Backpressure
High AvailabilityDefault HA
DevEx
CLI + Tinybird codeData as code
Environment branchingSecrets, consumer group IDs, etc.
Schema evolution
RaysDiamonds Background

Build for scale

Tinybird Kafka Connector Features

Ingest, scale, and ship a production-grade Kafka connector built for reliability, correctness, and developer velocity.

Scaling & performance

  • Serverless
  • Consumer autoscaling
  • Sub-second latency flushing*

Operations & failure handling

  • Zero ops (fully managed)
  • Debug metadata
  • Clear failure modes and recovery paths

Data correctness & offsets

  • Offset management
  • Builtin backpressure
  • Dead letter queue (quarantine)

Enterprise ready

  • Schema Registry integration
  • Kafka security protocols: SASL/SSL...
  • Private network supported*

Observability

  • Lag, latency and throughput metrics
  • Error logs
  • Grafana, Prometheus enabled

Developer experience

  • CLI + Tinybird Code
  • Environments & Data Branching
  • Local development + CI/CD

* Talk to us for sizing and enterprise requirements

Get started

Supported Kafka vendors

Pick your Kafka vendor and follow the guide. Tinybird will begin consuming your topics and exposing them as low-latency APIs in minutes.
[Read the docs]
Supported Kafka Vendors

FAQs

How do I create new connections?

Create a Kafka connection using the CLI with `tb datasource create --kafka` for a guided process, or manually create .connection and .datasource files. Your credentials are stored securely as secrets. Tinybird supports multiple authentication methods including SASL for cloud providers like Confluent Cloud, Amazon MSK, and others.

Is it a managed connector or do I have to run something myself?

Tinybird's Kafka connector is fully managed—you don't run anything. We handle consumer autoscaling, offset management, failure recovery, and high availability. Just provide your Kafka credentials and start ingesting.

Can I test ingestion locally or in different environments?

Yes. Tinybird Local is a Docker container that replicates the full Tinybird environment, including Kafka connector support. Use `tb dev` for local development, run tests with `tb test`, and deploy to staging/production environments.

How much can I ingest and how do I manage scale?

Tinybird's Kafka connector is serverless with automatic consumer autoscaling. There's no limit on topics or throughput—we scale based on your workload. The default message size limit is 10MB per message. You don't manage scale; we handle it automatically.

How is it priced?

Tinybird uses throughput-based pricing. You pay based on the volume of data you ingest and query, so costs scale directly with your actual usage. Tinybird offers a free tier to get started, with paid plans that scale elastically based on throughput.

What plan do I need for X rows/sec or Y GB/day?

The Kafka connector is included in all plans, including the Free tier. With throughput-based pricing, you can start free and scale as your data volume grows. For high-throughput requirements, Enterprise plans offer dedicated infrastructure optimized for large-scale streaming workloads.

What formats are supported?

Tinybird supports JSON and Avro message formats. You can configure the format using KAFKA_VALUE_FORMAT and KAFKA_KEY_FORMAT settings in your datasource file. Compressed messages (gzip) are automatically decoded on ingestion.

Do you support schema registry?

Yes. Tinybird integrates with Confluent Schema Registry for decoding Avro messages. For JSON messages with schema registry, use the KAFKA_VALUE_FORMAT and KAFKA_KEY_FORMAT settings (requires CLI 5.14.0+). Configure your schema registry URL in the .connection file using KAFKA_SCHEMA_REGISTRY_URL.

What happens on ingestion failure?

Messages that fail to match your schema are automatically sent to a quarantine data source (similar to a dead letter queue). Quarantine preserves the original message plus error details, allowing you to inspect issues and replay data after fixing problems. Built-in circuit breakers and backpressure prevent cascading failures.

How do you handle schema evolution?

Use the FORWARD_QUERY instruction in your .datasource file to define how existing data transforms to your new schema. When you deploy, Tinybird applies schema changes with zero downtime—your data remains queryable throughout the migration.

What are the main limits and quotas?

Key limits: 10MB max message size, rate limits on API requests (HTTP 429 when exceeded), and 10-second default query timeout (configurable on paid plans). Messages exceeding limits are quarantined, not dropped.

Tinybird wordmark