Open Source AutoMQ 1.6.0 Released !! 🚀🚀🚀17x Kafka Cost Reduction, Optimized Iceberg Support and Strmzi Operator Support. Learn more from here.
Kafka Linking

Seamlessly Migrate from any Apache Kafka® Compatible Service to AutoMQ with Zero Downtime

Kafka Linking plays a crucial role in various scenarios including hybrid cloud deployments, disaster recovery, and cluster migration. AutoMQ's Kafka Linking capability can connect AutoMQ with any other streaming system compatible with the Kafka protocol, enabling the replication of data and metadata between the two clusters. This feature supports a variety of Kafka Linking applications and use cases.

Zero Downtime Migration

AutoMQ, through innovative cluster proxy technology, allows the migration of data from the old cluster to the new cluster without affecting the applications currently running in the production environment, ensuring a seamless migration.

Offset-Preserving Migration

For data infrastructures that rely on Kafka consumer offsets, such as Flink jobs, it is particularly important to preserve these offsets during cluster migration. This allows users to migrate Kafka to AutoMQ without making any changes to these existing data infrastructures, significantly simplifying cluster migration.

Fully Built-in Capabilities

Kafka Linking is a fully built-in capability of AutoMQ, with no external dependencies. Users do not need to maintain connectors or linking pipelines themselves.

Synchronize All Data

All data and metadata of the Kafka cluster will be synchronized, including Topics, Consumer Groups and Offsets.

Compatible with Kafka Alternatives

AutoMQ is a new generation of Kafka that is 100% compatible with Apache Kafka. It is compatible with Kafka alternatives on the market, such as Confluent, Aiven, WarpStream, and Redpanda.

Easy-to-Use Operation UI

AutoMQ provides a user-friendly and intuitive UI. Even users unfamiliar with Kafka can easily configure Kafka Linking.

Seamless Cluster Migration Experience

AutoMQ supports seamless migration from existing Kafka or AutoMQ clusters to new AutoMQ clusters. Whether driven by the need for architectural evolution or the desire to migrate specific topics from old to new clusters, the entire process retains consumer offsets and synchronizes all data and metadata. This feature can be configured through the AutoMQ Enterprise Edition console UI.

Disaster Recovery Cluster

Establish a Disaster Recovery (DR) cluster to take over in case of an outage or catastrophe in your primary cluster. The synchronization feature ensures that the DR cluster is always up-to-date with data, metadata, topic structure, configurations, and consumer offsets. This enables low Recovery Point Objectives (RPOs) and Recovery Time Objectives (RTOs). Maintaining a DR cluster is straightforward without relying on external components or intricate management. Thanks to the Kafka Linking feature, which preserves consumer offsets during migration, consumers can failover and recover near the point of interruption, minimizing downtime without additional coding efforts.

Offload Workload to Read Replicas

By setting up a read-only Kafka cluster, you can offload analytical or batch jobs to isolated hardware, enhancing Kafka's fanout capabilities. This cluster is scalable in seconds, providing virtually infinite read throughput on demand.

Geo-Replication for Data Locality

Geographically distributing data allows consumers to access data locally, providing lower read latency and improving overall performance.

Start Your AutoMQ Journey Today

Contact us to schedule an online meeting to learn more, request PoC assistance, or arrange a demo.