Alternatives to Yandex Managed Service for Apache Kafka
Compare Yandex Managed Service for Apache Kafka alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Yandex Managed Service for Apache Kafka in 2026. Compare features, ratings, user reviews, pricing, and more from Yandex Managed Service for Apache Kafka competitors and alternatives in order to make an informed decision for your business.
-
1
Rocket iCluster
Rocket Software
Rocket iCluster high availability/disaster recovery (HA/DR) solutions ensure uninterrupted operation for your IBM i applications, providing continuous access by monitoring, identifying, and self-correcting replication problems. iCluster’s multiple-cluster administration console monitors events in real-time on the classic green screen and the modern web UI. Rocket iCluster reduces downtime related to unexpected IBM i system interruptions with real-time, fault-tolerant, object-level replication. In the event of an outage, you can bring a “warm” mirror of a clustered IBM i system into service within minutes. iCluster disaster recovery software ensures a high-availability environment by giving business applications concurrent access to both master and replicated data. This setup allows you to offload critical business tasks such as running reports and queries as well as ETL, EDI, and web tasks from your secondary system without affecting primary system performance. -
2
Get access to new functionalities, security patches, and other Elasticsearch improvements with official subscriptions. You can deploy a ready-to-use cluster in just a few minutes. Elasticsearch and Kibana stack settings are initially optimized for the cluster size you selected. Invest your time in your project, and we’ll take care of cluster maintenance, software backups, monitoring, fault tolerance, and updates. Index sharding reduces the load on each host and makes it easy to scale a cluster under peak loads. It’s much easier to develop an infrastructure if you have a visual representation of system behavior. Create trends, make forecasts, and evaluate system stability in a user-friendly interface. To create fault-tolerant geo-distributed Elasticsearch and Kibana clusters, just select the number of hosts and specify availability zones. Select the necessary computing capacity and create a ready-to-use Elasticsearch cluster.Starting Price: $117.79 per month
-
3
Apache Kafka
The Apache Software Foundation
Apache Kafka® is an open-source, distributed streaming platform. Scale production clusters up to a thousand brokers, trillions of messages per day, petabytes of data, hundreds of thousands of partitions. Elastically expand and contract storage and processing. Stretch clusters efficiently over availability zones or connect separate clusters across geographic regions. Process streams of events with joins, aggregations, filters, transformations, and more, using event-time and exactly-once processing. Kafka’s out-of-the-box Connect interface integrates with hundreds of event sources and event sinks including Postgres, JMS, Elasticsearch, AWS S3, and more. Read, write, and process streams of events in a vast array of programming languages. -
4
Apache Helix
Apache Software Foundation
Apache Helix is a generic cluster management framework used for the automatic management of partitioned, replicated and distributed resources hosted on a cluster of nodes. Helix automates reassignment of resources in the face of node failure and recovery, cluster expansion, and reconfiguration. To understand Helix, you first need to understand cluster management. A distributed system typically runs on multiple nodes for the following reasons: scalability, fault tolerance, load balancing. Each node performs one or more of the primary functions of the cluster, such as storing and serving data, producing and consuming data streams, and so on. Once configured for your system, Helix acts as the global brain for the system. It is designed to make decisions that cannot be made in isolation. While it is possible to integrate these functions into the distributed system, it complicates the code. -
5
Invest your time in your project, and we’ll take care of database maintenance: software backups, monitoring, fault tolerance, and updates. ClickHouse is great at handling queries to large amounts of data in real time, while column-based storage saves space due to strong data compression. All DBMS connections are encrypted using the TLS protocol. Data is secured in accordance with the requirements of local regulatory, GDPR, and ISO industry standards. Visualize the data structure in your ClickHouse cluster and send SQL queries to databases from the management console. The service also provides data replication between database hosts (both inside and between availability zones) and automatically switches the load over to a backup replica in the event of a failure.Starting Price: $42.51 per month
-
6
You can deploy a ready-to-use cluster in just a few minutes and just as easily set up automatic backups and replicas. Invest your time in your project, and we’ll take care of database maintenance: software backups, monitoring, fault tolerance, and updates. With Managed Service for MySQL®, you can easily create read replicas and build custom replication topologies. Data is stored in encrypted form in accordance with the requirements of local regulatory, GDPR, and ISO industry standards. Monitor your cluster status and resource consumption using the management console or export resource metrics via the API.Starting Price: $41.59 per month
-
7
SafeKit
Eviden
Evidian SafeKit is a high-availability software solution designed to ensure the redundancy of critical applications on Windows and Linux platforms. It provides an all-in-one approach by integrating load balancing, synchronous real-time file replication, automatic application failover, and automated failback after a server failure, all within a single software product. This eliminates the need for additional hardware components such as network load balancers or shared disks, as well as the necessity for enterprise editions of operating systems and databases. SafeKit's software clustering facilitates the creation of mirror clusters with real-time data replication and failover, farm clusters with load balancing and failover, and advanced architectures like farm+mirror clusters and active-active clusters. Its shared-nothing architecture simplifies deployment, even in remote sites, by avoiding the complexities associated with shared disk clusters. -
8
Submariner
Submariner
As Kubernetes gains adoption, teams are finding they must deploy and manage multiple clusters to facilitate features like geo-redundancy, scale, and fault isolation for their applications. With Submariner, your applications and services can span multiple cloud providers, data centers, and regions. The Broker must be deployed on a single Kubernetes cluster. This cluster’s API server must be reachable by all Kubernetes clusters connected by Submariner. It can be a dedicated cluster, or one of the connected clusters. Once Submariner is deployed on a cluster with the proper credentials to the Broker it will exchange Cluster and Endpoint objects with other clusters (via push/pull/watching), and start forming connections and routes to other clusters. Worker node IPs on all connected clusters must be outside of the Pod/Service CIDR ranges. -
9
Apache Geode
Apache
Build high-speed, data-intensive applications that elastically meet performance requirements at any scale. Take advantage of Apache Geode's unique technology that blends advanced techniques for data replication, partitioning and distributed processing. Apache Geode provides a database-like consistency model, reliable transaction processing and a shared-nothing architecture to maintain very low latency performance with high concurrency processing. Data can easily be partitioned (sharded) or replicated between nodes allowing performance to scale as needed. Durability is ensured through redundant in-memory copies and disk-based persistence. Super fast write-ahead-logging (WAL) persistence with a shared-nothing architecture that is optimized for fast parallel recovery of nodes or an entire cluster. -
10
Memurai
Memurai
Redis for Windows alternative, In-Memory Datastore Ready for the most demanding production workloads. Free for development and testing. Fully Redis-compatible. The core of Memurai is based on the Redis source code, ported to run natively on Windows. Memurai reliably supports all the features that make Redis the most popular NoSQL data store, including LRU eviction, persistence, replication, transactions, LUA scripting, high-availability, pub/sub, cluster, modules, and streams. A lot of attention has been put into ensuring full compatibility, including with the myriad of libraries and tools already available for Redis. You can even replicate data between Memurai and Redis, or use both within the same cluster! Seamless integration with Windows infrastructure and workflows. Whether it's used for development or production, Memurai seamlessly integrates with Windows best practices, tools and workflows. Engineering teams with an existing investments in the Windows infrastructure will be -
11
Tungsten Clustering
Continuent
Tungsten Clustering is the only complete, fully-integrated, fully-tested MySQL HA, DR and geo-clustering solution running on-premises and in the cloud combined with industry-best and fastest, 24/7 support for business-critical MySQL, MariaDB, & Percona Server applications. It allows enterprises running business-critical MySQL database applications to cost-effectively achieve continuous global operations with commercial-grade high availability (HA), geographically redundant disaster recovery (DR) and geographically distributed multi-master. Tungsten Clustering includes four core components for data replication, data connectivity, cluster management and cluster monitoring. Together, they handle all of the messaging and control of your Tungsten MySQL clusters in a seamlessly-orchestrated fashion. -
12
GridGain
GridGain Systems
The enterprise-grade platform built on Apache Ignite that provides in-memory speed and massive scalability for data-intensive applications and real-time data access across datastores and applications. Upgrade from Ignite to GridGain with no code changes and deploy your clusters securely at global scale with zero downtime. Perform rolling upgrades of your production clusters with no impact on application availability. Replicate across globally distributed data centers to load balance workloads and prevent downtime from regional outages. Secure your data at rest and in motion, and ensure compliance with security and privacy standards. Easily integrate with your organization's authentication and authorization system. Enable full data and user activity auditing. Create automated schedules for full and incremental backups. Restore your cluster to the last stable state with snapshots and point-in-time recovery. -
13
Amazon MSK
Amazon
Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed service that makes it easy for you to build and run applications that use Apache Kafka to process streaming data. Apache Kafka is an open-source platform for building real-time streaming data pipelines and applications. With Amazon MSK, you can use native Apache Kafka APIs to populate data lakes, stream changes to and from databases, and power machine learning and analytics applications. Apache Kafka clusters are challenging to setup, scale, and manage in production. When you run Apache Kafka on your own, you need to provision servers, configure Apache Kafka manually, replace servers when they fail, orchestrate server patches and upgrades, architect the cluster for high availability, ensure data is durably stored and secured, setup monitoring and alarms, and carefully plan scaling events to support load changes.Starting Price: $0.0543 per hour -
14
SIOS LifeKeeper
SIOS Technology Corp.
SIOS LifeKeeper for Windows is a comprehensive high-availability and disaster‑recovery solution that integrates failover clustering, continuous application monitoring, data replication, and flexible recovery policies to deliver 99.99 % uptime for Microsoft Windows Server environments—whether physical, virtual, cloud, hybrid‑cloud, or multicloud. Administrators can build SAN‑based or SANless clusters using a variety of storage types (direct‑attached SCSI, iSCSI, Fibre Channel, or local disk) and choose between local or remote standby servers that support both high availability and disaster recovery. LifeKeeper offers real‑time block‑level replication via bundled DataKeeper, with WAN‑optimized performance that includes nine levels of compression, bandwidth throttling, and integrated WAN acceleration, ensuring efficient replication across cloud regions or over WAN without hardware accelerators. -
15
FlashGrid
FlashGrid
FlashGrid's software solutions are designed to enhance the reliability and performance of mission-critical Oracle databases across various cloud platforms, including AWS, Azure, and Google Cloud. By enabling active-active clustering with Oracle Real Application Clusters (RAC), FlashGrid ensures a 99.999% uptime Service Level Agreement (SLA), effectively minimizing business disruptions caused by database outages. Their architecture supports multi-availability zone deployments, safeguarding against data center failures and local disasters. FlashGrid's Cloud Area Network software facilitates high-speed overlay networks with advanced high availability and performance management capabilities, while their Storage Fabric software transforms cloud storage into shared disks accessible by all nodes in a cluster. The FlashGrid Read-Local technology reduces storage network overhead by serving read operations from locally attached disks, thereby enhancing performance. -
16
Red Hat Data Grid
Red Hat
Red Hat® Data Grid is an in-memory, distributed, NoSQL datastore solution. Your applications can access, process, and analyze data at in-memory speed to deliver a superior user experience. High performance, elastic scalability, always available. Quickly access your data through fast, low-latency data processing using memory (RAM) and distributed parallel execution. Achieve linear scalability with data partitioning and distribution across cluster nodes. Gain high availability through data replication across cluster nodes. Attain fault tolerance and recover from disaster through cross-datacenter geo-replication and clustering. Gain development flexibly and greater productivity with a highly versatile, functionally rich NoSQL data store. Obtain comprehensive data security with encryption and role-based access. Data Grid 7.3.10 provides a security enhancement to address a CVE. You must upgrade any Data Grid 7.3 deployments to version 7.3.10 as soon as possible. -
17
Windows Server Failover Clustering
Microsoft
Failover Clustering in Windows Server (and Azure Local) enables a group of independent servers to work together to improve availability and scalability for clustered roles (formerly known as clustered applications and services). These nodes are interconnected via hardware and software, and if one node fails, another assumes its roles through an automated failover process. Clustered roles are actively monitored and, if they stop functioning, are restarted or migrated to maintain service continuity. The feature also supports Cluster Shared Volumes (CSVs), which provide a unified, distributed namespace and consistent shared storage access across nodes, reducing service disruptions. Typical uses include high‑availability file shares, SQL Server instances, and Hyper‑V virtual machines. Failover Clustering is supported on Windows Server 2016, 2019, 2022, and 2025, and in Azure Local environments. -
18
Google Cloud's Managed Service for Apache Kafka is a fully managed and scalable service that simplifies the deployment, management, and maintenance of Apache Kafka clusters. It automates operational tasks such as provisioning, patching, and scaling, allowing users to focus on building applications without the complexities of infrastructure management. It ensures high availability and reliability by replicating data across multiple zones, safeguarding against potential failures. It also offers seamless integration with other Google Cloud services, enabling users to create robust data processing pipelines. Security is a priority, with features like encryption at rest and in transit, identity, and access management, and network isolation to protect data. Google Cloud Managed Service for Kafka supports both public and private networking configurations, providing flexibility in connectivity options.Starting Price: $0.09 per hour
-
19
A service for managing OpenSearch clusters in Yandex Cloud infrastructure. Take advantage of this popular open source solution and integrate fast and scalable full-text search into your products. Deploy a ready-to-use OpenSearch cluster in just a few minutes, with product settings already optimized for the chosen cluster size. We take care of cluster maintenance: reserves, monitoring, fault tolerance, and software updates. Use our visualization tools to set up analytical dashboards, application monitoring, and alert systems. Connect third-party authentication and authorization services (SAML). The service supports granular configurations for data access levels. Open source code allows us to develop the service together with the community, be the first to provide timely updates and prevent vendor lock-in. OpenSearch is an easily scalable system of open source search and analytical tools. It offers a set of technologies for providing fast search and analytics.Starting Price: $0.012240 per GB
-
20
Get access to new MongoDB features and official releases that are 100% compatible with the platform. If the load on your cluster increases, you can add new servers or increase their capacity in a matter of minutes. Invest your time in your project, and we’ll take care of database maintenance: software backups, monitoring, fault tolerance, and updates. You can enable sharding for clusters that have MongoDB version 4.0 or higher. You can also add and configure individual shards to improve cluster performance. All DBMS connections are encrypted using the TLS protocol, and DB backups are GPG-encrypted. Data is secured in accordance with the requirements of local regulatory, GDPR, and ISO industry standards. MongoDB has no regular tables and stores data as collections of JSON-like documents. This is great for projects where data structures may change during development.
-
21
Aeron
Aeron
Aeron is a high-performance, open source messaging and clustering technology designed to power ultra-low-latency, fault-tolerant distributed systems, particularly in electronic trading and real-time data environments. It focuses on delivering predictable microsecond-level latency and extremely high throughput, enabling applications to process millions of messages per second while maintaining strong reliability. The Aeron suite includes Aeron Transport for high-performance unicast, multicast, and IPC messaging, Aeron Archive for ultra-fast message recording and replay with zero message loss, and Aeron Cluster for fault-tolerant distributed state replication using replicated log architecture. Its brokerless design reduces hardware overhead and operational costs while allowing systems to run on-premises, in the cloud, or in hybrid deployments. Aeron supports multiple programming languages, including Java, C/C++, and .NET.Starting Price: Free -
22
Aiven for Apache Kafka
Aiven
Apache Kafka as a fully managed service, with zero vendor lock-in and a full set of capabilities to build your streaming pipeline. Set up fully managed Kafka in less than 10 minutes — directly from our web console or programmatically via our API, CLI, Terraform provider or Kubernetes operator. Easily connect it to your existing tech stack with over 30 connectors, and feel confident in your setup with logs and metrics available out of the box via the service integrations. A fully managed distributed data streaming platform, deployable in the cloud of your choice. Ideal for event-driven applications, near-real-time data transfer and pipelines, stream analytics, and any other case where you need to move a lot of data between applications — and quickly. With Aiven’s hosted and managed-for-you Apache Kafka, you can set up clusters, deploy new nodes, migrate clouds, and upgrade existing versions — in a single mouse click — and monitor them through a simple dashboard.Starting Price: $200 per month -
23
Corosync Cluster Engine
Corosync
The Corosync Cluster Engine is a group communication system with additional features for implementing high availability within applications. The project provides four C application programming interface features. Closed process group communication model with extended virtual synchrony guarantees for creating replicated state machines; a simple availability manager that restarts the application process when it has failed; a configuration and statistics in-memory database that provides the ability to set, retrieve, and receive change notifications of information; and a quorum system that notifies applications when a quorum is achieved or lost. Our project is used as a high-availability framework by projects such as Pacemaker and Asterisk. We are always looking for developers or users interested in clustering or participating in our project. -
24
Waterstream
SimpleMatter
Waterstream turns your Kafka-compatible platform into a full-fledged MQTT broker. Connect millions of clients to your data streaming platform with no code, no integration pipelines, and no additional storage. Waterstream implements a bidirectional layer between Kafka and MQTT clients. Forget managing external MQTT clusters, integration pipelines to code, and data duplication. Waterstream scales out linearly. For most operations, its nodes don’t depend on each other. Add more instances to support an increasing number of clients. Waterstream requires only Kafka to operate. The built-in persistence benefits of using Kafka are all included: high availability, high throughput, and low latency. -
25
rqlite
rqlite
The lightweight, user-friendly, distributed relational database built on SQLite. Fault tolerance and high availability with zero hassle. rqlite is a distributed relational database that combines the simplicity of SQLite with the robustness of a fault-tolerant, highly available system. It's developer-friendly, its operation is straightforward, and it's designed for reliability with minimal complexity. Deploy in seconds, with no complex configurations. Seamlessly integrates with modern cloud infrastructures. Built on SQLite, the world’s most popular database. Supports full-text search, Vector Search, and JSON documents. Access controls and encryption for secure deployments. Rigorous, automated testing ensures high quality. Clustering provides high availability and fault tolerance. Automatic node discovery simplifies clustering. -
26
Google Cloud Bigtable
Google
Google Cloud Bigtable is a fully managed, scalable NoSQL database service for large analytical and operational workloads. Fast and performant: Use Cloud Bigtable as the storage engine that grows with you from your first gigabyte to petabyte-scale for low-latency applications as well as high-throughput data processing and analytics. Seamless scaling and replication: Start with a single node per cluster, and seamlessly scale to hundreds of nodes dynamically supporting peak demand. Replication also adds high availability and workload isolation for live serving apps. Simple and integrated: Fully managed service that integrates easily with big data tools like Hadoop, Dataflow, and Dataproc. Plus, support for the open source HBase API standard makes it easy for development teams to get started. -
27
Dqlite
Canonical
Dqlite is a fast, embedded, persistent SQL database with Raft consensus that is perfect for fault-tolerant IoT and Edge devices. Dqlite (“distributed SQLite”) extends SQLite across a cluster of machines, with automatic failover and high-availability to keep your application running. It uses C-Raft, an optimised Raft implementation in C, to gain high-performance transactional consensus and fault tolerance while preserving SQlite’s outstanding efficiency and tiny footprint. C-Raft is tuned to minimize transaction latency. C-Raft and dqlite are both written in C for maximum cross-platform portability. Published under the LGPLv3 license with a static linking exception for maximum compatibility. Includes common CLI pattern for database initialization and voting member joins and departures. Minimal, tunable delay for failover with automatic leader election. Disk-backed database with in-memory options and SQLite transactions. -
28
Altinity
Altinity
Altinity's expert engineering team can implement everything from core ClickHouse features to Kubernetes operator behavior to client library improvements. A flexible docker-based GUI manager for ClickHouse that can do the following: Install ClickHouse clusters; Add, delete, and replace nodes; Monitor cluster status; Help with troubleshooting and diagnostics. 3rd party tools and software integrations: Ingest: Kafka, ClickTail; APIs: Python, Golang, ODBC, Java; Kubernetes; UI tools: Grafana, Superset, Tabix, Graphite; Databases: MySQL, PostgreSQL; BI tools: Tableau and many more. Altinity.Cloud incorporates lessons from helping hundreds of customers operate ClickHouse-based analytics. Altinity.Cloud has a Kubernetes-based architecture that delivers portability and user choice of where to operate. Designed from the beginning to run anywhere without lock-in. Cost management is critical for SaaS businesses. -
29
Longhorn
Longhorn
In the past, ITOps and DevOps have found it hard to add replicated storage to Kubernetes clusters. As a result many non-cloud-hosted Kubernetes clusters don’t support persistent storage. External storage arrays are non-portable and can be extremely expensive. Longhorn delivers simplified, easy to deploy and upgrade, 100% open source, cloud-native persistent block storage without the cost overhead of open core or proprietary alternatives. Longhorn’s built-in incremental snapshot and backup features keep the volume data safe in or out of the Kubernetes cluster. Scheduled backups of persistent storage volumes in Kubernetes clusters is simplified with Longhorn’s intuitive, free management UI. External replication solutions will recover from a disk failure by re-replicating the entire data store. This can take days, during which time the cluster performs poorly and has a higher risk of failure. -
30
Yugabyte
Yugabyte
The Leading High-Performance Distributed SQL Database. Open source, cloud native relational DB for powering global, internet-scale apps. Single-Digit Millisecond Latency Build blazing fast cloud applications by serving queries directly from the DB. Massive Scale. Achieve millions of transactions per second and store multiple TB’s of data per node. Geo-Distribution. Deploy across regions and clouds with synchronous or multi-master replication. Built for Cloud Native Architectures. Develop, deploy and operationalize modern applications faster than ever before with YugabyteDB. Gain Developer Agility. Leverage full power of PostgreSQL-compatible SQL and distributed ACID transactions. Operate Resilient Services. Ensure continuous availability even when underlying compute, storage or network fails. Scale On-Demand. Add and remove nodes at will. Say no to over-provisioned clusters forever. Lower User Latency. -
31
DRBD
LINBIT
DRBD® (Distributed Replicated Block Device) is an open source, software‑based, shared‑nothing block storage replication solution for Linux, designed primarily to deliver high-performance, high‑availability (HA) data services by mirroring local block devices between nodes in real time, either synchronously or asynchronously. Implemented deep in the Linux kernel as a virtual block‑device driver, DRBD ensures local read performance with efficient write‑through replication to peer(s). User‑space utilities like drbdadm, drbdsetup, and drbdmeta enable declarative configuration, metadata management, and administration across installations. Originally built for two‑node HA clusters, DRBD 9.x extends support to multi‑node replication and integration into software‑defined storage (SDS) systems such as LINSTOR, making it suitable for cloud‑native environments.Starting Price: Free -
32
Apache Accumulo
Apache Corporation
With Apache Accumulo, users can store and manage large data sets across a cluster. Accumulo uses Apache Hadoop's HDFS to store its data and Apache ZooKeeper for consensus. While many users interact directly with Accumulo, several open source projects use Accumulo as their underlying store. To learn more about Accumulo, take the Accumulo tour, read the user manual and run the Accumulo example code. Feel free to contact us if you have any questions. Accumulo has a programming mechanism (called Iterators) that can modify key/value pairs at various points in the data management process. Every Accumulo key/value pair has its own security label which limits query results based off user authorizations. Accumulo runs on a cluster using one or more HDFS instances. Nodes can be added or removed as the amount of data stored in Accumulo changes. -
33
IPFS Cluster
IPFS Cluster
IPFS Cluster provides data orchestration across a swarm of IPFS daemons by allocating, replicating and tracking a global pinset distributed among multiple peers. IPFS has given the users the power of content-addressed storage. The permanent web requires, however, a data redundancy and availability solution that does not compromise on the distributed nature of the IPFS Network. IPFS Cluster is a distributed application that works as a sidecar to IPFS peers, maintaining a global cluster pinset and intelligently allocating its items to the IPFS peers. Cluster peers form a distributed network and maintain a global, replicated and conflict-free list of pins. Ingest IPFS content to multiple daemons directly. Each cluster peer provides an additional IPFS proxy API which performs cluster actions but behaves exactly like the IPFS daemon’s API does. Written in Go, Cluster peers can be programatically launched and controlled. -
34
SIOS DataKeeper
SIOS Technology Corp.
SIOS DataKeeper is a host‑based, block‑level replication solution that delivers real‑time, synchronous or asynchronous redundancy for Windows Server environments, integrating seamlessly with Windows Server Failover Clustering (WSFC). It enables "SANless" clusters—eliminating dependency on shared‑storage arrays—by replicating data across local, virtual, or cloud servers, including VMware, Hyper‑V, AWS, Azure, and Google Cloud Platform, while offering optimized performance without requiring hardware accelerators or compression devices. Once installed, it provides a new SIOS DataKeeper Volume resource in WSFC, supporting geographically dispersed clusters via cross‑subnet failover and configurable heartbeat parameters. Built-in WAN optimization and efficient compression maximize bandwidth use over local and wide‑area networks. -
35
pgEdge
pgEdge
Easily deploy a high availability solution for disaster recovery and failover between and within cloud regions and zero downtime for maintenance. Improve performance and availability with multiple master databases spread across different locations. Keep local data local and control which tables are globally replicated, and which stay local. Support higher throughput when workloads threaten to exceed available compute capacity. For organizations that need or prefer to self-host and self-manage their databases, pgEdge Platform runs on-premises or in self-managed cloud provider accounts. Runs on numerous OS and hardware combinations, and enterprise-class support is available. Self-hosted Edge Platform nodes can also be part of a pgEdge Cloud Postgres cluster. -
36
Eliminate unplanned downtime and minimize data loss due to corruption or failure. The SLE HA extension includes geo clustering to manage clustered servers on-premises or in the cloud anywhere in the world. Our policy-driven, highly available extension for Linux clusters helps you maintain business continuity and minimize unplanned downtime across locations and geographies. Flexible, policy-driven clustering and continuous data replication boost flexibility while improving service availability and resource utilization by supporting the mixed clustering of both physical and virtual Linux servers. Install, configure, manage, and monitor your clustered Linux environments with a powerful unified interface. Multi-tenancy can be used to manage geo clusters according to your business needs.
-
37
NetApp MetroCluster
NetApp
NetApp MetroCluster configurations implement two physically separated, mirrored ONTAP clusters that operate in concert to deliver continuous data and SVM protection. Each cluster synchronously replicates its data aggregates to its partner to maintain identical copies mirrored across both sites. In the event of a site failure, administrators can activate the mirrored SVM on the surviving cluster and resume data serving seamlessly. MetroCluster supports both fabric-attached (FC) and IP-based cluster setups: fabric-attached MetroCluster uses FC transport for SyncMirror between sites, while MetroCluster IP leverages layer‑2 stretched IP networks. Stretch MetroCluster deployments enable campus-wide coverage, MetroCluster IP supports configurations up to four nodes with NVMe/FC or NVMe/TCP starting in ONTAP 9.12.1/9.15.1, and front-end SAN protocols like FC, FCoE, and iSCSI are all supported. -
38
Webair
Webair
Webair provides Database-as-a-Service (DBaaS), a reliable and secure database management solution that gives your business simple, efficient and always available access to its mission-critical data. Our team has proven experience in managing the implementation, configuration, administration and optimization of database clusters, including business-critical, load-balanced and replicated MySQL clusters, including MariaDB, Galera and NoSQL. Webair’s Database Administrators are experts in helping you create a high-performance database environment. We work with you to design the best solution, matching the finest infrastructure to the right database configuration to meet your unique requirements. Free your business from routine database tasks such as performance monitoring, configuration, memory, storage and log file management, sizing, service pack upgrades and service patches. Focus on more important business aspects, including managing the critical data within your database. -
39
Percona XtraDB Cluster
Percona
Percona XtraDB Cluster (PXC) is a high availability, open-source, MySQL clustering solution that helps enterprises minimize unexpected downtime and data loss, reduce costs, and improve the performance and scalability of your database environments. PXC supports your critical business applications in the most demanding public, private, and hybrid cloud environments. Percona XtraDB Cluster (PXC) preserves, secures, and protects data and revenue streams by providing the highest level of availability for your business-critical applications. PXC helps you increase efficiency, eliminate license fees, and lower your total cost of investment, helping you meet budget constraints. Our integrated tools enable you to optimize, maintain, and monitor your cluster. This ensures you get the most out of your MySQL environment.Starting Price: Free -
40
JanusGraph
JanusGraph
JanusGraph is a scalable graph database optimized for storing and querying graphs containing hundreds of billions of vertices and edges distributed across a multi-machine cluster. JanusGraph is a project under The Linux Foundation, and includes participants from Expero, Google, GRAKN.AI, Hortonworks, IBM and Amazon. Elastic and linear scalability for a growing data and user base. Data distribution and replication for performance and fault tolerance. Multi-datacenter high availability and hot backups. All functionality is totally free. No need to buy commercial licenses. JanusGraph is fully open source under the Apache 2 license. JanusGraph is a transactional database that can support thousands of concurrent users executing complex graph traversals in real time. Support for ACID and eventual consistency. In addition to online transactional processing (OLTP), JanusGraph supports global graph analytics (OLAP) with its Apache Spark integration. -
41
Managed Service for PostgreSQL helps you deploy and maintain PostgreSQL server clusters in the Yandex Cloud infrastructure. You can deploy a ready-to-use cluster in just a few minutes. DB settings are optimized for the selected cluster size and you can change them if necessary. If the load on your cluster increases, you can add new servers or increase their capacity in a matter of minutes. With a user-friendly interface and intuitive visualization, you can monitor the status of your PostgreSQL cluster and DB load. All DBMS connections are encrypted using the TLS protocol, and DB backups are GPG-encrypted. Data is secured in accordance with the requirements of local regulatory, GDPR, and ISO industry standards.Starting Price: $40.09 per month
-
42
Apache Cassandra
Apache Software Foundation
The Apache Cassandra database is the right choice when you need scalability and high availability without compromising performance. Linear scalability and proven fault-tolerance on commodity hardware or cloud infrastructure make it the perfect platform for mission-critical data. Cassandra's support for replicating across multiple datacenters is best-in-class, providing lower latency for your users and the peace of mind of knowing that you can survive regional outages. -
43
Tencent Kubernetes Engine
Tencent
TKE is fully compatible with the entire range of Kubernetes capabilities and has been adapted to Tencent Cloud's fundamental IaaS capabilities such as CVM and CBS. In addition, Tencent Cloud’s Kubernetes-based cloud products such as CBS and CLB support one-click deployment to container clusters for a variety of open source applications, greatly improving deployment efficiency. Thanks to TKE, you can simplify the management of large-scale clusters and management and OPS of distributed applications without having to use cluster management software or design fault-tolerant cluster architecture. Simply launch TKE and specify the tasks you want to run, and then TKE will take care of all of the cluster management tasks, allowing you to focus on developing Dockerized applications. -
44
Baidu Messaging System
Baidu AI Cloud
Baidu Messaging System (BMS) is a distributed and scalable hosting message queue service with high throughputs. It collects massive data from websites, devices, or applications for real-time analysis, such as user browsing, clicks, and searches. BMS is a hosted service based on Apache Kafka. Kafka is a distributed, multi-partition, and multi-replica messaging service. The producer asynchronously interacts with the consumer through the message queue without waiting for each other. Compared with the traditional messaging service. BMS encapsulates the Kafka cluster details and provides them in the form of a hosted service. You can directly use BMS to integrate with massively distributed applications without the consideration of cluster operations and pay-per-use only. -
45
Spark Streaming
Apache Software Foundation
Spark Streaming brings Apache Spark's language-integrated API to stream processing, letting you write streaming jobs the same way you write batch jobs. It supports Java, Scala and Python. Spark Streaming recovers both lost work and operator state (e.g. sliding windows) out of the box, without any extra code on your part. By running on Spark, Spark Streaming lets you reuse the same code for batch processing, join streams against historical data, or run ad-hoc queries on stream state. Build powerful interactive applications, not just analytics. Spark Streaming is developed as part of Apache Spark. It thus gets tested and updated with each Spark release. You can run Spark Streaming on Spark's standalone cluster mode or other supported cluster resource managers. It also includes a local run mode for development. In production, Spark Streaming uses ZooKeeper and HDFS for high availability. -
46
Apache Gobblin
Apache Software Foundation
A distributed data integration framework that simplifies common aspects of Big Data integration such as data ingestion, replication, organization, and lifecycle management for both streaming and batch data ecosystems. Runs as a standalone application on a single box. Also supports embedded mode. Runs as an mapreduce application on multiple Hadoop versions. Also supports Azkaban for launching mapreduce jobs. Runs as a standalone cluster with primary and worker nodes. This mode supports high availability and can run on bare metals as well. Runs as an elastic cluster on public cloud. This mode supports high availability. Gobblin as it exists today is a framework that can be used to build different data integration applications like ingest, replication, etc. Each of these applications is typically configured as a separate job and executed through a scheduler like Azkaban. -
47
IBM Analytics Engine provides an architecture for Hadoop clusters that decouples the compute and storage tiers. Instead of a permanent cluster formed of dual-purpose nodes, the Analytics Engine allows users to store data in an object storage layer such as IBM Cloud Object Storage and spins up clusters of computing notes when needed. Separating compute from storage helps to transform the flexibility, scalability and maintainability of big data analytics platforms. Build on an ODPi compliant stack with pioneering data science tools with the broader Apache Hadoop and Apache Spark ecosystem. Define clusters based on your application's requirements. Choose the appropriate software pack, version, and size of the cluster. Use as long as required and delete as soon as an application finishes jobs. Configure clusters with third-party analytics libraries and packages. Deploy workloads from IBM Cloud services like machine learning.Starting Price: $0.014 per hour
-
48
Hadoop
Apache Software Foundation
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage. Rather than rely on hardware to deliver high-availability, the library itself is designed to detect and handle failures at the application layer, so delivering a highly-available service on top of a cluster of computers, each of which may be prone to failures. A wide variety of companies and organizations use Hadoop for both research and production. Users are encouraged to add themselves to the Hadoop PoweredBy wiki page. Apache Hadoop 3.3.4 incorporates a number of significant enhancements over the previous major release line (hadoop-3.2). -
49
Yandex Cloud Functions
Yandex
Run code as a function in a secure, fault-tolerant, and automatically scalable environment without creating or maintaining VMs. As the number of function calls increases, the service automatically creates additional instances of your function. All functions run in parallel. The runtime environment is hosted in three availability zones, ensuring availability even if one zone fails. Configure and prepare instances of functions always ready to process loads. This mode allows you to avoid cold starts and quickly process loads of any size. Give functions access to your VPC to accelerate interactions with private resources, database clusters, virtual machines, Kubernetes nodes, etc. Serverless Functions tracks and logs information about function calls and analyzes execution flow and performance. You can also describe logging mechanisms in your function code. Launch cloud functions in synchronized mode and delayed execution mode.Starting Price: $0.012240 per GB -
50
E-MapReduce
Alibaba
EMR is an all-in-one enterprise-ready big data platform that provides cluster, job, and data management services based on open-source ecosystems, such as Hadoop, Spark, Kafka, Flink, and Storm. Alibaba Cloud Elastic MapReduce (EMR) is a big data processing solution that runs on the Alibaba Cloud platform. EMR is built on Alibaba Cloud ECS instances and is based on open-source Apache Hadoop and Apache Spark. EMR allows you to use the Hadoop and Spark ecosystem components, such as Apache Hive, Apache Kafka, Flink, Druid, and TensorFlow, to analyze and process data. You can use EMR to process data stored on different Alibaba Cloud data storage service, such as Object Storage Service (OSS), Log Service (SLS), and Relational Database Service (RDS). You can quickly create clusters without the need to configure hardware and software. All maintenance operations are completed on its Web interface.