Alternatives to IBM Event Streams

Compare IBM Event Streams alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to IBM Event Streams in 2026. Compare features, ratings, user reviews, pricing, and more from IBM Event Streams competitors and alternatives in order to make an informed decision for your business.

  • 1
    HiveMQ

    HiveMQ

    HiveMQ

    HiveMQ is the Industrial AI Platform helping enterprises move from connected devices to intelligent operations. Built on the MQTT standard and a distributed edge-to-cloud architecture, HiveMQ connects and governs industrial data in real time, enabling organizations to act with intelligence. With proven reliability, scalability, and interoperability, HiveMQ provides the foundation industrial companies need to operationalize AI, powering the next generation of intelligent industry. Global leaders including Audi, BMW, Eli Lilly, Liberty Global, Mercedes-Benz, and Siemens trust HiveMQ to run their most mission-critical operations.
    Partner badge
    Compare vs. IBM Event Streams View Software
    Visit Website
  • 2
    Striim

    Striim

    Striim

    Data integration for your hybrid cloud. Modern, reliable data integration across your private and public cloud. All in real-time with change data capture and data streams. Built by the executive & technical team from GoldenGate Software, Striim brings decades of experience in mission-critical enterprise workloads. Striim scales out as a distributed platform in your environment or in the cloud. Scalability is fully configurable by your team. Striim is fully secure with HIPAA and GDPR compliance. Built ground up for modern enterprise workloads in the cloud or on-premise. Drag and drop to create data flows between your sources and targets. Process, enrich, and analyze your streaming data with real-time SQL queries.
  • 3
    Amazon EventBridge
    Amazon EventBridge is a serverless event bus that makes it easy to connect applications together using data from your own applications, integrated Software-as-a-Service (SaaS) applications, and AWS services. EventBridge delivers a stream of real-time data from event sources, such as Zendesk, Datadog, or Pagerduty, and routes that data to targets like AWS Lambda. You can set up routing rules to determine where to send your data to build application architectures that react in real time to all of your data sources. EventBridge makes it easy to build event-driven applications because it takes care of event ingestion and delivery, security, authorization, and error handling for you. As your applications become more interconnected through events, you need to spend more effort to find events and understand their structure in order to write code to react to those events.
  • 4
    EMQX

    EMQX

    EMQ Technologies

    EMQX is the world's most scalable and reliable MQTT messaging platform designed by EMQ. It supports 100M concurrent IoT device connections per cluster while maintaining extremely high throughput and sub-millisecond latency. EMQX boasts more than 20,000 global users from over 50 countries, connecting more than 100M IoT devices worldwide, and is trusted by over 300 customers in mission-critical IoT scenarios, including well-known brands like HPE, VMware, Verifone, SAIC Volkswagen, and Ericsson. Our edge-to-cloud IoT connectivity solutions are flexible to meet the demands of various industries towards digital transformation, including connected vehicles, Industrial IoT, oil & gas, carrier, finance, smart energy, and smart cities.
  • 5
    Apache Kafka

    Apache Kafka

    The Apache Software Foundation

    Apache Kafka® is an open-source, distributed streaming platform. Scale production clusters up to a thousand brokers, trillions of messages per day, petabytes of data, hundreds of thousands of partitions. Elastically expand and contract storage and processing. Stretch clusters efficiently over availability zones or connect separate clusters across geographic regions. Process streams of events with joins, aggregations, filters, transformations, and more, using event-time and exactly-once processing. Kafka’s out-of-the-box Connect interface integrates with hundreds of event sources and event sinks including Postgres, JMS, Elasticsearch, AWS S3, and more. Read, write, and process streams of events in a vast array of programming languages.
  • 6
    PubSub+ Platform
    Solace PubSub+ Platform helps enterprises design, deploy and manage event-driven systems across hybrid and multi-cloud and IoT environments so they can be more event-driven and operate in real-time. The PubSub+ Platform includes the powerful PubSub+ Event Brokers, event management capabilities with PubSub+ Event Portal, as well as monitoring and integration capabilities all available via a single cloud console. PubSub+ allows easy creation of an event mesh, an interconnected network of event brokers, allowing for seamless and dynamic data movement across highly distributed network environments. PubSub+ Event Brokers can be deployed as fully managed cloud services, self-managed software in private cloud or on-premises environments, or as turnkey hardware appliances for unparalleled performance and low TCO. PubSub+ Event Portal is a complimentary toolset for design and governance of event-driven systems including both Solace and Kafka-based event broker environments.
  • 7
    Azure Event Hubs
    Event Hubs is a fully managed, real-time data ingestion service that’s simple, trusted, and scalable. Stream millions of events per second from any source to build dynamic data pipelines and immediately respond to business challenges. Keep processing data during emergencies using the geo-disaster recovery and geo-replication features. Integrate seamlessly with other Azure services to unlock valuable insights. Allow existing Apache Kafka clients and applications to talk to Event Hubs without any code changes—you get a managed Kafka experience without having to manage your own clusters. Experience real-time data ingestion and microbatching on the same stream. Focus on drawing insights from your data instead of managing infrastructure. Build real-time big data pipelines and respond to business challenges right away.
    Starting Price: $0.03 per hour
  • 8
    Astra Streaming
    Responsive applications keep users engaged and developers inspired. Rise to meet these ever-increasing expectations with the DataStax Astra Streaming service platform. DataStax Astra Streaming is a cloud-native messaging and event streaming platform powered by Apache Pulsar. Astra Streaming allows you to build streaming applications on top of an elastically scalable, multi-cloud messaging and event streaming platform. Astra Streaming is powered by Apache Pulsar, the next-generation event streaming platform which provides a unified solution for streaming, queuing, pub/sub, and stream processing. Astra Streaming is a natural complement to Astra DB. Using Astra Streaming, existing Astra DB users can easily build real-time data pipelines into and out of their Astra DB instances. With Astra Streaming, avoid vendor lock-in and deploy on any of the major public clouds (AWS, GCP, Azure) compatible with open-source Apache Pulsar.
  • 9
    Confluent

    Confluent

    Confluent

    Infinite retention for Apache Kafka® with Confluent. Be infrastructure-enabled, not infrastructure-restricted Legacy technologies require you to choose between being real-time or highly-scalable. Event streaming enables you to innovate and win - by being both real-time and highly-scalable. Ever wonder how your rideshare app analyzes massive amounts of data from multiple sources to calculate real-time ETA? Ever wonder how your credit card company analyzes millions of credit card transactions across the globe and sends fraud notifications in real-time? The answer is event streaming. Move to microservices. Enable your hybrid strategy through a persistent bridge to cloud. Break down silos to demonstrate compliance. Gain real-time, persistent event transport. The list is endless.
  • 10
    Aiven for Apache Kafka
    Apache Kafka as a fully managed service, with zero vendor lock-in and a full set of capabilities to build your streaming pipeline. Set up fully managed Kafka in less than 10 minutes — directly from our web console or programmatically via our API, CLI, Terraform provider or Kubernetes operator. Easily connect it to your existing tech stack with over 30 connectors, and feel confident in your setup with logs and metrics available out of the box via the service integrations. A fully managed distributed data streaming platform, deployable in the cloud of your choice. Ideal for event-driven applications, near-real-time data transfer and pipelines, stream analytics, and any other case where you need to move a lot of data between applications — and quickly. With Aiven’s hosted and managed-for-you Apache Kafka, you can set up clusters, deploy new nodes, migrate clouds, and upgrade existing versions — in a single mouse click — and monitor them through a simple dashboard.
    Starting Price: $200 per month
  • 11
    Amazon MSK
    Amazon Managed Streaming for Apache Kafka (Amazon MSK) is a fully managed service that makes it easy for you to build and run applications that use Apache Kafka to process streaming data. Apache Kafka is an open-source platform for building real-time streaming data pipelines and applications. With Amazon MSK, you can use native Apache Kafka APIs to populate data lakes, stream changes to and from databases, and power machine learning and analytics applications. Apache Kafka clusters are challenging to setup, scale, and manage in production. When you run Apache Kafka on your own, you need to provision servers, configure Apache Kafka manually, replace servers when they fail, orchestrate server patches and upgrades, architect the cluster for high availability, ensure data is durably stored and secured, setup monitoring and alarms, and carefully plan scaling events to support load changes.
    Starting Price: $0.0543 per hour
  • 12
    Axual

    Axual

    Axual

    Axual is Kafka-as-a-Service for DevOps teams. Empower your team to unlock insights and drive decisions with our intuitive Kafka platform. Axual offers the ultimate solution for enterprises looking to seamlessly integrate data streaming into their core IT infrastructure. Our all-in-one Kafka platform is designed to eliminate the need for extensive technical knowledge or skills, and provides a ready-made solution that delivers all the benefits of event streaming without the hassle. The Axual Platform is a all-in-one solution, designed to help you simplify and enhance the deployment, management, and utilization of real-time data streaming with Apache Kafka. By providing an array of features that cater to the diverse needs of modern enterprises, the Axual Platform enables organizations to harness the full potential of data streaming while minimizing complexity and operational overhead.
  • 13
    Oracle Cloud Infrastructure Streaming
    Streaming service is a real-time, serverless, Apache Kafka-compatible event streaming platform for developers and data scientists. Streaming is tightly integrated with Oracle Cloud Infrastructure (OCI), Database, GoldenGate, and Integration Cloud. The service also provides out-of-the-box integrations for hundreds of third-party products across categories such as DevOps, databases, big data, and SaaS applications. Data engineers can easily set up and operate big data pipelines. Oracle handles all infrastructure and platform management for event streaming, including provisioning, scaling, and security patching. With the help of consumer groups, Streaming can provide state management for thousands of consumers. This helps developers easily build applications at scale.
  • 14
    StreamNative

    StreamNative

    StreamNative

    StreamNative redefines streaming infrastructure by seamlessly integrating Kafka, MQ, and other protocols into a single, unified platform, providing unparalleled flexibility and efficiency for modern data processing needs. StreamNative offers a unified solution that adapts to the diverse requirements of streaming and messaging in a microservices-driven environment. By providing a comprehensive and intelligent approach to messaging and streaming, StreamNative empowers organizations to navigate the complexities and scalability of the modern data ecosystem with efficiency and agility. Apache Pulsar’s unique architecture decouples the message serving layer from the message storage layer to deliver a mature cloud-native data-streaming platform. Scalable and elastic to adapt to rapidly changing event traffic and business needs. Scale-up to millions of topics with architecture that decouples computing and storage.
    Starting Price: $1,000 per month
  • 15
    WarpStream

    WarpStream

    WarpStream

    WarpStream is an Apache Kafka-compatible data streaming platform built directly on top of object storage, with no inter-AZ networking costs, no disks to manage, and infinitely scalable, all within your VPC. WarpStream is deployed as a stateless and auto-scaling agent binary in your VPC with no local disks to manage. Agents stream data directly to and from object storage with no buffering on local disks and no data tiering. Create new “virtual clusters” in our control plane instantly. Support different environments, teams, or projects without managing any dedicated infrastructure. WarpStream is protocol compatible with Apache Kafka, so you can keep using all your favorite tools and software. No need to rewrite your application or use a proprietary SDK. Just change the URL in your favorite Kafka client library and start streaming. Never again have to choose between reliability and your budget.
    Starting Price: $2,987 per month
  • 16
    Informatica Data Engineering Streaming
    AI-powered Informatica Data Engineering Streaming enables data engineers to ingest, process, and analyze real-time streaming data for actionable insights. Advanced serverless deployment option​ with integrated metering dashboard cuts admin overhead. Rapidly build intelligent data pipelines with CLAIRE®-powered automation, including automatic change data capture (CDC). Ingest thousands of databases and millions of files, and streaming events. Efficiently ingest databases, files, and streaming data for real-time data replication and streaming analytics. Find and inventory all data assets throughout your organization. Intelligently discover and prepare trusted data for advanced analytics and AI/ML projects.
  • 17
    Red Hat OpenShift Streams
    Red Hat® OpenShift® Streams for Apache Kafka is a managed cloud service that provides a streamlined developer experience for building, deploying, and scaling new cloud-native applications or modernizing existing systems. Red Hat OpenShift Streams for Apache Kafka makes it easy to create, discover, and connect to real-time data streams no matter where they are deployed. Streams are a key component for delivering event-driven and data analytics applications. The combination of seamless operations across distributed microservices, large data transfer volumes, and managed operations allows teams to focus on team strengths, speed up time to value, and lower operational costs. OpenShift Streams for Apache Kafka includes a Kafka ecosystem and is part of a family of cloud services—and the Red Hat OpenShift product family—which helps you build a wide range of data-driven solutions.
  • 18
    DeltaStream

    DeltaStream

    DeltaStream

    DeltaStream is a unified serverless stream processing platform that integrates with streaming storage services. Think about it as the compute layer on top of your streaming storage. It provides functionalities of streaming analytics(Stream processing) and streaming databases along with additional features to provide a complete platform to manage, process, secure and share streaming data. DeltaStream provides a SQL based interface where you can easily create stream processing applications such as streaming pipelines, materialized views, microservices and many more. It has a pluggable processing engine and currently uses Apache Flink as its primary stream processing engine. DeltaStream is more than just a query processing layer on top of Kafka or Kinesis. It brings relational database concepts to the data streaming world, including namespacing and role based access control enabling you to securely access, process and share your streaming data regardless of where they are stored.
  • 19
    Aiven

    Aiven

    Aiven

    Aiven manages your open source data infrastructure in the cloud - so you don't have to. Developers can do what they do best: create applications. We do what we do best: manage cloud data infrastructure. All solutions are open source. You can also freely move data between clouds or create multi-cloud environments. Know exactly how much you’ll be paying and why. We bundle networking, storage and basic support costs together. We are committed to keeping your Aiven software online. If there’s ever an issue, we’ll be there to fix it. Deploy a service on the Aiven platform in 10 minutes. Sign up - no credit card info needed. Select your open source service, and the cloud and region to deploy to. Choose your plan - you have $300 in free credits. Click "Create service" and go on to configure your data sources. Stay in control of your data using powerful open-source services.
    Starting Price: $200.00 per month
  • 20
    TIBCO Platform

    TIBCO Platform

    Cloud Software Group

    TIBCO delivers industrial-strength solutions that meet your performance, throughput, reliability, and scalability needs while offering a wide range of technology and deployment options to deliver real-time data where it’s needed most. The TIBCO Platform will bring together an evolving set of your TIBCO solutions wherever they are hosted—in the cloud, on-premises, and at the edge—into a single, unified experience so that you can more easily manage and monitor them. TIBCO helps build solutions that are essential to the success of the world’s largest enterprises.
  • 21
    Ably

    Ably

    Ably

    Ably is the definitive realtime experience platform. We power more WebSocket connections than any other pub/sub platform, serving over a billion devices monthly. Businesses like HubSpot, NASCAR and Webflow trust us to power their critical applications - reliably, securely and at serious scale. Ably’s products place composable realtime in the hands of developers. Simple APIs and SDKs for every tech stack, enable the creation of a host of live experiences - including chat, collaboration, notifications, broadcast and fan engagement. All powered by our scalable infrastructure.
    Starting Price: $49.99/month
  • 22
    Arroyo

    Arroyo

    Arroyo

    Scale from zero to millions of events per second. Arroyo ships as a single, compact binary. Run locally on MacOS or Linux for development, and deploy to production with Docker or Kubernetes. Arroyo is a new kind of stream processing engine, built from the ground up to make real-time easier than batch. Arroyo was designed from the start so that anyone with SQL experience can build reliable, efficient, and correct streaming pipelines. Data scientists and engineers can build end-to-end real-time applications, models, and dashboards, without a separate team of streaming experts. Transform, filter, aggregate, and join data streams by writing SQL, with sub-second results. Your streaming pipelines shouldn't page someone just because Kubernetes decided to reschedule your pods. Arroyo is built to run in modern, elastic cloud environments, from simple container runtimes like Fargate to large, distributed deployments on the Kubernetes logo Kubernetes.
  • 23
    Amazon Kinesis
    Easily collect, process, and analyze video and data streams in real time. Amazon Kinesis makes it easy to collect, process, and analyze real-time, streaming data so you can get timely insights and react quickly to new information. Amazon Kinesis offers key capabilities to cost-effectively process streaming data at any scale, along with the flexibility to choose the tools that best suit the requirements of your application. With Amazon Kinesis, you can ingest real-time data such as video, audio, application logs, website clickstreams, and IoT telemetry data for machine learning, analytics, and other applications. Amazon Kinesis enables you to process and analyze data as it arrives and respond instantly instead of having to wait until all your data is collected before the processing can begin. Amazon Kinesis enables you to ingest, buffer, and process streaming data in real-time, so you can derive insights in seconds or minutes instead of hours or days.
  • 24
    Nussknacker

    Nussknacker

    Nussknacker

    Nussknacker is a low-code visual tool for domain experts to define and run real-time decisioning algorithms instead of implementing them in the code. It serves where real-time actions on data have to be made: real-time marketing, fraud detection, Internet of Things, Customer 360, and Machine Learning inferring. An essential part of Nussknacker is a visual design tool for decision algorithms. It allows not-so-technical users – analysts or business people – to define decision logic in an imperative, easy-to-follow, and understandable way. Once authored, with a click of a button, scenarios are deployed for execution. And can be changed and redeployed anytime there’s a need. Nussknacker supports two processing modes: streaming and request-response. In streaming mode, it uses Kafka as its primary interface. It supports both stateful and stateless processing.
  • 25
    Spark Streaming

    Spark Streaming

    Apache Software Foundation

    Spark Streaming brings Apache Spark's language-integrated API to stream processing, letting you write streaming jobs the same way you write batch jobs. It supports Java, Scala and Python. Spark Streaming recovers both lost work and operator state (e.g. sliding windows) out of the box, without any extra code on your part. By running on Spark, Spark Streaming lets you reuse the same code for batch processing, join streams against historical data, or run ad-hoc queries on stream state. Build powerful interactive applications, not just analytics. Spark Streaming is developed as part of Apache Spark. It thus gets tested and updated with each Spark release. You can run Spark Streaming on Spark's standalone cluster mode or other supported cluster resource managers. It also includes a local run mode for development. In production, Spark Streaming uses ZooKeeper and HDFS for high availability.
  • 26
    Google Cloud Dataflow
    Unified stream and batch data processing that's serverless, fast, and cost-effective. Fully managed data processing service. Automated provisioning and management of processing resources. Horizontal autoscaling of worker resources to maximize resource utilization. OSS community-driven innovation with Apache Beam SDK. Reliable and consistent exactly-once processing. Streaming data analytics with speed. Dataflow enables fast, simplified streaming data pipeline development with lower data latency. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Allow teams to focus on programming instead of managing server clusters as Dataflow’s serverless approach removes operational overhead from data engineering workloads. Dataflow automates provisioning and management of processing resources to minimize latency and maximize utilization.
  • 27
    Lightstreamer

    Lightstreamer

    Lightstreamer

    ​Lightstreamer is an event broker optimized for the internet, ensuring seamless real-time data delivery across the web. Unlike traditional brokers, Lightstreamer automatically handles proxies, firewalls, disconnections, network congestion, and the general unpredictability of the internet. With its intelligent streaming feature, Lightstreamer guarantees real-time data transmission, always finding a way to deliver your data reliably and efficiently, ensuring robust last-mile messaging. Lightstreamer offers technology that is both mature and cutting-edge, continuously evolving to stay at the forefront of innovation. With a proven track record and years of field-tested performance, Lightstreamer ensures your data is delivered reliably and efficiently. Experience unparalleled reliability in any scenario with Lightstreamer.
  • 28
    Leo

    Leo

    Leo

    Turn your data into a realtime stream, making it immediately available and ready to use. Leo reduces the complexity of event sourcing by making it easy to create, visualize, monitor, and maintain your data flows. Once you unlock your data, you are no longer limited by the constraints of your legacy systems. Dramatically reduced dev time keeps your developers and stakeholders happy. Adopt microservice architectures to continuously innovate and improve agility. In reality, success with microservices is all about data. An organization must invest in a reliable and repeatable data backbone to make microservices a reality. Implement full-fledged search in your custom app. With data flowing, adding and maintaining a search database will not be a burden.
    Starting Price: $251 per month
  • 29
    IBM Event Automation
    ​IBM Event Automation is a fully composable event-driven solution designed to enable users to detect situations, act in real time, automate decisions, and maximize revenue potential. It allows businesses to respond in real time using Apache Flink, leveraging AI to anticipate critical business patterns. It facilitates the development of scalable applications to meet evolving business needs and handle increasing workloads seamlessly. It enables self-service access with approval controls, field redaction, and schema filtering, enforced by a Kafka-native event gateway via policy administration. IBM Event Automation unifies and accelerates event management by using policy administration for self-service access, enabling control definitions for approval processes, field-level redaction, and schema-based filtering. Use cases include transaction data analysis, inventory optimization, detecting suspicious activity, enhancing customer understanding, predictive maintenance, etc.
  • 30
    Macrometa

    Macrometa

    Macrometa

    We deliver a geo-distributed real-time database, stream processing and compute runtime for event-driven applications across up to 175 worldwide edge data centers. App & API builders love our platform because we solve the hardest problems of sharing mutable state across 100s of global locations, with strong consistency & low latency. Macrometa enables you to surgically extend your existing infrastructure to bring part of or your entire application closer to your end users. This allows you to improve performance, user experience, and comply with global data governance laws. Macrometa is a serverless, streaming NoSQL database, with integrated pub/sub and stream data processing and compute engine. Create stateful data infrastructure, stateful functions & containers for long running workloads, and process data streams in real time. You do the code, we do all the ops and orchestration.
  • 31
    Amazon Managed Service for Apache Flink
    Thousands of customers use Amazon Managed Service for Apache Flink to run stream processing applications. With Amazon Managed Service for Apache Flink, you can transform and analyze streaming data in real-time using Apache Flink and integrate applications with other AWS services. There are no servers and clusters to manage, and there is no computing and storage infrastructure to set up. You pay only for the resources you use. Build and run Apache Flink applications, without setting up infrastructure and managing resources and clusters. Process gigabytes of data per second with subsecond latencies and respond to events in real-time. Deploy highly available and durable applications with Multi-AZ deployments and APIs for application lifecycle management. Develop applications that transform and deliver data to Amazon Simple Storage Service (Amazon S3), Amazon OpenSearch Service, and more.
    Starting Price: $0.11 per hour
  • 32
    Alibaba Cloud EventBridge
    EventBridge is a serverless event bus service that connects to Alibaba Cloud services, custom applications, and SaaS applications as a centralized hub. EventBridge can also use the CloudEvents 1.0 specification to route events among these services and applications. EventBridge helps you build loosely coupled and distributed event-driven architectures. Provides comprehensive event rule management, including creating, updating, and querying event rules, and enabling or disabling these rules. Supports an ever-growing range of events from Alibaba Cloud services. Region-specific, cross-zone distributed cluster deployment provides powerful disaster recovery capabilities and delivers up to 99.95% service availability. Provides event governance capabilities and supports event flow control, event replay, and event retry policies.
  • 33
    Apache Beam

    Apache Beam

    Apache Software Foundation

    The easiest way to do batch and streaming data processing. Write once, run anywhere data processing for mission-critical production workloads. Beam reads your data from a diverse set of supported sources, no matter if it’s on-prem or in the cloud. Beam executes your business logic for both batch and streaming use cases. Beam writes the results of your data processing logic to the most popular data sinks in the industry. A simplified, single programming model for both batch and streaming use cases for every member of your data and application teams. Apache Beam is extensible, with projects such as TensorFlow Extended and Apache Hop built on top of Apache Beam. Execute pipelines on multiple execution environments (runners), providing flexibility and avoiding lock-in. Open, community-based development and support to help evolve your application and meet the needs of your specific use cases.
  • 34
    Eventarc

    Eventarc

    Google

    ​Google Cloud's Eventarc is a fully managed platform that enables developers to build event-driven architectures by routing events from various sources to supported destinations. It allows for the collection of events occurring within a system and publishes them to a specified destination, facilitating the creation of loosely coupled services that react to state changes. ​Eventarc supports events from Google Cloud services, custom applications, and third-party SaaS providers, providing flexibility in event-driven application design. Developers can create triggers to route events to various destinations, such as Cloud Run services, allowing for responsive and scalable application architectures. Eventarc ensures secure event delivery by integrating with Identity and Access Management (IAM), enabling fine-grained access control over event ingestion and processing.
  • 35
    IBM MQ on Cloud
    IBM® MQ on Cloud is the gold standard for enterprise messaging, providing security-rich and reliable messaging on-premises and across multiple clouds. Use IBM MQ on Cloud as a managed offering. IBM will handle upgrades, patches and many of the operational management tasks, allowing you to focus on integrations with your applications. Your company uses a mobile app on the cloud to facilitate e-commerce transactions. IBM MQ on Cloud connects the on-premises stock system with the consumer application to give users real-time information about what products are available. Your company hosts its core IT systems in San Francisco, but packages are processed in a depot in London. IBM MQ on Cloud reliably transmits messages from one location to another. It lets the London office encrypt "send" data about every package that needs to be tracked, and lets the San Francisco office receive and process that information more securely. Both offices can trust that information won’t be lost.
  • 36
    Amazon MQ
    Amazon MQ is a managed message broker service for Apache ActiveMQ that makes it easy to set up and operate message brokers in the cloud. Message brokers allow different software systems–often using different programming languages, and on different platforms–to communicate and exchange information. Amazon MQ reduces your operational load by managing the provisioning, setup, and maintenance of ActiveMQ, a popular open-source message broker. Connecting your current applications to Amazon MQ is easy because it uses industry-standard APIs and protocols for messaging, including JMS, NMS, AMQP, STOMP, MQTT, and WebSocket. Using standards means that in most cases, there’s no need to rewrite any messaging code when you migrate to AWS. With a few clicks in the Amazon MQ Console, Amazon MQ provisions your broker with support for version upgrades, so you can always use the latest version that Amazon MQ supports. Once you configure your broker, your applications can produce and consume messages.
  • 37
    Apache Flume

    Apache Flume

    Apache Software Foundation

    Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. It has a simple and flexible architecture based on streaming data flows. It is robust and fault-tolerant with tunable reliability mechanisms and many failovers and recovery mechanisms. It uses a simple extensible data model that allows for online analytic applications. The Apache Flume team is pleased to announce the release of Flume 1.8.0. Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of streaming event data.
  • 38
    Apache Flink

    Apache Flink

    Apache Software Foundation

    Apache Flink is a framework and distributed processing engine for stateful computations over unbounded and bounded data streams. Flink has been designed to run in all common cluster environments, perform computations at in-memory speed and at any scale. Any kind of data is produced as a stream of events. Credit card transactions, sensor measurements, machine logs, or user interactions on a website or mobile application, all of these data are generated as a stream. Apache Flink excels at processing unbounded and bounded data sets. Precise control of time and state enable Flink’s runtime to run any kind of application on unbounded streams. Bounded streams are internally processed by algorithms and data structures that are specifically designed for fixed sized data sets, yielding excellent performance. Flink is designed to work well each of the previously listed resource managers.
  • 39
    Cloudera DataFlow
    Cloudera DataFlow for the Public Cloud (CDF-PC) is a cloud-native universal data distribution service powered by Apache NiFi ​​that lets developers connect to any data source anywhere with any structure, process it, and deliver to any destination. CDF-PC offers a flow-based low-code development paradigm that aligns best with how developers design, develop, and test data distribution pipelines. With over 400+ connectors and processors across the ecosystem of hybrid cloud services—including data lakes, lakehouses, cloud warehouses, and on-premises sources—CDF-PC provides indiscriminate data distribution. These data distribution flows can then be version-controlled into a catalog where operators can self-serve deployments to different runtimes.
  • 40
    Samza

    Samza

    Apache Software Foundation

    Samza allows you to build stateful applications that process data in real-time from multiple sources including Apache Kafka. Battle-tested at scale, it supports flexible deployment options to run on YARN or as a standalone library. Samza provides extremely low latencies and high throughput to analyze your data instantly. Scales to several terabytes of state with features like incremental checkpoints and host-affinity. Samza is easy to operate with flexible deployment options - YARN, Kubernetes or standalone. Ability to run the same code to process both batch and streaming data. Integrates with several sources including Kafka, HDFS, AWS Kinesis, Azure Eventhubs, K-V stores and ElasticSearch.
  • 41
    Azure Stream Analytics
    Discover Azure Stream Analytics, the easy-to-use, real-time analytics service that is designed for mission-critical workloads. Build an end-to-end serverless streaming pipeline with just a few clicks. Go from zero to production in minutes using SQL—easily extensible with custom code and built-in machine learning capabilities for more advanced scenarios. Run your most demanding workloads with the confidence of a financially backed SLA.
  • 42
    Apache Pulsar

    Apache Pulsar

    Apache Software Foundation

    Apache Pulsar is a cloud-native, distributed messaging and streaming platform originally created at Yahoo! and now a top-level Apache Software Foundation project. Easy to deploy, lightweight compute process, developer-friendly APIs, no need to run your own stream processing engine. Run in production at Yahoo! scale for over 5 years, with millions of messages per second across millions of topics. Built from the ground up as a multi-tenant system. Supports isolation, authentication, authorization and quotas. Configurable replication between data centers across multiple geographic regions. Persistent message storage based on Apache BookKeeper. IO-level isolation between write and read operations. Rest admin API for provisioning, administration, tools and monitoring.
  • 43
    Google Cloud Managed Service for Kafka
    ​Google Cloud's Managed Service for Apache Kafka is a fully managed and scalable service that simplifies the deployment, management, and maintenance of Apache Kafka clusters. It automates operational tasks such as provisioning, patching, and scaling, allowing users to focus on building applications without the complexities of infrastructure management. It ensures high availability and reliability by replicating data across multiple zones, safeguarding against potential failures. It also offers seamless integration with other Google Cloud services, enabling users to create robust data processing pipelines. Security is a priority, with features like encryption at rest and in transit, identity, and access management, and network isolation to protect data. Google Cloud Managed Service for Kafka supports both public and private networking configurations, providing flexibility in connectivity options.
    Starting Price: $0.09 per hour
  • 44
    Baidu AI Cloud Stream Computing
    Baidu Stream Computing (BSC) provides real-time streaming data processing capacity with low delay, high throughput and high accuracy. It is fully compatible with Spark SQL; and can realize the logic data processing of complicated businesses through SQL statement, which is easy to use; provides users with full life cycle management for the streaming-oriented computing jobs. Integrate deeply with multiple storage products of Baidu AI Cloud as the upstream and downstream of stream computing, including Baidu Kafka, RDS, BOS, IOT Hub, Baidu ElasticSearch, TSDB, SCS and others. Provide a comprehensive job monitoring indicator, and the user can view the monitoring indicators of the job and set the alarm rules to protect the job.
  • 45
    IBM MQ
    Massive amounts of data move as messages between applications, systems and services at any given time. If an application isn’t ready or if there’s a service interruption, messages and transactions can be lost or duplicated, costing businesses time and money to make things right. IBM has expertly refined IBM MQ over 25 years on the market. With MQ, if a message can’t be delivered immediately, it’s secured in a queue, where it waits until delivery is assured. Where competitors may deliver messages twice or not at all, MQ moves data, including file data, once — and once only. Never lose a message with MQ. IBM MQ is available as software to run in public or private clouds, in containers or on your mainframe. IBM also offers an IBM-managed cloud service (IBM MQ on Cloud) hosted on IBM Cloud or Amazon, and even as a purpose-built Appliance (IBM MQ Appliance) to simplify deployment and maintenance.
  • 46
    RabbitMQ

    RabbitMQ

    RabbitMQ

    RabbitMQ is lightweight and easy to deploy on-premises and in the cloud. It supports multiple messaging protocols. RabbitMQ can be deployed in distributed and federated configurations to meet high-scale, high-availability requirements. With tens of thousands of users, RabbitMQ is one of the most popular open-source message brokers. From T-Mobile to Runtastic, RabbitMQ is used worldwide at small startups and large enterprises. RabbitMQ is lightweight and easy to deploy on-premises and in the cloud. It supports multiple messaging protocols. RabbitMQ can be deployed in distributed and federated configurations to meet high-scale, high-availability requirements. RabbitMQ runs on many operating systems and cloud environments and provides a wide range of developer tools for most popular languages. Deploy with Kubernetes, BOSH, Chef, Docker and Puppet. Develop cross-language messaging with favorite programming languages such as Java, .NET, PHP, Python, JavaScript, Ruby, Go, etc.
  • 47
    Anypoint MQ

    Anypoint MQ

    MuleSoft

    With Anypoint MQ, perform advanced asynchronous messaging — such as queueing and pub/sub — with fully hosted and managed cloud message queues and exchanges. As a service of Anypoint Platform™, Anypoint MQ supports environments, business groups, and role-based access control (RBAC) with enterprise-grade functionality.
  • 48
    IBM Cloud Messages for RabbitMQ
    IBM® Messages for RabbitMQ on IBM Cloud® supports multiple messaging protocols as a broker. It lets you route, track and queue messages with customizable persistence levels, delivery settings and publish confirmations. Get to global scale with integrated, infrastructure-as-code tools, such as IBM Cloud Schematics with Terraform and Red Hat® Ansible® support at no additional charge. IBM® Key Protect lets you can bring your own encryption key. Each deployment supports private networking, in-database auditing and more. Messages for RabbitMQ allows you to scale disk and RAM independently to fit your requirements. Grow with elasticity just an API call away. The service is compatible with RabbitMQ APIs, data formats and clients. You can use Messages for RabbitMQ as a drop-in replacement for RabbitMQ. The standard configuration includes three data members configured for high availability. Deployments use multiple availability zones.
  • 49
    PubNub

    PubNub

    PubNub

    Innovate with Realtime Features: We take care of realtime communication infrastructure so you can focus on your app. Our Platform for Realtime Communication: A platform to build and operate real-time interactivity for web, mobile, AI/ML, IoT, and Edge computing applications Faster & Easier Deployments: SDK support for 50+ mobile, web, server, and IoT environments (PubNub and community supported) and more than 65 pre-built integrations with external and third-party APIs to give developers the features they need regardless of programming language or tech stack. Scalability: The industry’s most scalable platform capable of supporting millions of concurrent users and allows for rapid growth with low latency, high uptime, and without financial penalties. Security & Compliance: Enterprise-grade security and compliance with the most stringent regulations worldwide, including GDPR, SOC 2, HIPAA, ISO 27001, and CCPA.
  • 50
    Apache Storm

    Apache Storm

    Apache Software Foundation

    Apache Storm is a free and open source distributed realtime computation system. Apache Storm makes it easy to reliably process unbounded streams of data, doing for realtime processing what Hadoop did for batch processing. Apache Storm is simple, can be used with any programming language, and is a lot of fun to use! Apache Storm has many use cases: realtime analytics, online machine learning, continuous computation, distributed RPC, ETL, and more. Apache Storm is fast: a benchmark clocked it at over a million tuples processed per second per node. It is scalable, fault-tolerant, guarantees your data will be processed, and is easy to set up and operate. Apache Storm integrates with the queueing and database technologies you already use. An Apache Storm topology consumes streams of data and processes those streams in arbitrarily complex ways, repartitioning the streams between each stage of the computation however needed. Read more in the tutorial.