Showing 681 open source projects for "latency"

View related business solutions
  • Powerful App Monitoring Without Surprise Bills Icon
    Powerful App Monitoring Without Surprise Bills

    AppSignal starts at $23/month with all features included. No overages, no hidden fees. 30-day free trial.

    Tired of monitoring tools that punish you for scaling? AppSignal offers transparent, predictable pricing with every feature unlocked on every plan. Track errors, monitor performance, detect anomalies, and manage logs across Ruby, Python, Node.js, and more. Trusted by developers since 2012 with free dev-to-dev support. No credit card required to start your 30-day trial.
    Try AppSignal Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Sunshine

    Sunshine

    Self-hosted game stream host for Moonlight

    Sunshine is an open-source self‑hosted cloud gaming server that implements NVIDIA’s GameStream protocol. Compatible with Moonlight clients across platforms, it supports low‑latency streaming via software or hardware encoding (AMD/Intel/NVIDIA) and offers a browser‑based control UI for pairing.
    Downloads: 1,231 This Week
    Last Update:
    See Project
  • 2
    Ant Media Server

    Ant Media Server

    Adaptive, ultra low latency streaming

    Real-time streaming engine delivers content with sub-0.5secs latency. Ant Media Server supports WebRTC, CMAF, HLS, RTMP, RTSP, SRT, Zixi and more for your needs. Ant Media Server is a real-time streaming engine software that provides adaptive, ultra low latency streaming by using WebRTC technology with ~0.5 seconds latency. Ant Media Server is auto-scalable and can run on-premise or on-cloud. Increase interaction and experience real-time video streaming.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 3
    SmokePing

    SmokePing

    The Active Monitoring System

    SmokePing is a latency monitoring tool that measures, visualizes, and alerts on network latency using configurable probes such as ping, fping, or HTTP checks. Developed by Tobi Oetiker, it provides detailed, historical graphs of network performance, making it ideal for diagnosing intermittent issues or verifying SLAs. It uses RRDtool for data storage and graphing.
    Downloads: 53 This Week
    Last Update:
    See Project
  • 4
    Mumble

    Mumble

    Mumble is an open-source, low-latency, high quality voice chat

    Mumble is an open-source, low-latency, high-quality voice chat software. There are two modules in Mumble; the client (mumble) and the server (murmur). The client works on Windows, Linux, FreeBSD, OpenBSD, and macOS, while the server should work on anything Qt can be installed on. Low-latency and high-quality voice-chat program written on top of Qt and Opus. Administrators appreciate Mumble for being able to self-host and have control over data security and privacy.
    Downloads: 15 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    BlackHole

    BlackHole

    BlackHole is a modern macOS audio loopback driver

    BlackHole is a modern virtual audio loopback driver for macOS that allows audio to be routed between applications with zero additional latency. It is widely used by musicians, streamers, developers, and audio engineers who need reliable internal audio routing without external hardware. BlackHole supports a wide range of sample rates and configurable channel counts, making it flexible for both simple and advanced audio workflows. The driver integrates directly with macOS Core Audio and appears in Audio MIDI Setup and supported audio applications. ...
    Downloads: 99 This Week
    Last Update:
    See Project
  • 6
    Atlas OS

    Atlas OS

    An open and lightweight modification to Windows

    Atlas is an open-source, script-based modification for Windows that enhances usability, privacy, and performance. It strips unnecessary background services, disables telemetry, and includes gaming and latency optimizations to make Windows feel lean and fast for enthusiasts.
    Downloads: 2,149 This Week
    Last Update:
    See Project
  • 7
    SysDVR

    SysDVR

    Stream switch games to your PC via USB or network

    ...Video quality is fixed to 720p @ 30fps with h264 compression, this is a hardware limit. Audio quality is fixed to 16bit PCM @ 48kHz stereo. Not compressed. Very low latency with an optimal setup, most games are playable.
    Downloads: 110 This Week
    Last Update:
    See Project
  • 8
    scrcpy

    scrcpy

    Display and control your Android device

    ...It is cross-platform (GNU/Linux, macOS and Windows) and does not require any root access. scrcpy displays only the device screen but offers great performance (30~60fps) and quality (1920×1080 or above). It’s got low latency (35~70ms) and a very low startup time (less than a second). It offers plenty of great features and is non-intrusive, with nothing left installed inside the device. scrcpy works with Android devices with at least API 21 (Android 5.0) and adb debugging must be enabled on the device.
    Downloads: 10,297 This Week
    Last Update:
    See Project
  • 9
    DeepSpeed MII

    DeepSpeed MII

    MII makes low-latency and high-throughput inference possible

    ...While open-sourcing has democratized access to AI capabilities, their application is still restricted by two critical factors: inference latency and cost. DeepSpeed-MII is a new open-source python library from DeepSpeed, aimed towards making low-latency, low-cost inference of powerful models not only feasible but also easily accessible. MII offers access to the highly optimized implementation of thousands of widely used DL models. MII-supported models achieve significantly lower latency and cost compared to their original implementation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    RealtimeSTT

    RealtimeSTT

    A robust, efficient, low-latency speech-to-text library

    RealtimeSTT is a Python-based realtime speech-to-text engine emphasizing low latency, wake-word detection, voice activity detection, and automatic speech segmentation. It provides asynchronous callbacks, nanosecond-precision timestamps, and CLI tools, suitable for building voice assistants, meeting transcribers, or live caption systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    LiteRT

    LiteRT

    LiteRT is the new name for TensorFlow Lite (TFLite)

    LiteRT is an experimental, real-time inference runtime built by Google AI Edge to run lightweight ML models on edge devices with ultra-low latency. It focuses on delivering predictable and consistent performance for models used in time-critical applications like robotics, AR/VR, and IoT. LiteRT is designed to be hardware-agnostic, with minimal dependencies and tight control over execution scheduling.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 12
    Infinispan

    Infinispan

    Infinispan is an open source data grid platform

    Infinispan is a distributed in-memory data grid and caching system designed for high-performance computing. It allows applications to scale dynamically by distributing data across multiple nodes, reducing latency and improving resilience.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Easy GPU PV

    Easy GPU PV

    A Project dedicated to making GPU Partitioning on Windows easier

    Easy-GPU-PV is an open-source project that automates the relatively complex steps required to set up GPU Paravirtualization (GPU-PV) on Windows Hyper-V, letting you partition a physical GPU and assign slices to virtual machines for low-latency remote use. It creates and installs a Windows VM, copies the necessary GPU-PV driver files into the VM, and helps install Parsec (an ultra-low-latency remote-desktop client) so you can connect to the VM at high framerates. The repo is practical and script-driven: the author provides PowerShell scripts (PreChecks.ps1, CopyFilesToVM.ps1, Update-VMGpuPartitionDriver.ps1, etc.) plus configuration values and an unattended install template to minimize manual steps. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 14
    SpacetimeDB

    SpacetimeDB

    Multiplayer at the speed of light

    SpacetimeDB is a high-performance, distributed database optimized for real-time multiplayer applications. It offers low-latency data synchronization across clients, ensuring a seamless and responsive user experience. With built-in conflict resolution and support for complex data types, SpacetimeDB simplifies the development of interactive, collaborative applications.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Selkies-GStreamer

    Selkies-GStreamer

    Open-Source Low-Latency Accelerated Linux WebRTC HTML5 Remote Desktop

    selkies-gstreamer is a GStreamer-based media streaming component used in the Selkies project, a cloud-native platform designed for interactive desktop and application streaming. This module acts as a high-performance media pipeline that captures video, encodes it with low latency, and streams it via WebRTC to client browsers. It is optimized for GPU-accelerated encoding and integrates with Kubernetes-based deployments to enable scalable, real-time remote desktop sessions. This component plays a critical role in delivering smooth, responsive experiences for cloud-based workstations, gaming, or visualization tools.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Qbot

    Qbot

    AI-powered Quantitative Investment Research Platform

    ...It bundles a lightweight GUI client (built with wxPython) and a modular backend so researchers can iterate on strategies, run batch backtests, and validate ideas in a near-real simulated environment that models latency and slippage. The project places special emphasis on AI-driven strategies — including supervised learning, reinforcement learning and multi-factor models — and offers a “model zoo” and example strategies to help users get started. For evaluation and analysis, Qbot integrates reporting and visualization (tearsheets, metrics) so you can compare performance across runs and inspect trade-level behavior. ...
    Downloads: 32 This Week
    Last Update:
    See Project
  • 17
    PeerDB

    PeerDB

    Fast, Simple and a cost effective tool to replicate data from Postgres

    PeerDB is an open-source platform for real-time replication and transformation of data from PostgreSQL to analytical warehouses like BigQuery and Snowflake. It supports Change Data Capture (CDC) and provides seamless syncing and transformation logic with low latency. PeerDB is ideal for teams building real-time data pipelines without relying on expensive proprietary solutions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Canal

    Canal

    MySQL binlog

    ...It enables real-time data synchronization and change data capture (CDC) between MySQL and other systems such as Elasticsearch, Kafka, or HBase. Canal is widely used for data integration, replication, and monitoring across distributed systems, offering high performance and low-latency log parsing.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Lantern Database

    Lantern Database

    PostgreSQL vector database extension for building AI applications

    Lantern is a real-time data transformation engine that enables data engineers to build, run, and monitor streaming data pipelines with SQL. It’s designed to process events in motion, offering low-latency stream transformations, aggregations, and enrichment in a declarative way. Lantern is especially suited for modern data infrastructure and analytics platforms.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    FlashInfer

    FlashInfer

    FlashInfer: Kernel Library for LLM Serving

    FlashInfer is a kernel library designed to enhance the serving of Large Language Models (LLMs) by optimizing inference performance. It provides a high-performance framework that integrates seamlessly with existing systems, aiming to reduce latency and improve efficiency in LLM deployments. FlashInfer supports various hardware architectures and is built to scale with the demands of production environments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Kimi K2.5

    Kimi K2.5

    Moonshot's most powerful AI model

    ...Based on a 1T-parameter Mixture-of-Experts (MoE) architecture with 32B activated parameters, it integrates advanced language reasoning with strong visual understanding. K2.5 supports both “Thinking” and “Instant” modes, enabling either deep step-by-step reasoning or low-latency responses depending on the task. Designed for agentic workflows, it features an Agent Swarm mechanism that decomposes complex problems into coordinated sub-agents executing in parallel. With a 256K context length and MoonViT vision encoder, the model excels across reasoning, coding, long-context comprehension, image, and video benchmarks. ...
    Downloads: 60 This Week
    Last Update:
    See Project
  • 22
    ScaleLLM

    ScaleLLM

    A high-performance inference system for large language models

    ScaleLLM is a high-performance inference system tailored for Large Language Models (LLMs), specifically designed for production environments. It focuses on optimizing inference processes to handle large-scale deployments efficiently, ensuring low latency and high throughput. ScaleLLM supports various LLM architectures and integrates with existing infrastructures, providing a scalable solution for deploying LLMs in real-world applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    hls4ml

    hls4ml

    Machine learning on FPGAs using HLS

    hls4ml is an open-source framework that enables machine learning models to be implemented directly on hardware such as FPGAs and ASICs using high-level synthesis techniques. The system converts trained neural network models from common machine learning frameworks into hardware description code suitable for ultra-low-latency inference. This approach allows machine learning algorithms to run directly on specialized hardware, making them suitable for applications that require extremely fast response times and minimal power consumption. The framework was originally developed for high-energy physics experiments where real-time decision systems must process large volumes of data with strict latency constraints. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    SlateDB

    SlateDB

    A cloud native embedded storage engine built on object storage

    ...Leveraging object storage allows SlateDB to provide bottomless storage capacity, high durability, and easy replication. The trade-off is that object storage has a higher latency and higher API cost than local disk. To mitigate high write API costs (PUTs), SlateDB batches writes. Rather than writing every put() call to object storage, MemTables are flushed periodically to object storage as a string-sorted table (SST). The flush interval is configurable. To mitigate write latency, SlateDB provides an async put method.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    GuitarPedal

    GuitarPedal

    Linus learns analog circuits

    GuitarPedal is an experimental repository exploring a digital guitar-effects signal chain implemented with lean, low-level code. The project demonstrates how to read audio input, process it through simple transformations, and write the result out in real time with minimal latency. It emphasizes straightforward, inspectable DSP so developers can follow the math and tweak parameters without a giant framework in the way. The codebase favors portability and simplicity, focusing on a handful of canonical effects rather than a sprawling plugin architecture. It doubles as a teaching aid for musicians who code, showing how buffers, sampling rates, and numerical stability affect tone. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB