Showing 488 open source projects for "hybrid"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Yampa

    Yampa

    Functional Reactive Programming domain-specific language

    Yampa is a Functional Reactive Programming (FRP) library for Haskell, specifically designed for modeling hybrid systems that involve both continuous and discrete time behaviors, such as games, simulations, robotics, and reactive systems. Based on the concept of signal functions, Yampa offers a declarative way to model time-varying values and their transformations, making it easier to manage complex time-based logic without resorting to imperative state management.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Amazon SSM Agent

    Amazon SSM Agent

    An agent to enable remote management of your EC2 instances

    AWS Systems Manager is a secure end-to-end management solution for hybrid cloud environments. AWS Systems Manager is the operations hub for your AWS applications and resources, and is broken into four core feature groups. Aggregate data in a single console and gain actionable insights across AWS services such as Amazon CloudWatch, AWS CloudTrail, and AWS Config, as well as third-party tools. Leverage operational data to easily manage applications and identify issues quickly across associated AWS resource groups. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 3
    NullClaw

    NullClaw

    Fastest, smallest, and fully autonomous AI assistant infrastructure

    ...At just 678 KB with ~1 MB peak RAM usage, it boots in under 2 milliseconds and runs on virtually any hardware, including low-cost ARM boards. Despite its size, it delivers a complete AI stack with 22+ model providers, 18+ communication channels, integrated tools, hybrid memory, and sandboxed runtime support. Its architecture is fully modular, using vtable interfaces that allow providers, channels, tools, memory backends, and runtimes to be swapped without code changes. NullClaw is secure by design, enforcing pairing-based authentication, strict sandboxing, encrypted secrets, resource limits, and workspace scoping by default. ...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 4
    Advanced + Agentic RAG Cookbooks

    Advanced + Agentic RAG Cookbooks

    Advanced RAG cookbooks for building accurate LLM applications

    ...Athina AI’s RAG Cookbooks covers the full RAG pipeline, including indexing, retrieval, augmentation, and generation, while also addressing evaluation to measure accuracy and relevance. It includes multiple approaches such as hybrid search, contextual compression, and agent-based retrieval strategies, allowing users to experiment and compare methods. It is designed to reduce development time by offering practical examples and references to research papers, making it useful for both learning and production use. Overall, it serves as a hands-on resource for improving LLM outputs using external data sources.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    GLM-4.5

    GLM-4.5

    GLM-4.5: Open-source LLM for intelligent agents by Z.ai

    GLM-4.5 is a cutting-edge open-source large language model designed by Z.ai for intelligent agent applications. The flagship GLM-4.5 model has 355 billion total parameters with 32 billion active parameters, while the compact GLM-4.5-Air version offers 106 billion total parameters and 12 billion active parameters. Both models unify reasoning, coding, and intelligent agent capabilities, providing two modes: a thinking mode for complex reasoning and tool usage, and a non-thinking mode for...
    Downloads: 45 This Week
    Last Update:
    See Project
  • 6
    CUDA-Q

    CUDA-Q

    C++ and Python support for the CUDA Quantum programming model

    CUDA-Q is an open-source platform for developing hybrid quantum-classical applications using a unified programming model across CPUs, GPUs, and quantum processing units. It provides a full toolchain that includes compilers, runtimes, and libraries for writing quantum programs in both C++ and Python. The platform is designed to be hardware-agnostic, allowing developers to run applications on different quantum backends or simulate them efficiently using GPU acceleration when physical quantum hardware is unavailable. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Lithops

    Lithops

    A multi-cloud framework for big data analytics

    ...Lithops is ideal for data processing, ML inference, and embarrassingly parallel workloads, giving you the power of FaaS (Function-as-a-Service) without vendor lock-in. It also supports hybrid cloud setups, object storage access, and simple integration with Jupyter notebooks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    MiniMax-M1

    MiniMax-M1

    Open-weight, large-scale hybrid-attention reasoning model

    MiniMax-M1 is presented as the world’s first open-weight, large-scale hybrid-attention reasoning model, designed to push the frontier of long-context, tool-using, and deeply “thinking” language models. It is built on the MiniMax-Text-01 foundation and keeps the same massive parameter budget, but reworks the attention and training setup for better reasoning and test-time compute scaling. Architecturally, it combines Mixture-of-Experts layers with lightning attention, enabling the model to support a native context length of 1 million tokens while using far fewer FLOPs than comparable reasoning models for very long generations. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    HeavyDB

    HeavyDB

    HeavyDB (formerly MapD/OmniSciDB)

    ...HeavyDB was originally developed as part of the OmniSci platform (formerly MapD) and is commonly used for large-scale analytics and geospatial data processing. The database compiles queries into optimized machine code that executes efficiently on GPU hardware, significantly accelerating analytical workloads. It supports hybrid deployment environments where queries can run on both CPU and GPU architectures depending on the available resources.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    DreamFactory

    DreamFactory

    DreamFactory API Management Platform

    DreamFactory can run on private, public, or hybrid cloud infrastructure. DreamFactory is easy to deploy with Docker, Kubernetes, or directly from GitHub using our automated installers. Eliminate the biggest bottleneck in modern IT. Launch your project in weeks rather than months. DreamFactory instantly creates a secure, standardized, reusable, fully documented, and live REST API.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Youtu-Agent

    Youtu-Agent

    A simple yet powerful agent framework that delivers with models

    ...The framework supports automated generation of agent components, enabling the system to synthesize prompts, tool interfaces, and workflow configurations automatically. Youtu-Agent also incorporates hybrid learning strategies that combine experience accumulation with reinforcement learning to improve agent performance over time. These learning mechanisms allow agents to refine their reasoning, coding, and search capabilities as they interact with environments and tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Pathway AI Pipelines

    Pathway AI Pipelines

    Ready-to-run cloud templates for RAG

    ...It supports numerous connectors including local files, Google Drive, SharePoint, Kafka, PostgreSQL, and real-time APIs, making it suitable for enterprise data environments. The templates include built-in indexing, vector search, hybrid search, and caching capabilities that remove the need to assemble separate infrastructure components. Developers can run the applications locally or deploy them to cloud platforms using Docker with minimal setup. Overall, llm-app functions as a practical accelerator for teams building real-time, production-ready AI knowledge systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ownCloud Infinite Scale

    ownCloud Infinite Scale

    ownCloud Infinite Scale Stack

    ...Because of its modular design, oCIS can be deployed efficiently in public, private, or hybrid infrastructures while reducing operational overhead.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    HunyuanVideo-Foley

    HunyuanVideo-Foley

    Multimodal Diffusion with Representation Alignment

    ...The model architecture aligns audio, video, and text representations to produce realistic synchronized soundtracks. Produces high-quality 48 kHz audio output suitable for professional use. Hybrid architecture combining multimodal transformer blocks and unimodal refinement blocks. Temporal alignment via frame-level synchronization modules (e.g. Synchformer).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    trieve

    trieve

    All-in-one infrastructure for search, recommendations, RAG

    Trieve is an all-in-one infrastructure for building hybrid vector search, recommendations, and RAG.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Tempesta FW

    Tempesta FW

    All-in-one solution for high performance web content delivery

    ...This is a drop-in-replacement for the whole web server frontend infrastructure: an HTTPS load balancer, a web accelerator, a DDoS mitigation system, and a web application firewall (WAF). Tempesta FW is the first and only hybrid of a Web accelerator and a multi-layer firewall. This unique architecture provides seamless integration with the Linux iptables or nftables. Tempesta FW services up to 1.8M HTTP requests per second on the cheapest hardware, which is x3 faster than Nginx or HAProxy. Tempesta TLS is about 40-80% faster than Nginx/OpenSSL and provides up to x4 lower latency.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Wan2.2

    Wan2.2

    Wan2.2: Open and Advanced Large-Scale Video Generative Model

    ...The model is trained on significantly larger datasets than its predecessor, greatly enhancing motion complexity, semantic understanding, and aesthetic diversity. Wan2.2 also open-sources a 5-billion parameter high-compression VAE-based hybrid text-image-to-video (TI2V) model that supports 720P video generation at 24fps on consumer-grade GPUs like the RTX 4090. It supports multiple video generation tasks including text-to-video.
    Downloads: 173 This Week
    Last Update:
    See Project
  • 18
    NVIDIA cuOpt

    NVIDIA cuOpt

    GPU accelerated decision optimization

    ...The platform provides multiple interfaces, including C, Python, and server APIs, allowing developers to integrate optimization capabilities into applications and services. cuOpt is designed for high-performance environments and can be deployed across cloud, hybrid, or on-premise infrastructures. By combining GPU acceleration with scalable APIs, cuOpt enables organizations to solve large optimization challenges in logistics, operations research, and decision-making systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    SeaGOAT

    SeaGOAT

    local-first semantic code search engine

    ...The tool runs locally on a developer’s machine and processes repositories using a combination of embedding models and conventional search utilities, enabling both semantic and text-based retrieval methods. By combining vector search with tools like ripgrep, SeaGOAT provides a hybrid approach that supports both natural language queries and precise keyword matching in source files. It is built primarily in Python and is intended to work on common operating systems such as Linux, macOS, and Windows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    uzu

    uzu

    A high-performance inference engine for AI models

    ...Written primarily in Rust and leveraging Apple’s Metal framework, the project focuses on maximizing performance when executing large language models and other AI workloads on devices such as Mac computers with M-series chips. The engine implements a hybrid architecture in which model layers can be executed either as custom GPU kernels or through Apple’s MPSGraph API, allowing it to balance performance and compatibility depending on the workload. By utilizing Apple’s unified memory architecture, uzu reduces memory copying overhead and improves inference throughput for local AI workloads. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    PowerInfer

    PowerInfer

    High-speed Large Language Model Serving for Local Deployment

    ...Its architecture exploits the observation that only a subset of neurons in large models are frequently activated, allowing the system to preload frequently used neurons into GPU memory while processing less common activations on the CPU. This hybrid execution strategy significantly reduces memory bottlenecks and improves overall inference speed. PowerInfer incorporates specialized algorithms and sparse operators to manage neuron activation patterns and minimize data transfers between hardware components. As a result, it enables powerful language models to run on consumer hardware while achieving performance comparable to more expensive server-grade systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    GLM-Image

    GLM-Image

    GLM-Image: Auto-regressive for Dense-knowledge and High-fidelity Image

    GLM-Image is an open-source generative AI model designed to create high-fidelity images from text prompts using a hybrid architecture that combines autoregressive semantic understanding with diffusion-based detail refinement. It excels at generating images that include complex layouts and detailed text content, making it especially useful for posters, diagrams, info-graphics, social media graphics, and visual content that requires precise text placement and semantic alignment. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    ZeusDB Vector Database

    ZeusDB Vector Database

    Blazing-fast vector DB with similarity search and metadata filtering

    ...The storage layer is designed for durability and growth, supporting sharding, replication, and background compaction while keeping query tails predictable. Developers get multiple ingestion paths—batch, streaming, and upsert—making it easy to keep embeddings synchronized as content changes. Hybrid search is a core design goal, allowing you to mix vector, keyword, and filter queries in a single request for practical relevance. Observability and safety round out the system, with metrics, tracing, and guardrails to manage recalls, deletions, and privacy-sensitive data at scale.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    FastVLM

    FastVLM

    This repository contains the official implementation of FastVLM

    FastVLM is an efficiency-focused vision-language modeling stack that introduces FastViTHD, a hybrid vision encoder engineered to emit fewer visual tokens and slash encoding time, especially for high-resolution images. Instead of elaborate pruning stages, the design trades off resolution and token count through input scaling, simplifying the pipeline while maintaining strong accuracy. Reported results highlight dramatic speedups in time-to-first-token and competitive quality versus contemporary open VLMs, including comparisons across small and larger variants. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ML Ferret

    ML Ferret

    Refer and Ground Anything Anywhere at Any Granularity

    Ferret is Apple’s end-to-end multimodal large language model designed specifically for flexible referring and grounding: it can understand references of any granularity (boxes, points, free-form regions) and then ground open-vocabulary descriptions back onto the image. The core idea is a hybrid region representation that mixes discrete coordinates with continuous visual features, so the model can fluidly handle “any-form” referring while maintaining precise spatial localization. The repo presents the vision-language pipeline, model assets, and paper resources that show how Ferret answers questions, follows instructions, and returns grounded outputs rather than just text. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB