Alternatives to Infervision

Compare Infervision alternatives for your business or organization using the curated list below. SourceForge ranks the best alternatives to Infervision in 2026. Compare features, ratings, user reviews, pricing, and more from Infervision competitors and alternatives in order to make an informed decision for your business.

  • 1
    Artrya

    Artrya

    Artrya

    Artrya partners with clinics receiving chest pain patients to deliver new models of cardiovascular care based on AI. Easily integrate plaque features of high prognostic value into the evaluation of coronary artery disease. Rapid assessment of chest pain patients in emergency and primary care settings. We see a world where communities do not have to be plagued by a high rate of a heart attack. AI-driven insights into the type and volume of arterial plaque detected through coronary computed tomography angiography (CCTA) facilitate the accurate stratification of chest pain patients within minutes. Quickly determine patients with minimal to zero coronary artery disease with the confidence there are no underlying plaque features that could cause future major adverse cardiac events. Catch the early warning signs of a potential heart attack. Effectively rule in at-risk patients with acute or atypical chest pain who warrant further investigation and therapy.
  • 2
    ClearRead Xray

    ClearRead Xray

    Riverain Technologies

    Riverain Technologies' ClearRead Xray is a suite of five FDA-cleared applications designed to enhance the efficiency and accuracy of chest X-ray interpretation across healthcare enterprises without necessitating additional equipment, procedures, or radiation exposure. The platform employs unique suppression technology to provide an unobstructed view of the chest, facilitating the rapid and accurate detection of cardiothoracic diseases. Key features include the production of bone-suppressed images for improved visualization, identification of potential lung cancer nodules, reduction in portable X-ray reading time, automatic comparison of findings with prior exams to highlight changes, and enterprise-wide compatibility across various manufacturers and imaging protocols. Notably, ClearRead Xray Bone Suppress increases the visibility of soft tissue by suppressing bone structures in digital images, aiding radiologists in detecting previously missed nodules.
  • 3
    qCT

    qCT

    Qure.ai

    Qure.ai's qLC-Suite is an AI-powered solution designed to enhance early detection and management of lung nodules, thereby facilitating timely intervention for lung cancer. The suite offers precise quantification, comprehensive characterization, and 3D visualization of lung nodules, ensuring no missed opportunities for early intervention. It supports both incidental and targeted screening by locating nodules and measuring their volume with a single click. Additionally, it tracks volumetric growth over time, aiding in monitoring nodule progression. The qLC-Suite integrates seamlessly into existing workflows, providing rapid analysis and reporting to support clinical decision-making. Comprehensive lung nodule management platform that enables care coordination through smart prompts, hardware-agnostic image viewing for AI-annotated chest X-rays and CTs, cross-departmental scan sharing, and custom notifications for suspect cases.
  • 4
    Aidoc

    Aidoc

    Aidoc Medical

    Aidoc develops advanced healthcare-grade AI based decision support software. Our technology analyzes medical imaging to provide one of the most comprehensive solutions for flagging acute abnormalities across the body, helping radiologists prioritize life threatening cases and expedite patient care. Empowering radiologists to detect urgent cases faster and reducing overall report turnaround time directly in the workflow. 510(k) Triage and notification software indicated for use in the analysis of non-enhanced head CT images; flags and communicates suspected positive findings of pathologies in head CT images, namely Intracranial Hemorrhage (ICH). 510(k) triage and notification software indicated for use in the analysis of Head CTA images; flags and communicates Brain Aneurysm (BA). 510(k) triage and notification software indicated for use in the analysis of Chest X-Ray images; flags and communicates Pneumothorax (Ptx).
  • 5
    KeyChest

    KeyChest

    KeyChest

    Website certificate expiration is easily forgotten—causing costly downtime. Our expert service automatically checks and renews your certificates, on time, and correctly, so you can start every day with confidence. KeyChest is free for personal use. KeyChest is the tool you need to stay on top of all your certificates. Plan your renewals, get your weekly summary and present your certificate performance indicators (KPIs) to your boss. KeyChest has been designed and developed as a tool to help with minimum effort so it can automatically detect new servers without you doing anything. The burden of certificates increases every day. Work stacks up when you track certificates manually. You're strapped for time, and short of qualified people. You’re blindsided by expiring certificates. Downtime costs you money, customers, and peace of mind. KeyChest requests certificates in regular intervals and measures latency of the process.
  • 6
    Join AI

    Join AI

    Miracle Advance Technologies

    Join AI offers advanced artificial intelligence solutions for the healthcare industry, particularly focusing on radiology and endoscopy. Their AI services include chest imaging, breast imaging, stroke detection, and gastrointestinal endoscopy, providing clinicians with enhanced diagnostic capabilities. By utilizing cutting-edge algorithms, Join AI helps healthcare professionals make more accurate and timely diagnoses, improving patient outcomes.
  • 7
    Lunit

    Lunit

    Lunit

    By detecting early-stage cancers, AI enables timely management that can increase the chance of survival through AI. Cancer is complex. Every cancer is unique in its own way, containing vast information that needs to be processed. Complex is what AI does best. With its analysis that extends human capacity, AI can unlock cancer information that leads us to new revelations that can serve as a key strategy in fighting cancer. Armed with the accurate and appropriate information, we can take a step ahead in the war on cancer. Through AI, we can conquer cancer. We develop AI that detects early-stage cancer and optimizes cancer treatment. Detects with 97-99% accuracy. Accurately detects 10 of the most common findings in a chest x-ray. Improving the most basic and universal diagnostic test. Alleviating the burden in radiology workflow. Lunit INSIGHT CXR covers the vast majority of findings. Accurately detects 10 of the most common findings in a chest x-ray.
  • 8
    Augmento

    Augmento

    DeepTek

    X-Ray and CT Scans based AI-powered solution for Public Health Screening. Covers chest pathologies including Tuberculosis and COVID 19 like infections. A state of the art technology to augment Xray for infectious diseases like TB and Covid 19 patterns and instantly prescreen and triage. We have evolved the entire workflow around them making imaging assessment, diagnosis, and reporting seamless. A smart notification removes the stress out of the process of reporting. Three key highlights - classify, localize, and quantify lesions. Leveraging Artificial Intelligence for providing prompt and efficient teleradiology services to the global imaging marketplace.
  • 9
    HeartFlow

    HeartFlow

    HeartFlow

    HeartFlow's non-invasive personalized cardiac test provides unprecedented visualization of each patient's coronary arteries, enabling physicians to create more effective treatment plans for their patients. The HeartFlow FFRCT Analysis starts when a patient undergoes a standard coronary computed tomography scan at a hospital or imaging center. The CT images are securely uploaded to our cloud. Next, we use advanced algorithms incorporating artificial intelligence to build a personalized, digital model of that patient’s coronary arteries. Our team of highly trained analysts then inspects this model, making any needed edits. Once this patient-specific model is completed, the HeartFlow pathway applies physiologic principles and computational fluid dynamics to compute the blood flow and FFRCT values at every point in the model. Throughout the process, we follow rigorous and well-established protocols to ensure consistent processing for every patient.
  • 10
    Medecom

    Medecom

    Medecom

    Medecom develop software solutions for radiology and mammography since 2000. We have 3 families of products : diagnostic solutions for X Rays and mammography including stitching and tomosynthesis, Mini PACS and PACS, RIS and flat panels solutions. Artificial intelligence is available for chest X Rays and mammography.
  • 11
    MedGemma

    MedGemma

    Google DeepMind

    MedGemma is a collection of Gemma 3 variants that are trained for performance on medical text and image comprehension. Developers can use MedGemma to accelerate building healthcare-based AI applications. MedGemma currently comes in two variants: a 4B multimodal version and a 27B text-only version. MedGemma 4B utilizes a SigLIP image encoder that has been specifically pre-trained on a variety of de-identified medical data, including chest X-rays, dermatology images, ophthalmology images, and histopathology slides. Its LLM component is trained on a diverse set of medical data, including radiology images, histopathology patches, ophthalmology images, and dermatology images. MedGemma 4B is available in both pre-trained (suffix: -pt) and instruction-tuned (suffix -it) versions. The instruction-tuned version is a better starting point for most applications.
  • 12
    ROKAPAX

    ROKAPAX

    ROKAPAX

    Our product portfolio includes: VNA, PACS systems, Radiology Information System (RIS), Diagnostic Viewers, AI-based Medical Diagnostic supporting Solutions (MDSS) for MMG, CT chest, ECG and endoprosthesis (on Xray). There are more then 300 000 monthly active users, 20 000 units of diagnostic equipment been integrated with our PACS|VNA. We offer you our innovative technological solutions for your patients and are ready to share our experience of Healthcare digitalization to improve the lives of patients. We are open to ensure the implementation of our solutions to improve the diagnostic process for end users - radiologists, oncologists and cardiologists. Our core competencies developed over 14 years of experience working with 5,000 clinics: - comprehensive and modular platform for the entire imaging process (vendor independent) AI-based medical decision support systems for all scenarios: - warning about the priority of cases with pathologies (red-yellow-green) on one screen
  • 13
    Medis Suite XA

    Medis Suite XA

    Medis Medical Imaging Systems

    Medis Suite XA is our solution for X-Ray angiography. With a workflow based on 30+ years of experience in cardiovascular image analysis, Medis Suite XA is a complete package that provides various modules including a highly practical viewer, various coronary and vascular analyses that allow for anatomical evaluation of the arteries, various analyses for both ventricles, as well as integrated reporting. The analysis of coronary and peripheral vessels (QCA and QVA) in angiograms. The analysis of left and right ventriculograms (QLV and QRV). Innovative QFR® analysis, a measure of the functional significance of lesions, without the need for adenosine and/or a pressure wire. Seamless integration in the healthcare IT environment and an easy connection with the DICOM network. Innovative QFR® analysis, a measure of the functional significance of lesions, without the need for adenosine and/or a pressure wire.
  • 14
    Genki

    Genki

    DEEPTEK

    X-Ray and CT scans-based AI-powered solution for public health screening. It covers chest pathologies, including Tuberculosis and COVID-19 like infections. Genki solution is "AI+Expert in the loop" and provides an "end-to-end imaging workflow." A unique human-in-the-loop system where AI augments imaging experts generates reports faster and enhances productivity. ​State-of-the-art is used technology to augment Xray for infectious diseases like TB and Covid 19 patterns and instantly prescreen and triage. We have evolved the entire workflow around them, making imaging assessment, diagnosis, and reporting seamless. A smart notification removes the stress from the process of reporting. Three key highlights, classify, localize, and quantify lesions. Genki applies point-of-care diagnostics, which ushers the much-needed last-mile approach in bridging the gaps in disease elimination. Its solutions can be embedded in mobile X-ray units / CT scanners to ensure instant triage/ pre-screening.
  • 15
    Amazon Elastic Inference
    Amazon Elastic Inference allows you to attach low-cost GPU-powered acceleration to Amazon EC2 and Sagemaker instances or Amazon ECS tasks, to reduce the cost of running deep learning inference by up to 75%. Amazon Elastic Inference supports TensorFlow, Apache MXNet, PyTorch and ONNX models. Inference is the process of making predictions using a trained model. In deep learning applications, inference accounts for up to 90% of total operational costs for two reasons. Firstly, standalone GPU instances are typically designed for model training - not for inference. While training jobs batch process hundreds of data samples in parallel, inference jobs usually process a single input in real time, and thus consume a small amount of GPU compute. This makes standalone GPU inference cost-inefficient. On the other hand, standalone CPU instances are not specialized for matrix operations, and thus are often too slow for deep learning inference.
  • 16
    Tensormesh

    Tensormesh

    Tensormesh

    Tensormesh is a caching layer built specifically for large-language-model inference workloads that enables organizations to reuse intermediate computations, drastically reduce GPU usage, and accelerate time-to-first-token and latency. It works by capturing and reusing key-value cache states that are normally thrown away after each inference, thereby cutting redundant compute and delivering “up to 10x faster inference” while substantially lowering GPU load. It supports deployments in public cloud or on-premises, with full observability and enterprise-grade control, SDKs/APIs, and dashboards for integration into existing inference pipelines, and compatibility with inference engines such as vLLM out of the box. Tensormesh emphasizes performance at scale, including sub-millisecond repeated queries, while optimizing every layer of inference from caching through computation.
  • 17
    PaliGemma 2
    PaliGemma 2, the next evolution in tunable vision-language models, builds upon the performant Gemma 2 models, adding the power of vision and making it easier than ever to fine-tune for exceptional performance. With PaliGemma 2, these models can see, understand, and interact with visual input, opening up a world of new possibilities. It offers scalable performance with multiple model sizes (3B, 10B, 28B parameters) and resolutions (224px, 448px, 896px). PaliGemma 2 generates detailed, contextually relevant captions for images, going beyond simple object identification to describe actions, emotions, and the overall narrative of the scene. Our research demonstrates leading performance in chemical formula recognition, music score recognition, spatial reasoning, and chest X-ray report generation, as detailed in the technical report. Upgrading to PaliGemma 2 is a breeze for existing PaliGemma users.
  • 18
    Amazon SageMaker Model Deployment
    Amazon SageMaker makes it easy to deploy ML models to make predictions (also known as inference) at the best price-performance for any use case. It provides a broad selection of ML infrastructure and model deployment options to help meet all your ML inference needs. It is a fully managed service and integrates with MLOps tools, so you can scale your model deployment, reduce inference costs, manage models more effectively in production, and reduce operational burden. From low latency (a few milliseconds) and high throughput (hundreds of thousands of requests per second) to long-running inference for use cases such as natural language processing and computer vision, you can use Amazon SageMaker for all your inference needs.
  • 19
    NVIDIA Triton Inference Server
    NVIDIA Triton™ inference server delivers fast and scalable AI in production. Open-source inference serving software, Triton inference server streamlines AI inference by enabling teams deploy trained AI models from any framework (TensorFlow, NVIDIA TensorRT®, PyTorch, ONNX, XGBoost, Python, custom and more on any GPU- or CPU-based infrastructure (cloud, data center, or edge). Triton runs models concurrently on GPUs to maximize throughput and utilization, supports x86 and ARM CPU-based inferencing, and offers features like dynamic batching, model analyzer, model ensemble, and audio streaming. Triton helps developers deliver high-performance inference aTriton integrates with Kubernetes for orchestration and scaling, exports Prometheus metrics for monitoring, supports live model updates, and can be used in all major public cloud machine learning (ML) and managed Kubernetes platforms. Triton helps standardize model deployment in production.
  • 20
    kluster.ai

    kluster.ai

    kluster.ai

    Kluster.ai is a developer-centric AI cloud platform designed to deploy, scale, and fine-tune large language models (LLMs) with speed and efficiency. Built for developers by developers, it offers Adaptive Inference, a flexible and scalable service that adjusts seamlessly to workload demands, ensuring high-performance processing and consistent turnaround times. Adaptive Inference provides three distinct processing options: real-time inference for ultra-low latency needs, asynchronous inference for cost-effective handling of flexible timing tasks, and batch inference for efficient processing of high-volume, bulk tasks. It supports a range of open-weight, cutting-edge multimodal models for chat, vision, code, and more, including Meta's Llama 4 Maverick and Scout, Qwen3-235B-A22B, DeepSeek-R1, and Gemma 3 . Kluster.ai's OpenAI-compatible API allows developers to integrate these models into their applications seamlessly.
    Starting Price: $0.15per input
  • 21
    Amazing.photos

    Amazing.photos

    Amazing.photos

    We help you create a great impression - using AI to give you an excellent profile picture. We use your photos to train an AI model for you, this AI model is trained on your photos and is private to you. We then create AI avatars/profile pictures for you. The output is highly realistic, your model is not shared with anyone else. You can delete your model and your photos at any time. You can download them, share them, delete them, sell them, get tattoos of them on your chest, build giant heroic stone statues of them - the whole lot. Our business relies on our reputation of treating your data with respect.
    Starting Price: $21 one-time payment
  • 22
    Hugging Face Transformers
    ​Transformers is a library of pretrained natural language processing, computer vision, audio, and multimodal models for inference and training. Use Transformers to train models on your data, build inference applications, and generate text with large language models. Explore the Hugging Face Hub today to find a model and use Transformers to help you get started right away.​ Simple and optimized inference class for many machine learning tasks like text generation, image segmentation, automatic speech recognition, document question answering, and more. A comprehensive trainer that supports features such as mixed precision, torch.compile, and FlashAttention for training and distributed training for PyTorch models.​ Fast text generation with large language models and vision language models. Every model is implemented from only three main classes (configuration, model, and preprocessor) and can be quickly used for inference or training.
    Starting Price: $9 per month
  • 23
    KServe

    KServe

    KServe

    Highly scalable and standards-based model inference platform on Kubernetes for trusted AI. KServe is a standard model inference platform on Kubernetes, built for highly scalable use cases. Provides performant, standardized inference protocol across ML frameworks. Support modern serverless inference workload with autoscaling including a scale to zero on GPU. Provides high scalability, density packing, and intelligent routing using ModelMesh. Simple and pluggable production serving for production ML serving including prediction, pre/post-processing, monitoring, and explainability. Advanced deployments with the canary rollout, experiments, ensembles, and transformers. ModelMesh is designed for high-scale, high-density, and frequently-changing model use cases. ModelMesh intelligently loads and unloads AI models to and from memory to strike an intelligent trade-off between responsiveness to users and computational footprint.
  • 24
    NVIDIA DGX Cloud Serverless Inference
    NVIDIA DGX Cloud Serverless Inference is a high-performance, serverless AI inference solution that accelerates AI innovation with auto-scaling, cost-efficient GPU utilization, multi-cloud flexibility, and seamless scalability. With NVIDIA DGX Cloud Serverless Inference, you can scale down to zero instances during periods of inactivity to optimize resource utilization and reduce costs. There's no extra cost for cold-boot start times, and the system is optimized to minimize them. NVIDIA DGX Cloud Serverless Inference is powered by NVIDIA Cloud Functions (NVCF), which offers robust observability features. It allows you to integrate your preferred monitoring tools, such as Splunk, for comprehensive insights into your AI workloads. NVCF offers flexible deployment options for NIM microservices while allowing you to bring your own containers, models, and Helm charts.
  • 25
    Infermedica API

    Infermedica API

    Infermedica

    Infermedica is a leading digital health company, specializing in AI-powered solutions for early symptoms assessment, digital triage and care navigation. Through its technology healthcare organizations can introduce solutions for symptoms checking, patient intake, follow-up, call center support and more. Infermedica’s Medical Guidance Platform is a Class IIb Medical Device under the MDR in the European Union, and is developed with the highest standards of quality and data protection, compliant with ISO, HIPAA, GDPR, SOC2. Infermedica has been adeptly interweaving cutting edge technologies, including AI, LLMs, and NLP, with the proven Medical Knowledge and Inference Engine that lies in the core of its technologies. Those technologies can be accessed through various touchpoints, including web, mobile, call centers, voice agents and chat bots. Infermedica’s solutions achieve 94% of accuracy. To learn more, visit our webpage.
  • 26
    EdgeCortix

    EdgeCortix

    EdgeCortix

    Breaking the limits in AI processors and edge AI inference acceleration. Where AI inference acceleration needs it all, more TOPS, lower latency, better area and power efficiency, and scalability, EdgeCortix AI processor cores make it happen. General-purpose processing cores, CPUs, and GPUs, provide developers with flexibility for most applications. However, these general-purpose cores don’t match up well with workloads found in deep neural networks. EdgeCortix began with a mission in mind: redefining edge AI processing from the ground up. With EdgeCortix technology including a full-stack AI inference software development environment, run-time reconfigurable edge AI inference IP, and edge AI chips for boards and systems, designers can deploy near-cloud-level AI performance at the edge. Think about what that can do for these and other applications. Finding threats, raising situational awareness, and making vehicles smarter.
  • 27
    NetMind AI

    NetMind AI

    NetMind AI

    NetMind.AI is a decentralized computing platform and AI ecosystem designed to accelerate global AI innovation. By leveraging idle GPU resources worldwide, it offers accessible and affordable AI computing power to individuals, businesses, and organizations of all sizes. The platform provides a range of services, including GPU rental, serverless inference, and an AI ecosystem that encompasses data processing, model training, inference, and agent development. Users can rent GPUs at competitive prices, deploy models effortlessly with on-demand serverless inference, and access a wide array of open-source AI model APIs with high-throughput, low-latency performance. NetMind.AI also enables contributors to add their idle GPUs to the network, earning NetMind Tokens (NMT) as rewards. These tokens facilitate transactions on the platform, allowing users to pay for services such as training, fine-tuning, inference, and GPU rentals.
  • 28
    Together AI

    Together AI

    Together AI

    Together AI provides an AI-native cloud platform built to accelerate training, fine-tuning, and inference on high-performance GPU clusters. Engineered for massive scale, the platform supports workloads that process trillions of tokens without performance drops. Together AI delivers industry-leading cost efficiency by optimizing hardware, scheduling, and inference techniques, lowering total cost of ownership for demanding AI workloads. With deep research expertise, the company brings cutting-edge models, hardware, and runtime innovations—like ATLAS runtime-learning accelerators—directly into production environments. Its full-stack ecosystem includes a model library, inference APIs, fine-tuning capabilities, pre-training support, and instant GPU clusters. Designed for AI-native teams, Together AI helps organizations build and deploy advanced applications faster and more affordably.
    Starting Price: $0.0001 per 1k tokens
  • 29
    NVIDIA NeMo Megatron
    NVIDIA NeMo Megatron is an end-to-end framework for training and deploying LLMs with billions and trillions of parameters. NVIDIA NeMo Megatron, part of the NVIDIA AI platform, offers an easy, efficient, and cost-effective containerized framework to build and deploy LLMs. Designed for enterprise application development, it builds upon the most advanced technologies from NVIDIA research and provides an end-to-end workflow for automated distributed data processing, training large-scale customized GPT-3, T5, and multilingual T5 (mT5) models, and deploying models for inference at scale. Harnessing the power of LLMs is made easy through validated and converged recipes with predefined configurations for training and inference. Customizing models is simplified by the hyperparameter tool, which automatically searches for the best hyperparameter configurations and performance for training and inference on any given distributed GPU cluster configuration.
  • 30
    MaiaOS

    MaiaOS

    Zyphra Technologies

    Zyphra is an artificial intelligence company based in Palo Alto with a growing presence in Montreal and London. We’re building MaiaOS, a multimodal agent system combining advanced research in next-gen neural network architectures (SSM hybrids), long-term memory & reinforcement learning. We believe the future of AGI will involve a combination of cloud and on-device deployment strategies with an increasing shift toward local inference. MaiaOS is built around a deployment framework that maximizes inference efficiency for real-time intelligence. Our AI & product teams come from leading organizations and institutions including Google DeepMind, Anthropic, StabilityAI, Qualcomm, Neuralink, Nvidia, and Apple. We have deep expertise across AI models, learning algorithms, and systems/infrastructure with a focus on inference efficiency and AI silicon performance. Zyphra's team is committed to democratizing advanced AI systems.
  • 31
    Stanhope AI

    Stanhope AI

    Stanhope AI

    Active Inference is a novel framework for agentic AI based on world models, emerging from over 30 years of research in computational neuroscience. From this paradigm, we offer an AI built for power and computational efficiency, designed to live on-device and on the edge. Integrating with traditional computer vision stacks our intelligent decision-making systems provide an explainable output that allows organizations to build accountability into their AI tools and products. We are taking active inference from neuroscience into AI as the foundation for software that will allow robots and embodied platforms to make autonomous decisions like the human brain.
  • 32
    Dash0

    Dash0

    Dash0

    Dash0 is an OpenTelemetry-native observability platform that unifies metrics, logs, traces, and resources into one intuitive interface, enabling fast and context-rich monitoring without vendor lock-in. It centralizes Prometheus and OpenTelemetry metrics, supports powerful filtering of high-cardinality attributes, and provides heatmap drilldowns and detailed trace views to pinpoint errors and bottlenecks in real time. Users benefit from fully customizable dashboards built on Perses, with support for code-based configuration and Grafana import, plus seamless integration with predefined alerts, checks, and PromQL queries. Dash0's AI-enhanced tools, such as Log AI for automated severity inference and pattern extraction, enrich telemetry data without requiring users to even notice that AI is working behind the scenes. These AI capabilities power features like log classification, grouping, inferred severity tagging, and streamlined triage workflows through the SIFT framework.
    Starting Price: $0.20 per month
  • 33
    Watchman

    Watchman

    Watchman

    Watchman AI is an AI-native demand inference platform that captures invisible B2B buyers by identifying, researching, enriching, and qualifying company accounts and person-level leads in real time, turning anonymous visitors into an endless stream of qualified prospects without forms or manual work. It operates full-autopilot, using dynamic inference agents to lock in precision prospects the moment they land, automatically curating and syncing those leads into an existing stack with zero friction, and replacing underperforming tools by surfacing the accounts that move revenue. The system emphasizes efficiency and growth, continuously capturing and converting stealth traffic 24/7 while automating workflows so teams save hours of manual effort and scale demand capture.
    Starting Price: $32 per month
  • 34
    AutoGen

    AutoGen

    Microsoft

    An Open-Source Programming Framework for Agentic AI. AutoGen provides multi-agent conversation framework as a high-level abstraction. With this framework, one can conveniently build LLM workflows. AutoGen offers a collection of working systems spanning a wide range of applications from various domains and complexities. AutoGen supports enhanced LLM inference APIs, which can be used to improve inference performance and reduce cost.
  • 35
    Amazon EC2 G4 Instances
    Amazon EC2 G4 instances are optimized for machine learning inference and graphics-intensive applications. It offers a choice between NVIDIA T4 GPUs (G4dn) and AMD Radeon Pro V520 GPUs (G4ad). G4dn instances combine NVIDIA T4 GPUs with custom Intel Cascade Lake CPUs, providing a balance of compute, memory, and networking resources. These instances are ideal for deploying machine learning models, video transcoding, game streaming, and graphics rendering. G4ad instances, featuring AMD Radeon Pro V520 GPUs and 2nd-generation AMD EPYC processors, deliver cost-effective solutions for graphics workloads. Both G4dn and G4ad instances support Amazon Elastic Inference, allowing users to attach low-cost GPU-powered inference acceleration to Amazon EC2 and reduce deep learning inference costs. They are available in various sizes to accommodate different performance needs and are integrated with AWS services such as Amazon SageMaker, Amazon ECS, and Amazon EKS.
  • 36
    eXprts

    eXprts

    Cobite

    eXprts is a lot more than just an invoice processing and expense management system. We complement a core system based on a fully-automated invoice processing, validation and audit engine, with numerous integrated features that will provide your company with a full medicine chest for your company's wireless management headaches. eXprts includes a battle-tested Web 2.0 procurement portal, which integrates into your existing approval systems and business processes. When devices arrive on site, our arrival and inventory tracking tool provides the safeguards that ensure that your commitments to end-users are being met, you can even maintain your own device pools to cut procurement time to hours instead of days. Our comprehensive package of software and services for managing Market Data entitlements, inventory, and invoices provides a single provider solution for simplifying the dizzying complexity of procurement, management, and cost control in the area of market data.
    Starting Price: $1000 per month
  • 37
    NVIDIA TensorRT
    NVIDIA TensorRT is an ecosystem of APIs for high-performance deep learning inference, encompassing an inference runtime and model optimizations that deliver low latency and high throughput for production applications. Built on the CUDA parallel programming model, TensorRT optimizes neural network models trained on all major frameworks, calibrating them for lower precision with high accuracy, and deploying them across hyperscale data centers, workstations, laptops, and edge devices. It employs techniques such as quantization, layer and tensor fusion, and kernel tuning on all types of NVIDIA GPUs, from edge devices to PCs to data centers. The ecosystem includes TensorRT-LLM, an open source library that accelerates and optimizes inference performance of recent large language models on the NVIDIA AI platform, enabling developers to experiment with new LLMs for high performance and quick customization through a simplified Python API.
  • 38
    FriendliAI

    FriendliAI

    FriendliAI

    FriendliAI is a generative AI infrastructure platform that offers fast, efficient, and reliable inference solutions for production environments. It provides a suite of tools and services designed to optimize the deployment and serving of large language models (LLMs) and other generative AI workloads at scale. Key offerings include Friendli Endpoints, which allow users to build and serve custom generative AI models, saving GPU costs and accelerating AI inference. It supports seamless integration with popular open source models from the Hugging Face Hub, enabling lightning-fast, high-performance inference. FriendliAI's cutting-edge technologies, such as Iteration Batching, Friendli DNN Library, Friendli TCache, and Native Quantization, contribute to significant cost savings (50–90%), reduced GPU requirements (6× fewer GPUs), higher throughput (10.7×), and lower latency (6.2×).
    Starting Price: $5.9 per hour
  • 39
    Amazon EC2 Inf1 Instances
    Amazon EC2 Inf1 instances are purpose-built to deliver high-performance and cost-effective machine learning inference. They provide up to 2.3 times higher throughput and up to 70% lower cost per inference compared to other Amazon EC2 instances. Powered by up to 16 AWS Inferentia chips, ML inference accelerators designed by AWS, Inf1 instances also feature 2nd generation Intel Xeon Scalable processors and offer up to 100 Gbps networking bandwidth to support large-scale ML applications. These instances are ideal for deploying applications such as search engines, recommendation systems, computer vision, speech recognition, natural language processing, personalization, and fraud detection. Developers can deploy their ML models on Inf1 instances using the AWS Neuron SDK, which integrates with popular ML frameworks like TensorFlow, PyTorch, and Apache MXNet, allowing for seamless migration with minimal code changes.
    Starting Price: $0.228 per hour
  • 40
    Astria

    Astria

    Astria

    Tailor-made AI image generation, start creating your unique images. Align your crew with the most detailed, custom-made visual references. Previs to the max. Find the most attractive visualization for your product. Instant realization of your vision, with limitless variations. Realize your super-specific concepts with augmented creativity. Experiment, modify, and fine-tune. Upload 10-20 pictures of your subject. Preferably shot or cropped to a 1:1 aspect ratio. We recommend uploading 3 photos of the full body or entire object + 5 medium shot photos from the chest up + 10 close-ups. Change body poses for every picture, use pictures from different days' backgrounds and lighting, and show a variety of expressions and emotions. Make sure you capture the subject's eyes looking in different directions for different images, take one with closed eyes. Every picture of your subject should introduce new info about your subject.
    Starting Price: $0.10 per prompt
  • 41
    Protopia AI

    Protopia AI

    Protopia AI

    Protopia AI’s Stained Glass Transform (SGT) offers a cutting-edge solution to secure sensitive data in AI workloads by preventing data exposure during processing and inference. It enables enterprises to maximize the value of their data by breaking down silos while retaining full ownership and security. SGT supports deployment across diverse environments, including on-premises, hybrid, and multi-tenant clouds, optimizing GPU use for performance. It runs up to 14,000 times faster than traditional cryptographic methods, ensuring AI inference adds only minimal latency. The platform is designed to meet the needs of industries with strict data privacy requirements, such as finance, defense, and healthcare. Protopia’s technology integrates with AWS Marketplace and partners like Lambda and vLLM to provide comprehensive, high-performance, secure AI inference solutions.
  • 42
    Mirai

    Mirai

    Mirai

    Mirai is a developer-focused on-device AI infrastructure platform designed to convert, optimize, and run machine learning models directly on Apple devices with high performance and privacy. It provides a unified pipeline that enables teams to convert and quantize models, benchmark them, distribute them, and execute inference locally. It is built specifically for Apple Silicon and aims to deliver near-zero latency, zero inference cost, and full data privacy by keeping sensitive processing on the user’s device. Through its SDK and inference engine, developers can integrate AI features into applications quickly, using hardware-aware optimizations that unlock the full power of the GPU and Neural Engine. Mirai also includes dynamic routing capabilities that automatically decide whether a request should run locally or in the cloud based on latency, privacy, or workload requirements.
  • 43
    StatXact
    StatXact® 12 provides statisticians in disciplines across the social and natural sciences with the world’s most expansive toolkit for exact inference and power analysis. Offering more than 160 tests and procedures, the StatXact® suite has transformed the complex algorithms of modern statistical analysis into a validated software package with a user-friendly interface. StatXact offers more tests and procedures for exact inference and power analysis than any other software package on the market. Fast and accurate inferences are compiled on an easy-to-use interface, drawing on complex algorithms created by Cytel’s team of statistical experts. To help leaders in life sciences unlock the power of data, solve their toughest problems, and make evidence-driven decisions with confidence. Ready to discuss your project with Cytel? Speak to an expert about your statistical and advanced analytics requirements.
    Starting Price: $995.00
  • 44
    Tinfoil

    Tinfoil

    Tinfoil

    Tinfoil is a verifiably private AI platform built to deliver zero-trust, zero-data-retention inference by running open-source or custom models inside secure hardware enclaves in the cloud, giving you the data-privacy assurances of on-premises systems with the scalability and convenience of the cloud. All user inputs and inference operations are processed in confidential-computing environments so that no one, not even Tinfoil or the cloud provider, can access or retain your data. It supports private chat, private data analysis, user-trained fine-tuning, and an OpenAI-compatible inference API, covers workloads such as AI agents, private content moderation, and proprietary code models, and provides features like public verification of enclave attestation, “provable zero data access,” and full compatibility with major open source models.
  • 45
    HPC-AI

    HPC-AI

    HPC-AI

    HPC-AI is an enterprise AI infrastructure and GPU cloud platform designed to accelerate deep learning training, inference, and large-scale compute workloads with high performance and cost efficiency. It delivers a pre-configured AI-optimized stack that enables rapid deployment and real-time inference while supporting demanding workloads that require high IOPS, ultra-low latency, and massive throughput. It provides a robust GPU cloud environment built for artificial intelligence, high-performance computing, and other compute-intensive applications, giving teams the tools needed to run complex workflows efficiently. At its core, the company’s software focuses on parallel and distributed training, inference, and fine-tuning of large neural networks, helping organizations reduce infrastructure costs while maintaining performance. It is powered in part by technologies such as Colossal-AI, which significantly accelerates model training and improves productivity.
    Starting Price: $3.05 per hour
  • 46
    SmartSig

    SmartSig

    DrFirst

    An estimated 66% of data from the nation’s largest medication history database is missing essential sig information. That can leave your patients at significant risk of an adverse drug event or worse. It doesn’t have to be that way. SmartSigSM from DrFirst fills the information gaps left from missing sig data quickly, efficiently and accurately. Building on robust predictive data delivered by SmartStrings, which automatically identifies the most commonly prescribed medications, SmartSig intelligently translates free text data, even safely inferring missing components, using statistics and clinical analytics to produce a real-time translation in a click. This essential interoperability tool takes less than two hours to implement and is compatible with major electronic records giving you more time to spend with your patients and a proven way to improve patient safety.
  • 47
    Dreamtonics Synthesizer V
    Warmth and tonality are hallmarks of the human singing voice. Behind the scenes, Synthesize V leverages a deep neural network-based synthesis engine capable of generating incredibly life-like singing voices. Plus, unlike other solutions that utilize neural networks, our first-of-its-kind synthesizer is 100% offline yet runs at lightning-fast speeds. Bad connection? No worries, you will never lose access to your work. Experiment with an expanding inventory of voices ready to plug and play with Synthesizer V Studio. Dive deeper and customize voices with dynamic vocal modes like chest, belt, and breathy. Visualize your modifications in waveforms in real-time via the live rendering feature, helping you minimize hearing fatigue and reduce the idea-to-sound cycle. Synthesizer V AI voices are available natively in English, Japanese and Chinese. Plus, the cross-lingual synthesis feature breaks the language barrier, empowering any voice to sing in any of our three languages!
    Starting Price: $79 one-time payment
  • 48
    Size-Me

    Size-Me

    Bodi.Me

    Size-Me is a fast, cost-effective, and accurate garment size recommendation tool for B2B and B2C apparel businesses. It uses touch-free body measurements and requires no intrusive wearer photos or body scans. Our garment fit solution uses high-quality data from more than 350,000 3D body scans and models, the latest machine learning, and optimized algorithms to extrapolate a complete body profile from a few simple measurements with a high degree of accuracy. The software supports plugins for the major e-commerce platforms. From just height, weight, and chest or bra size, we can recommend the best size for each wearer from your range, based on your garment data. This means fewer returns, improved customer satisfaction and retention, more accurate inventory, and better sell-through. This ultra-user-friendly solution simplifies the fitting process with a set of straightforward questions for each wearer, eliminating the need for intrusive photos or body scans.
  • 49
    Inferable

    Inferable

    Inferable

    Create your first AI automation in 60 seconds. Inferable seamlessly integrates with your existing codebase and infrastructure, allowing you to create powerful AI automation without compromising on control or security. Works with your existing codebase. Integrates with your existing services via opt-in. Enforce determinism through source code. Create and manage automation programmatically. You own the computer, in your own infrastructure. Inferable comes out of the box with delightful DX to kickstart your AI automation journey. We bring the best in class vertically integrated LLM orchestration. You bring your product and domain expertise. At the core of Inferable is a distributed message queue that ensures your AI automation is scalable and reliable. It makes sure your automations are executed correctly, and that failures are handled gracefully. Decorate your existing functions, REST APIs, and GraphQL endpoints with decorators to require human approval.
    Starting Price: $0.006 per KB
  • 50
    Amazon SageMaker Feature Store
    Amazon SageMaker Feature Store is a fully managed, purpose-built repository to store, share, and manage features for machine learning (ML) models. Features are inputs to ML models used during training and inference. For example, in an application that recommends a music playlist, features could include song ratings, listening duration, and listener demographics. Features are used repeatedly by multiple teams and feature quality is critical to ensure a highly accurate model. Also, when features used to train models offline in batch are made available for real-time inference, it’s hard to keep the two feature stores synchronized. SageMaker Feature Store provides a secured and unified store for feature use across the ML lifecycle. Store, share, and manage ML model features for training and inference to promote feature reuse across ML applications. Ingest features from any data source including streaming and batch such as application logs, service logs, clickstreams, sensors, etc.