Related Products
|
||||||
About
NVIDIA Triton™ inference server delivers fast and scalable AI in production. Open-source inference serving software, Triton inference server streamlines AI inference by enabling teams deploy trained AI models from any framework (TensorFlow, NVIDIA TensorRT®, PyTorch, ONNX, XGBoost, Python, custom and more on any GPU- or CPU-based infrastructure (cloud, data center, or edge). Triton runs models concurrently on GPUs to maximize throughput and utilization, supports x86 and ARM CPU-based inferencing, and offers features like dynamic batching, model analyzer, model ensemble, and audio streaming. Triton helps developers deliver high-performance inference aTriton integrates with Kubernetes for orchestration and scaling, exports Prometheus metrics for monitoring, supports live model updates, and can be used in all major public cloud machine learning (ML) and managed Kubernetes platforms. Triton helps standardize model deployment in production.
|
About
An end-to-end open source machine learning platform. TensorFlow is an end-to-end open source platform for machine learning. It has a comprehensive, flexible ecosystem of tools, libraries and community resources that lets researchers push the state-of-the-art in ML and developers easily build and deploy ML powered applications. Build and train ML models easily using intuitive high-level APIs like Keras with eager execution, which makes for immediate model iteration and easy debugging. Easily train and deploy models in the cloud, on-prem, in the browser, or on-device no matter what language you use. A simple and flexible architecture to take new ideas from concept to code, to state-of-the-art models, and to publication faster. Build, deploy, and experiment easily with TensorFlow.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Developers and companies searching for an inference server solution to improve AI production
|
Audience
Organizations interested in a powerful open source machine learning platform
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
Free
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationNVIDIA
United States
developer.nvidia.com/nvidia-triton-inference-server
|
Company InformationTensorFlow
Founded: 2015
United States
www.tensorflow.org
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
|
|||||
|
|
|
|||||
|
|
|
|||||
Categories |
Categories |
|||||
Machine Learning Features
Deep Learning
ML Algorithm Library
Model Training
Natural Language Processing (NLP)
Predictive Modeling
Statistical / Mathematical Tools
Templates
Visualization
|
||||||
Integrations
HPE Ezmeral
NVIDIA DeepStream SDK
Vertex AI
AWS Marketplace
Amazon EKS
Amazon SageMaker
Amazon SageMaker Studio Lab
Azure Data Science Virtual Machines
Collimator
Deep Lake
|
Integrations
HPE Ezmeral
NVIDIA DeepStream SDK
Vertex AI
AWS Marketplace
Amazon EKS
Amazon SageMaker
Amazon SageMaker Studio Lab
Azure Data Science Virtual Machines
Collimator
Deep Lake
|
|||||
|
|
|