Browse free open source LLM Inference tools and projects for Windows and Linux below. Use the toggles on the left to filter open source LLM Inference tools by OS, license, language, programming language, and project status.
ONNX Runtime: cross-platform, high performance ML inferencing
High-performance neural network inference framework for mobile
C++ library for high performance inference on NVIDIA GPUs
OpenVINO™ Toolkit repository
Open standard for machine learning interoperability
Deep learning optimization library: makes distributed training easy
Guide to deploying deep-learning inference networks
The deep learning toolkit for speech-to-text
Easy-to-use deep learning framework with 3 key features
A set of Docker images for training and serving models in TensorFlow
Implementation of model parallel autoregressive transformers on GPUs
PArallel Distributed Deep LEarning: Machine Learning Framework
Powering Amazon custom machine learning chips
A GPU-accelerated library containing highly optimized building blocks
Deep Learning API and Server in C++14 support for Caffe, PyTorch
MII makes low-latency and high-throughput inference possible
Deep learning inference framework optimized for mobile platforms
OpenMMLab Model Deployment Framework
Trainable models and NN optimization tools
A unified framework for scalable computing
Library for serving Transformers models on Amazon SageMaker
Toolkit for allowing inference and serving with MXNet in SageMaker
Uniform deep learning inference framework for mobile
Probabilistic reasoning and statistical analysis in TensorFlow
A computer vision framework to create and deploy apps in minutes