Browse free open source LLM Inference tools and projects for Windows below. Use the toggles on the left to filter open source LLM Inference tools by OS, license, language, programming language, and project status.
Port of OpenAI's Whisper model in C/C++
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
User-friendly AI Interface
ONNX Runtime: cross-platform, high performance ML inferencing
High-performance neural network inference framework for mobile
C++ library for high performance inference on NVIDIA GPUs
A high-throughput and memory-efficient inference and serving engine
Ready-to-use OCR with 80+ supported languages
OpenVINO™ Toolkit repository
Self-hosted, community-driven, local OpenAI compatible API
Open standard for machine learning interoperability
Open-Source AI Camera. Empower any camera/CCTV
Database system for building simpler and faster AI-powered application
Tensor search for humans
A high-performance ML model serving framework, offers dynamic batching
A scalable inference server for models optimized with OpenVINO
Uncover insights, surface problems, monitor, and fine tune your LLM
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
Protect and discover secrets using Gitleaks
Sequence-to-sequence framework, focused on Neural Machine Translation
The official Python client for the Huggingface Hub
Guide to deploying deep-learning inference networks
Lightweight anchor-free object detection model
Everything you need to build state-of-the-art foundation models