Bring the notion of Model-as-a-Service to life
Powering Amazon custom machine learning chips
The deep learning toolkit for speech-to-text
A GPU-accelerated library containing highly optimized building blocks
LLMs as Copilots for Theorem Proving in Lean
Official inference library for Mistral models
Lightweight anchor-free object detection model
Easy-to-use Speech Toolkit including Self-Supervised Learning model
A RWKV management and startup tool, full automation, only 8MB
Data manipulation and transformation for audio signal processing
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
The AI-native (edge and LLM) proxy for agents
An Open-Source Programming Framework for Agentic AI
Deep learning optimization library: makes distributed training easy
State-of-the-art diffusion models for image and audio generation
Standardized Serverless ML Inference Platform on Kubernetes
Easy-to-use deep learning framework with 3 key features
A library for accelerating Transformer models on NVIDIA GPUs
AICI: Prompts as (Wasm) Programs
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
Build Production-ready Agentic Workflow with Natural Language
GPU environment management and cluster orchestration
C#/.NET binding of llama.cpp, including LLaMa/GPT model inference
A Pythonic framework to simplify AI service building
Framework for Accelerating LLM Generation with Multiple Decoding Heads