Self-hosted, community-driven, local OpenAI compatible API. Drop-in replacement for OpenAI running LLMs on consumer-grade hardware. Free Open Source OpenAI alternative. No GPU is required. Runs ggml, GPTQ, onnx, TF compatible models: llama, gpt4all, rwkv, whisper, vicuna, koala, gpt4all-j, cerebras, falcon, dolly, starcoder, and many others. LocalAI is a drop-in replacement REST API that’s compatible with OpenAI API specifications for local inferencing. It allows you to run LLMs (and not only) locally or on-prem with consumer-grade hardware, supporting multiple model families that are compatible with the ggml format. Does not require GPU.

Features

  • Local, OpenAI drop-in alternative REST API
  • NO GPU required
  • Supports multiple models
  • Once loaded the first time, it keep models loaded in memory for faster inference
  • Doesn’t shell-out, but uses C++ bindings for a faster inference and better performance
  • You own your data

Project Samples

Project Activity

See All Activity >

License

MIT License

Follow LocalAI

LocalAI Web Site

Other Useful Business Software
Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
Try Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of LocalAI!

Additional Project Details

Programming Language

Go

Related Categories

Go Large Language Models (LLM), Go LLM Inference Tool

Registered

2023-08-21