Showing 101 open source projects for "ollama"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    Ollama

    Ollama

    Get up and running with Llama 2 and other large language models

    Run, create, and share large language models (LLMs). Get up and running with large language models, locally. Run Llama 2 and other models on macOS. Customize and create your own.
    Downloads: 567 This Week
    Last Update:
    See Project
  • 2
    Ollama Python

    Ollama Python

    Ollama Python library

    ollama-python is an open-source Python SDK that wraps the Ollama CLI, allowing seamless interaction with local large language models (LLMs) managed by Ollama. Developers use it to load models, send prompts, manage sessions, and stream responses directly from Python code. It simplifies integration of Ollama-based models into applications, supporting synchronous and streaming modes.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    handy-ollama

    handy-ollama

    Implement CPU from scratch and play with large model deployments

    handy-ollama is an open-source educational project designed to help developers and AI enthusiasts learn how to deploy and run large language models locally using the Ollama platform. The repository serves as a structured tutorial that explains how to install, configure, and use Ollama to run modern language models on personal hardware without requiring advanced infrastructure.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    NextJS Ollama LLM UI

    NextJS Ollama LLM UI

    Fully-featured web interface for Ollama LLMs

    NextJS Ollama LLM UI is a web-based frontend interface built with Next.js to make interacting with Ollama-hosted large language models easy and fast. Its goal is to remove the complexity of setting up and managing UI components for local or offline LLM usage by providing a straightforward chat experience with support for responsive layouts, light and dark themes, and local chat history storage in the browser.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Powerful App Monitoring Without Surprise Bills Icon
    Powerful App Monitoring Without Surprise Bills

    AppSignal starts at $23/month with all features included. No overages, no hidden fees. 30-day free trial.

    Tired of monitoring tools that punish you for scaling? AppSignal offers transparent, predictable pricing with every feature unlocked on every plan. Track errors, monitor performance, detect anomalies, and manage logs across Ruby, Python, Node.js, and more. Trusted by developers since 2012 with free dev-to-dev support. No credit card required to start your 30-day trial.
    Try AppSignal Free
  • 5
    Ollamac

    Ollamac

    Mac app for Ollama

    Ollamac is an open-source native macOS application that provides a graphical interface for interacting with local large language models running through the Ollama inference framework. The project was created to simplify the process of using local AI models, which typically require command-line interaction, by offering a clean and intuitive desktop interface. Through this interface, users can run and chat with a variety of LLM models installed through Ollama directly on their own machines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    OllamaSharp

    OllamaSharp

    The easiest way to use Ollama in .NET

    OllamaSharp is an open-source .NET library that provides strongly typed bindings for interacting with the Ollama API, making it easier for developers to integrate local large language models into C# and .NET applications. The project acts as a wrapper around the Ollama API, exposing all endpoints through asynchronous methods that allow developers to perform tasks such as generating text, creating embeddings, and managing models. It supports both local and remote Ollama instances, enabling developers to run AI models on their own hardware or connect to remote model servers. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Open WebUI

    Open WebUI

    User-friendly AI Interface

    ...Additionally, Open WebUI offers a Progressive Web App (PWA) for mobile devices, providing offline access and a native app-like experience. The platform also includes a Model Builder, allowing users to create custom models from base Ollama models directly within the interface. With over 156,000 users, Open WebUI is a versatile solution for deploying and managing AI models in a secure, offline environment.
    Downloads: 111 This Week
    Last Update:
    See Project
  • 8
    Chipper

    Chipper

    AI interface for tinkerers (Ollama, Haystack RAG, Python)

    Chipper is an AI interface designed for tinkerers and developers, providing a platform to experiment with various AI models and techniques. It offers integration with tools like Ollama and Haystack for Retrieval-Augmented Generation (RAG), enabling users to build and test AI applications efficiently. Chipper supports Python and provides a modular architecture, allowing for customization and extension based on specific project requirements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Maid

    Maid

    Maid is a cross-platform Flutter app for interfacing with GGUF

    Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely. Maid is a cross-platform free and open source application for interfacing with llama.cpp models locally, and remotely with Ollama, Mistral, Google Gemini and OpenAI models remotely. Maid supports Sillytavern character cards to allow you to interact with all your favorite characters. Maid supports downloading a curated list of Models in-app directly from hugging face.
    Downloads: 10 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 10
    oterm

    oterm

    the terminal client for Ollama

    Oterm is an open-source terminal client designed to provide a lightweight command-line interface for interacting with large language models through the Ollama ecosystem. The tool allows users to chat with local AI models directly from the terminal without needing a graphical interface or web application. Its interface is designed to be simple and intuitive, enabling developers to launch conversations quickly using a single command. Oterm supports persistent chat sessions that store conversations, system prompts, and parameter configurations locally in a database. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    CodeCompanion

    CodeCompanion

    AI-powered coding, seamlessly in Neovim. Supports Anthropic, etc.

    Currently supports Anthropic, Copilot, Gemini, Ollama and OpenAI adapters.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 12
    DeepChat

    DeepChat

    A smart assistant that connects powerful AI to your personal world

    ...DeepChat is a powerful open-source AI chat platform providing a unified interface for interacting with various large language models. Whether you're using cloud APIs like OpenAI, Gemini, Anthropic, or locally deployed Ollama models, DeepChat delivers a smooth user experience. As a cross-platform AI assistant application, DeepChat not only supports basic chat functionality but also offers advanced features such as search enhancement, tool calling, and multimodal interaction, making AI capabilities more accessible and efficient.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    ai-renamer

    ai-renamer

    A Node.js CLI that uses Ollama and LM Studio models

    ...Instead of relying on manual naming or metadata, the tool analyzes the actual content of files, including images, videos, and documents, to generate descriptive and context-aware filenames. It integrates with local and cloud-based AI providers such as Ollama, LM Studio, and OpenAI, allowing users to choose between offline and API-based workflows depending on their needs. The tool supports batch processing, making it particularly useful for organizing large collections of files quickly and efficiently. It also provides customization options such as naming conventions, language preferences, and prompt modifications to tailor the output to specific use cases. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    Gollama

    Gollama

    Go manage your Ollama models

    Gollama is a macOS and Linux tool for managing Ollama models through an interactive terminal-based interface. It provides a TUI that lets users list, inspect, sort, filter, edit, run, unload, copy, rename, delete, and push models from one place rather than relying entirely on manual command-line workflows. The project is aimed at developers and local AI users who frequently work with multiple Ollama models and want a more efficient operational layer for everyday maintenance. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    chatd

    chatd

    Chat with your documents using local AI

    ...The software focuses on privacy and security by ensuring that all document processing and inference occur entirely on the user’s computer without sending data to external cloud services. It includes a built-in integration with the Ollama runtime, which provides a cross-platform environment for running large language models locally. The application typically runs models such as Mistral-7B and allows users to load and analyze documents while asking questions in natural language. Unlike many document-chat tools that require manual installation of model servers, chatd packages the model runner with the application so that users can start interacting with documents immediately after launching the program.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    ...The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows developers to use powerful open-source models such as Llama, Phi, DeepSeek, and Qwen while maintaining privacy and avoiding external service dependencies. The system supports contextual queries where the AI analyzes files within a directory and generates answers based on project documentation or source code. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    bolt.diy

    bolt.diy

    Prompt, run, edit, & deploy full-stack web applications using any LLM

    bolt.diy is an open-source platform that allows you to easily create, run, edit, and deploy full-stack web applications using a variety of large language models (LLMs). It supports popular models like OpenAI, Anthropic, Ollama, OpenRouter, Gemini, LMStudio, Mistral, xAI, HuggingFace, DeepSeek, and Groq, and provides the flexibility to integrate additional models through the Vercel AI SDK. Whether you’re experimenting with pre-built models or developing custom AI-driven applications, bolt.diy offers a smooth and intuitive experience for building AI-powered web apps. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 18
    Large Language Models (LLMs)

    Large Language Models (LLMs)

    Connect MATLAB to LLM APIs, including OpenAI® Chat Completions

    This repository enables MATLAB to connect with large language models (LLMs) such as OpenAI's ChatGPT, DALL-E, Azure OpenAI, and Ollama, integrating their natural language processing and image generation capabilities directly within MATLAB environments. It facilitates creating chatbots, summarizing text, and image generation, among other tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Enchanted

    Enchanted

    Enchanted is iOS and macOS app for chatting with language models

    Enchanted is an open-source, cross-platform application built to let users chat with privately hosted large language models from Apple devices, including macOS, iOS, and visionOS. Designed to work seamlessly with servers like Ollama, it provides a privacy-focused alternative to traditional cloud AI UIs by connecting directly to your own LLM endpoints such as Llama, Mistral, Vicuna, and more. The interface resembles familiar commercial chat apps but emphasizes local control, offline capabilities, and multimodal support, making it ideal for users who want rich AI interaction without exposing sensitive prompts or conversations to third-party services. ...
    Downloads: 16 This Week
    Last Update:
    See Project
  • 20
    Perplexica

    Perplexica

    Perplexica is an AI-powered answering engine.

    Perplexica is a privacy-focused AI answering engine like Perplexity that you can self-host on your own hardware for private, source-cited web research. It combines live internet search results with AI models, letting you use local LLMs via Ollama or connect to providers like OpenAI, Claude, Gemini, and Groq. Powered by SearxNG, it aggregates results from multiple search engines while keeping your identity and queries private. Perplexica offers multiple search modes—Speed, Balanced, and Quality—so you can trade off latency and depth depending on the task. It also enhances the experience with widgets and rich search types like images, videos, and domain-limited queries. ...
    Downloads: 26 This Week
    Last Update:
    See Project
  • 21
    OpenJarvis

    OpenJarvis

    Personal AI, On Personal Devices

    ...The framework provides shared primitives for building local-first agents, along with evaluation tools that measure performance using metrics such as energy consumption, latency, cost, and accuracy. OpenJarvis integrates with local inference engines like Ollama, vLLM, SGLang, and llama.cpp to run language models directly on personal hardware. It also includes a learning loop that allows models to improve over time using locally generated interaction traces. By prioritizing local execution and efficiency, OpenJarvis aims to provide a foundation for privacy-preserving personal AI assistants.
    Downloads: 20 This Week
    Last Update:
    See Project
  • 22
    Dive

    Dive

    Dive is an open-source MCP Host Desktop Application

    Dive is an open‑source MCP host desktop application that serves as a bridge between MCP servers and any large language models supporting function calling, designed to deliver a seamless AI agent experience across environments. Compatible with ChatGPT, Anthropic, Ollama and OpenAI-compatible models. Enabling seamless MCP AI agent integration on both stdio and SSE mode. One-click access to managed MCP servers via OAPHub.ai - eliminates complex local deployments. Modern Tauri version alongside traditional Electron version for optimal performance.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24

    ollama-kit-v1.1

    Portable AI stack with Ollama container, models, and Gramified GUI

    Full portable AI stack with container + models Includes: Pre-built Ollama container (v0.11.4) All included models: Mistral, LLaMA3, Phi, Gemma, Qwen, CodeLLaMA, etc. Gramified GUI for model interaction Launch scripts for Linux, WSL, and Windows Usage: Extract the archive to your preferred location. Load the container and start Ollama. Launch the Gramified GUI as described in the README. Linux extract example: unzip ollama-kit-v1.1-full.zip Note: This release is large due to included model files. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    KubeAI

    KubeAI

    Private Open AI on Kubernetes

    ...Admins can configure ML models by using the Model Kubernetes Custom Resources. KubeAI can be thought of as a Model Operator (See Operator Pattern) that manages vLLM and Ollama servers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB