Showing 87 open source projects for "ollama mac"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Ollama

    Ollama

    Get up and running with Llama 2 and other large language models

    Run, create, and share large language models (LLMs). Get up and running with large language models, locally. Run Llama 2 and other models on macOS. Customize and create your own.
    Downloads: 523 This Week
    Last Update:
    See Project
  • 2
    Ollama Python

    Ollama Python

    Ollama Python library

    ollama-python is an open-source Python SDK that wraps the Ollama CLI, allowing seamless interaction with local large language models (LLMs) managed by Ollama. Developers use it to load models, send prompts, manage sessions, and stream responses directly from Python code. It simplifies integration of Ollama-based models into applications, supporting synchronous and streaming modes. This tool is ideal for those building AI-driven apps with local model deployment.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    handy-ollama

    handy-ollama

    Implement CPU from scratch and play with large model deployments

    handy-ollama is an open-source educational project designed to help developers and AI enthusiasts learn how to deploy and run large language models locally using the Ollama platform. The repository serves as a structured tutorial that explains how to install, configure, and use Ollama to run modern language models on personal hardware without requiring advanced infrastructure. A key focus of the project is enabling users to run large models even without GPUs by leveraging optimized CPU-based...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    NextJS Ollama LLM UI

    NextJS Ollama LLM UI

    Fully-featured web interface for Ollama LLMs

    NextJS Ollama LLM UI is a web-based frontend interface built with Next.js to make interacting with Ollama-hosted large language models easy and fast. Its goal is to remove the complexity of setting up and managing UI components for local or offline LLM usage by providing a straightforward chat experience with support for responsive layouts, light and dark themes, and local chat history storage in the browser. The interface stores conversations in local storage, so no separate backend...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Ollamac

    Ollamac

    Mac app for Ollama

    Ollamac is an open-source native macOS application that provides a graphical interface for interacting with local large language models running through the Ollama inference framework. The project was created to simplify the process of using local AI models, which typically require command-line interaction, by offering a clean and intuitive desktop interface. Through this interface, users can run and chat with a variety of LLM models installed through Ollama directly on their own machines....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    OllamaSharp

    OllamaSharp

    The easiest way to use Ollama in .NET

    OllamaSharp is an open-source .NET library that provides strongly typed bindings for interacting with the Ollama API, making it easier for developers to integrate local large language models into C# and .NET applications. The project acts as a wrapper around the Ollama API, exposing all endpoints through asynchronous methods that allow developers to perform tasks such as generating text, creating embeddings, and managing models. It supports both local and remote Ollama instances, enabling...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Open WebUI

    Open WebUI

    User-friendly AI Interface

    Open WebUI is an extensible, feature-rich, and user-friendly self-hosted AI platform designed to operate entirely offline. It supports various LLM runners like Ollama and OpenAI-compatible APIs, with a built-in inference engine for Retrieval Augmented Generation (RAG), making it a powerful AI deployment solution. Key features include effortless setup via Docker or Kubernetes, seamless integration with OpenAI-compatible APIs, granular permissions and user groups for enhanced security,...
    Downloads: 90 This Week
    Last Update:
    See Project
  • 8
    Chipper

    Chipper

    AI interface for tinkerers (Ollama, Haystack RAG, Python)

    Chipper is an AI interface designed for tinkerers and developers, providing a platform to experiment with various AI models and techniques. It offers integration with tools like Ollama and Haystack for Retrieval-Augmented Generation (RAG), enabling users to build and test AI applications efficiently. Chipper supports Python and provides a modular architecture, allowing for customization and extension based on specific project requirements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    oterm

    oterm

    the terminal client for Ollama

    Oterm is an open-source terminal client designed to provide a lightweight command-line interface for interacting with large language models through the Ollama ecosystem. The tool allows users to chat with local AI models directly from the terminal without needing a graphical interface or web application. Its interface is designed to be simple and intuitive, enabling developers to launch conversations quickly using a single command. Oterm supports persistent chat sessions that store...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 10
    DeepChat

    DeepChat

    A smart assistant that connects powerful AI to your personal world

    DeepChat is an open‑source, multi‑model AI chat platform featuring a unified interface for cloud and local language models, enriched with tool‑calling capabilities, search enhancements, privacy protection, and extensive model support. DeepChat is a powerful open-source AI chat platform providing a unified interface for interacting with various large language models. Whether you're using cloud APIs like OpenAI, Gemini, Anthropic, or locally deployed Ollama models, DeepChat delivers a smooth...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 11
    ai-renamer

    ai-renamer

    A Node.js CLI that uses Ollama and LM Studio models

    ai-renamer is a Node.js-based command-line tool that uses large language models to automatically rename files based on their content, enabling more meaningful and organized file management. Instead of relying on manual naming or metadata, the tool analyzes the actual content of files, including images, videos, and documents, to generate descriptive and context-aware filenames. It integrates with local and cloud-based AI providers such as Ollama, LM Studio, and OpenAI, allowing users to...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 12
    CodeCompanion

    CodeCompanion

    AI-powered coding, seamlessly in Neovim. Supports Anthropic, etc.

    Currently supports Anthropic, Copilot, Gemini, Ollama and OpenAI adapters.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    tlm is an open-source command-line AI assistant designed to provide intelligent terminal support using locally running large language models. The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Gollama

    Gollama

    Go manage your Ollama models

    Gollama is a macOS and Linux tool for managing Ollama models through an interactive terminal-based interface. It provides a TUI that lets users list, inspect, sort, filter, edit, run, unload, copy, rename, delete, and push models from one place rather than relying entirely on manual command-line workflows. The project is aimed at developers and local AI users who frequently work with multiple Ollama models and want a more efficient operational layer for everyday maintenance. Beyond standard...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    bolt.diy

    bolt.diy

    Prompt, run, edit, & deploy full-stack web applications using any LLM

    bolt.diy is an open-source platform that allows you to easily create, run, edit, and deploy full-stack web applications using a variety of large language models (LLMs). It supports popular models like OpenAI, Anthropic, Ollama, OpenRouter, Gemini, LMStudio, Mistral, xAI, HuggingFace, DeepSeek, and Groq, and provides the flexibility to integrate additional models through the Vercel AI SDK. Whether you’re experimenting with pre-built models or developing custom AI-driven applications, bolt.diy...
    Downloads: 13 This Week
    Last Update:
    See Project
  • 16
    Enchanted

    Enchanted

    Enchanted is iOS and macOS app for chatting with language models

    Enchanted is an open-source, cross-platform application built to let users chat with privately hosted large language models from Apple devices, including macOS, iOS, and visionOS. Designed to work seamlessly with servers like Ollama, it provides a privacy-focused alternative to traditional cloud AI UIs by connecting directly to your own LLM endpoints such as Llama, Mistral, Vicuna, and more. The interface resembles familiar commercial chat apps but emphasizes local control, offline...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 17
    chatd

    chatd

    Chat with your documents using local AI

    chatd is an open-source desktop application that allows users to interact with their documents through a locally running large language model. The software focuses on privacy and security by ensuring that all document processing and inference occur entirely on the user’s computer without sending data to external cloud services. It includes a built-in integration with the Ollama runtime, which provides a cross-platform environment for running large language models locally. The application...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Large Language Models (LLMs)

    Large Language Models (LLMs)

    Connect MATLAB to LLM APIs, including OpenAI® Chat Completions

    This repository enables MATLAB to connect with large language models (LLMs) such as OpenAI's ChatGPT, DALL-E, Azure OpenAI, and Ollama, integrating their natural language processing and image generation capabilities directly within MATLAB environments. It facilitates creating chatbots, summarizing text, and image generation, among other tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Perplexica

    Perplexica

    Perplexica is an AI-powered answering engine.

    Perplexica is a privacy-focused AI answering engine like Perplexity that you can self-host on your own hardware for private, source-cited web research. It combines live internet search results with AI models, letting you use local LLMs via Ollama or connect to providers like OpenAI, Claude, Gemini, and Groq. Powered by SearxNG, it aggregates results from multiple search engines while keeping your identity and queries private. Perplexica offers multiple search modes—Speed, Balanced, and...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 20
    Dive

    Dive

    Dive is an open-source MCP Host Desktop Application

    Dive is an open‑source MCP host desktop application that serves as a bridge between MCP servers and any large language models supporting function calling, designed to deliver a seamless AI agent experience across environments. Compatible with ChatGPT, Anthropic, Ollama and OpenAI-compatible models. Enabling seamless MCP AI agent integration on both stdio and SSE mode. One-click access to managed MCP servers via OAPHub.ai - eliminates complex local deployments. Modern Tauri version alongside...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 21
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    OpenJarvis

    OpenJarvis

    Personal AI, On Personal Devices

    OpenJarvis is an open-source framework designed to build personal AI agents that run primarily on local devices rather than relying on cloud infrastructure. Developed as part of the Intelligence Per Watt research initiative, it focuses on improving the efficiency and practicality of on-device AI systems. The framework provides shared primitives for building local-first agents, along with evaluation tools that measure performance using metrics such as energy consumption, latency, cost, and...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 23
    KubeAI

    KubeAI

    Private Open AI on Kubernetes

    Get inferencing running on Kubernetes: LLMs, Embeddings, Speech-to-Text. KubeAI serves an OpenAI compatible HTTP API. Admins can configure ML models by using the Model Kubernetes Custom Resources. KubeAI can be thought of as a Model Operator (See Operator Pattern) that manages vLLM and Ollama servers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Void Editor

    Void Editor

    Open source AI IDE and Cursor alternative

    Void is an open-source, AI-powered code editor built as a fork of Visual Studio Code. Designed as a fully transparent and privacy-focused alternative to Cursor or GitHub Copilot, it lets you use AI models locally or via APIs (OpenAI, Claude, Gemini, Ollama, etc.)—without routing data through proprietary servers. Developed by YC-backed startup Glass Devtools, it supports traditional coding features inherited from VS Code, enhanced with in-editor LLM capabilities—autocomplete, inline quick...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 25
    Chatbox

    Chatbox

    The Ultimate AI Copilot on Your Desktop

    Chatbox is a cross-platform desktop AI client designed to give you a fast, polished, and private way to work with modern language models. It runs locally on Windows, macOS, and Linux, keeping your conversations and data stored on your own device. Chatbox acts as a unified interface for popular LLMs like ChatGPT, Claude, Gemini, and local models via Ollama, making it easy to switch providers without changing tools. Built with an ergonomic UI, it’s optimized for long sessions, prompt...
    Downloads: 21 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • Next
MongoDB Logo MongoDB