Showing 1147 open source projects for "llm"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 1
    LLM Datasets

    LLM Datasets

    Curated list of datasets and tools for post-training

    LLM Datasets curates and standardizes datasets commonly used to train and fine-tune large language models, reducing the overhead of hunting down sources and normalizing formats. The repository aims to make datasets easy to inspect and transform, with scripts for downloading, deduping, cleaning, and converting to formats like JSONL that slot into training pipelines.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    tiny-llm

    tiny-llm

    A course of learning LLM inference serving on Apple Silicon

    ...The project demonstrates how to load and run models such as Qwen-style architectures while progressively implementing performance improvements like KV caching, request batching, and optimized attention mechanisms. It also introduces concepts behind modern LLM serving systems that resemble simplified versions of production inference engines such as vLLM.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    Superset LLM

    Superset LLM

    Run an army of Claude Code, Codex, etc. on your machine

    Superset is a development environment and terminal-based platform designed to orchestrate multiple AI coding agents simultaneously within a single workspace. The tool enables developers to run many autonomous coding agents in parallel without the typical overhead of manually managing multiple terminals, repositories, or branches. Each agent task is isolated in its own Git worktree, ensuring that code changes from different agents do not interfere with each other while allowing developers to...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    MLC LLM

    MLC LLM

    Universal LLM Deployment Engine with ML Compilation

    MLC LLM is a machine learning compiler and deployment framework designed to enable efficient execution of large language models across a wide range of hardware platforms. The project focuses on compiling models into optimized runtimes that can run natively on devices such as GPUs, mobile processors, browsers, and edge hardware. By leveraging machine learning compilation techniques, mlc-llm produces high-performance inference engines that maintain consistent APIs across platforms. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    LLM CLI

    LLM CLI

    Access large language models from the command-line

    A CLI utility and Python library for interacting with Large Language Models, both via remote APIs and models that can be installed and run on your own machine.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    Harbor LLM

    Harbor LLM

    Run a full local LLM stack with one command using Docker

    ...It is intended for local development and experimentation rather than production deployment, giving developers a flexible way to explore AI systems, test configurations, and manage complex LLM stacks without manual wiring or setup overhead.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    LLM Scraper

    LLM Scraper

    Extract structured data from webpages using LLM-powered scraping

    LLM Scraper also provides streaming output and code generation capabilities that help developers build reusable scraping workflows.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    local-llm

    local-llm

    Run LLMs locally on Cloud Workstations

    local-llm is a development framework that enables developers to run large language models locally within Google Cloud Workstations or standard environments without requiring GPU hardware. It focuses on making generative AI development more accessible by leveraging quantized models and CPU-based execution, eliminating the dependency on expensive GPU infrastructure.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    TensorRT LLM

    TensorRT LLM

    TensorRT LLM provides users with an easy-to-use Python API

    TensorRT-LLM is an open-source high-performance inference library specifically designed to optimize and accelerate large language model deployment on NVIDIA GPUs. It provides a Python-based API built on top of PyTorch that allows developers to define, customize, and deploy LLMs efficiently across a variety of hardware configurations, from single GPUs to large multi-node clusters.
    Downloads: 1 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    LLM Vision

    LLM Vision

    Visual intelligence for your home.

    LLM Vision is an open-source integration for Home Assistant that adds multimodal large language model capabilities to smart home environments. The project enables Home Assistant to analyze images, video files, and live camera feeds using vision-capable AI models. Instead of relying only on traditional object detection pipelines, it allows users to send prompts about visual content and receive contextual descriptions or answers about what is happening in camera footage.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    SWIFT LLM

    SWIFT LLM

    Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs

    SWIFT LLM is a comprehensive framework developed within the ModelScope ecosystem for training, fine-tuning, evaluating, and deploying large language models and multimodal models. The platform provides a full machine learning pipeline that supports tasks ranging from model pre-training to reinforcement learning alignment techniques. It integrates with popular inference engines such as vLLM and LMDeploy to accelerate deployment and runtime performance.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    LLM TLDR

    LLM TLDR

    95% token savings. 155x faster queries. 16 languages

    LLM TLDR is a tool that leverages large language models (LLMs) to generate concise, coherent summaries (TL;DRs) of long documents, articles, or text files, helping users quickly understand large amounts of content without reading every word. It integrates with LLM APIs to handle input texts of varying lengths and complexity, applying techniques like chunking, context management, and multi-pass summarization to preserve accuracy even when the source is very large.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    RTP-LLM

    RTP-LLM

    Alibaba's high-performance LLM inference engine for diverse apps

    RTP-LLM is an open-source large language model inference acceleration engine developed by Alibaba to provide high-performance serving infrastructure for modern LLM deployments. The system focuses on improving throughput, latency, and resource utilization when running large models in production environments. It achieves this by implementing optimized GPU kernels, batching strategies, and memory management techniques tailored for transformer inference workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    LLM-Pruner

    LLM-Pruner

    On the Structural Pruning of Large Language Models

    LLM-Pruner is an open-source framework designed to compress large language models through structured pruning techniques while maintaining their general capabilities. Large language models often require enormous computational resources, making them expensive to deploy and inefficient for many practical applications. LLM-Pruner addresses this issue by identifying and removing non-essential components within transformer architectures, such as redundant attention heads or feed-forward structures. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    TAME LLM

    TAME LLM

    Traditional Mandarin LLMs for Taiwan

    ...The training pipeline leverages high-performance computing infrastructure and frameworks such as NVIDIA NeMo and Megatron to enable large-scale model training. Taiwan-LLM aims to improve language understanding and generation for Traditional Mandarin users by incorporating region-specific datasets and evaluation benchmarks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    LLM Colosseum

    LLM Colosseum

    Benchmark LLMs by fighting in Street Fighter 3

    LLM-Colosseum is an experimental benchmarking framework designed to evaluate the capabilities of large language models through gameplay interactions rather than traditional text-based benchmarks. The system places language models inside the environment of the classic video game Street Fighter III, where they must interpret the game state and decide which actions to perform during combat.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    LLM Guard

    LLM Guard

    The Security Toolkit for LLM Interactions

    ...The toolkit also helps prevent sensitive information leaks by identifying secrets such as API keys or credentials before they are processed by the model. LLM Guard supports both input and output filtering pipelines, allowing developers to sanitize prompts and validate generated responses in real time. The library integrates easily with existing AI frameworks and can be deployed in production environments to enhance the security posture of LLM-based applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    LLM-Finetuning

    LLM-Finetuning

    LLM Finetuning with peft

    LLM-Finetuning is an open educational repository that provides practical notebooks and tutorials for fine-tuning large language models using modern machine learning frameworks. The project focuses on parameter-efficient fine-tuning methods such as LoRA and QLoRA, which allow large models to be adapted to new tasks without requiring full retraining.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Reader LLM

    Reader LLM

    Convert any URL to an LLM-friendly input with a simple prefix

    Reader LLM is an open-source tool designed to convert web content into formats that are easier for large language models to process. The system works by transforming a webpage into a clean text or Markdown representation that removes unnecessary formatting and highlights the core information within the page. Developers can use a simple URL prefix to retrieve a version of a webpage that has been optimized for machine consumption, making it suitable for use in AI agents or retrieval-augmented generation pipelines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Happy-LLM

    Happy-LLM

    Large Language Model Principles and Practice Tutorial from Scratch

    Happy-LLM is an open-source educational project created by the Datawhale AI community that provides a structured and comprehensive tutorial for understanding and building large language models from scratch. The project guides learners through the entire conceptual and practical pipeline of modern LLM development, starting with foundational natural language processing concepts and gradually progressing to advanced architectures and training techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    self-llm

    self-llm

    Tutorial tailored for Chinese babies on rapid fine-tuning

    self-llm is an open source educational project created by the Datawhale community that serves as a practical guide for deploying, fine-tuning, and using open-source large language models on Linux systems. The repository focuses on helping beginners and developers understand how to run and customize modern LLMs locally rather than relying solely on hosted APIs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    slime LLM

    slime LLM

    slime is an LLM post-training framework for RL Scaling

    slime is an open-source large language model (LLM) post-training framework developed to support reinforcement learning (RL)-based scaling and high-performance training workflows for advanced LLMs, blending training and rollout modules into an extensible system. It offers a flexible architecture that connects high-throughput training (e.g., via Megatron-LM) with a customizable data generation pipeline, enabling researchers and engineers to iterate on new RL training paradigms effectively. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    LLM Council

    LLM Council

    LLM Council works together to answer your hardest questions

    LLM Council is a creative open-source web application by Andrej Karpathy that lets you consult multiple large language models together to answer questions more reliably than querying a single model. Instead of relying on one provider, this application sends your query simultaneously to several LLMs supported via OpenRouter, collects each model’s independent response, and then orchestrates a multi-stage evaluation where the models critique and rank each other’s outputs anonymously. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    LLM Course

    LLM Course

    Course to get into Large Language Models (LLMs)

    LLM Course is a hands-on, notebook-driven path for learning how large language models work in practice, from data curation to training, fine-tuning, evaluating, and deploying. It emphasizes reproducible experiments: each step is demonstrated with runnable code, clear dependencies, and references to commonly used open-source models and libraries.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    LLM Action

    LLM Action

    Technical principles related to large models

    LLM-Action is a knowledge/tutorial/repository that shares principles, techniques, and real-world experience related to large language models (LLMs), focusing on LLM engineering, deployment, optimization, inference, compression, and tooling. It organizes content in domains like training, inference, compression, alignment, evaluation, pipelines, and applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB