Open Source Mac Reinforcement Learning Frameworks

Reinforcement Learning Frameworks for Mac

View 1 business solution

Browse free open source Reinforcement Learning Frameworks and projects for Mac below. Use the toggles on the left to filter open source Reinforcement Learning Frameworks by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 99.99% Uptime for MySQL and PostgreSQL on Google Cloud Icon
    99.99% Uptime for MySQL and PostgreSQL on Google Cloud

    Enterprise Plus edition delivers sub-second maintenance downtime and 2x read/write performance. Built for critical apps.

    Cloud SQL Enterprise Plus gives you a 99.99% availability SLA with near-zero downtime maintenance—typically under 10 seconds. Get 2x better read/write performance, intelligent data caching, and 35 days of point-in-time recovery. Supports MySQL, PostgreSQL, and SQL Server with built-in vector search for gen AI apps. New customers get $300 in free credit.
    Try Cloud SQL Free
  • 1
    DeepSeek R1

    DeepSeek R1

    Open-source, high-performance AI model with advanced reasoning

    DeepSeek-R1 is an open-source large language model developed by DeepSeek, designed to excel in complex reasoning tasks across domains such as mathematics, coding, and language. DeepSeek R1 offers unrestricted access for both commercial and academic use. The model employs a Mixture of Experts (MoE) architecture, comprising 671 billion total parameters with 37 billion active parameters per token, and supports a context length of up to 128,000 tokens. DeepSeek-R1's training regimen uniquely integrates large-scale reinforcement learning (RL) without relying on supervised fine-tuning, enabling the model to develop advanced reasoning capabilities. This approach has resulted in performance comparable to leading models like OpenAI's o1, while maintaining cost-efficiency. To further support the research community, DeepSeek has released distilled versions of the model based on architectures such as LLaMA and Qwen.
    Downloads: 60 This Week
    Last Update:
    See Project
  • 2
    DeepSeek-V3

    DeepSeek-V3

    Powerful AI language model (MoE) optimized for efficiency/performance

    DeepSeek-V3 is a robust Mixture-of-Experts (MoE) language model developed by DeepSeek, featuring a total of 671 billion parameters, with 37 billion activated per token. It employs Multi-head Latent Attention (MLA) and the DeepSeekMoE architecture to enhance computational efficiency. The model introduces an auxiliary-loss-free load balancing strategy and a multi-token prediction training objective to boost performance. Trained on 14.8 trillion diverse, high-quality tokens, DeepSeek-V3 underwent supervised fine-tuning and reinforcement learning to fully realize its capabilities. Evaluations indicate that it outperforms other open-source models and rivals leading closed-source models, achieving this with a training duration of 55 days on 2,048 Nvidia H800 GPUs, costing approximately $5.58 million.
    Downloads: 59 This Week
    Last Update:
    See Project
  • 3
    AirSim

    AirSim

    A simulator for drones, cars and more, built on Unreal Engine

    AirSim is an open-source, cross platform simulator for drones, cars and more vehicles, built on Unreal Engine with an experimental Unity release in the works. It supports software-in-the-loop simulation with popular flight controllers such as PX4 & ArduPilot and hardware-in-loop with PX4 for physically and visually realistic simulations. It is developed as an Unreal plugin that can simply be dropped into any Unreal environment. AirSim's development is oriented towards the goal of creating a platform for AI research to experiment with deep learning, computer vision and reinforcement learning algorithms for autonomous vehicles. For this purpose, AirSim also exposes APIs to retrieve data and control vehicles in a platform independent way. AirSim is fully enabled for multiple vehicles. This capability allows you to create multiple vehicles easily and use APIs to control them.
    Downloads: 57 This Week
    Last Update:
    See Project
  • 4
    Bullet Physics SDK

    Bullet Physics SDK

    Real-time collision detection and multi-physics simulation for VR

    This is the official C++ source code repository of the Bullet Physics SDK: real-time collision detection and multi-physics simulation for VR, games, visual effects, robotics, machine learning etc. We are developing a new differentiable simulator for robotics learning, called Tiny Differentiable Simulator, or TDS. The simulator allows for hybrid simulation with neural networks. It allows different automatic differentiation backends, for forward and reverse mode gradients. TDS can be trained using Deep Reinforcement Learning, or using Gradient based optimization (for example LFBGS). In addition, the simulator can be entirely run on CUDA for fast rollouts, in combination with Augmented Random Search. This allows for 1 million simulation steps per second. It is highly recommended to use PyBullet Python bindings for improved support for robotics, reinforcement learning and VR. Use pip install pybullet and checkout the PyBullet Quickstart Guide.
    Downloads: 15 This Week
    Last Update:
    See Project
  • $300 in Free Credit Across 150+ Cloud Services Icon
    $300 in Free Credit Across 150+ Cloud Services

    VMs, containers, AI, databases, storage | build anything. No commitment to start.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale with Google Cloud.
    Start Building Free
  • 5
    Vowpal Wabbit

    Vowpal Wabbit

    Machine learning system which pushes the frontier of machine learning

    Vowpal Wabbit is a machine learning system that pushes the frontier of machine learning with techniques such as online, hashing, allreduce, reductions, learning2search, active, and interactive learning. There is a specific focus on reinforcement learning with several contextual bandit algorithms implemented and the online nature lending to the problem well. Vowpal Wabbit is a destination for implementing and maturing state-of-the-art algorithms with performance in mind. The input format for the learning algorithm is substantially more flexible than might be expected. Examples can have features consisting of free-form text, which is interpreted in a bag-of-words way. There can even be multiple sets of free-form text in different namespaces. Similar to the few other online algorithm implementations out there. There are several optimization algorithms available with the baseline being sparse gradient descent (GD) on a loss function.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    Agent S

    Agent S

    Agent S: an open agentic framework that uses computers like a human

    Agent S is an open-source agentic framework designed to enable autonomous computer use through an Agent-Computer Interface (ACI). Built to operate graphical user interfaces like a human, it allows AI agents to perceive screens, reason about tasks, and execute actions across macOS, Windows, and Linux systems. The latest version, Agent S3, surpasses human-level performance on the OSWorld benchmark, demonstrating state-of-the-art results in complex multi-step computer tasks. Agent S combines powerful foundation models (such as GPT-5) with grounding models like UI-TARS to translate visual inputs into precise executable actions. It supports flexible deployment via CLI, SDK, or cloud, and integrates with multiple model providers including OpenAI, Anthropic, Gemini, Azure, and Hugging Face endpoints. With optional local code execution, reflection mechanisms, and compositional planning, Agent S provides a scalable and research-driven framework for building advanced computer-use agents.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    LightZero

    LightZero

    [NeurIPS 2023 Spotlight] LightZero

    LightZero is an efficient, scalable, and open-source framework implementing MuZero, a powerful model-based reinforcement learning algorithm that learns to predict rewards and transitions without explicit environment models. Developed by OpenDILab, LightZero focuses on providing a highly optimized and user-friendly platform for both academic research and industrial applications of MuZero and similar algorithms.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    Tensorforce

    Tensorforce

    A TensorFlow library for applied reinforcement learning

    Tensorforce is an open-source deep reinforcement learning framework built on TensorFlow, emphasizing modularized design and straightforward usability for applied research and practice.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9
    Pwnagotchi

    Pwnagotchi

    Deep Reinforcement learning instrumenting bettercap for WiFi pwning

    Pwnagotchi is an A2C-based “AI” powered by bettercap and running on a Raspberry Pi Zero W that learns from its surrounding WiFi environment in order to maximize the crackable WPA key material it captures (either through passive sniffing or by performing deauthentication and association attacks). This material is collected on disk as PCAP files containing any form of handshake supported by hashcat, including full and half WPA handshakes as well as PMKIDs. Instead of merely playing Super Mario or Atari games like most reinforcement learning based “AI” (yawn), Pwnagotchi tunes its own parameters over time to get better at pwning WiFi things in the real world environments you expose it to. To give hackers an excuse to learn about reinforcement learning and WiFi networking, and have a reason to get out for more walks.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Deploy Apps in Seconds with Cloud Run Icon
    Deploy Apps in Seconds with Cloud Run

    Host and run your applications without the need to manage infrastructure. Scales up from and down to zero automatically.

    Cloud Run is the fastest way to deploy containerized apps. Push your code in Go, Python, Node.js, Java, or any language and Cloud Run builds and deploys it automatically. Get fast autoscaling, pay only when your code runs, and skip the infrastructure headaches. Two million requests free per month. And new customers get $300 in free credit.
    Try Cloud Run Free
  • 10
    RL Games

    RL Games

    RL implementations

    rl_games is a high-performance reinforcement learning framework optimized for GPU-based training, particularly in environments like robotics and continuous control tasks. It supports advanced algorithms and is built with PyTorch.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    SLM Lab

    SLM Lab

    Modular Deep Reinforcement Learning framework in PyTorch

    SLM Lab is a modular and extensible deep reinforcement learning framework designed for research and practical applications. It provides implementations of various state-of-the-art RL algorithms and emphasizes reproducibility, scalability, and detailed experiment tracking. SLM Lab is structured around a flexible experiment management system, allowing users to define, run, and analyze RL experiments efficiently.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    TorchRL

    TorchRL

    A modular, primitive-first, python-first PyTorch library

    TorchRL is an open-source Reinforcement Learning (RL) library for PyTorch. TorchRL provides PyTorch and python-first, low and high-level abstractions for RL that are intended to be efficient, modular, documented, and properly tested. The code is aimed at supporting research in RL. Most of it is written in Python in a highly modular way, such that researchers can easily swap components, transform them, or write new ones with little effort.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    ViZDoom

    ViZDoom

    Doom-based AI research platform for reinforcement learning

    ViZDoom allows developing AI bots that play Doom using only the visual information (the screen buffer). It is primarily intended for research in machine visual learning, and deep reinforcement learning, in particular. ViZDoom is based on ZDOOM, the most popular modern source-port of DOOM. This means compatibility with a huge range of tools and resources that can be used to create custom scenarios, availability of detailed documentation of the engine and tools and support of Doom community. Async and sync single-player and multi-player modes. Fast (up to 7000 fps in sync mode, single-threaded). Lightweight (few MBs). Customizable resolution and rendering parameters. Access to the depth buffer (3D vision). Automatic labeling of game objects visible in the frame. Access to the list of actors/objects and map geometry.ViZDoom API is reinforcement learning friendly (suitable also for learning from demonstration, apprenticeship learning or apprenticeship via inverse reinforcement learning.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    DI-engine

    DI-engine

    OpenDILab Decision AI Engine

    DI-engine is a unified reinforcement learning (RL) platform for reproducible and scalable RL research. It offers modular pipelines for various RL algorithms, with an emphasis on production-level training and evaluation.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    EvoTorch

    EvoTorch

    Advanced evolutionary computation library built on top of PyTorch

    EvoTorch is an evolutionary optimization framework built on top of PyTorch, developed by NNAISENSE. It is designed for large-scale optimization problems, particularly those that require evolutionary algorithms rather than gradient-based methods.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    H2O LLM Studio

    H2O LLM Studio

    Framework and no-code GUI for fine-tuning LLMs

    Welcome to H2O LLM Studio, a framework and no-code GUI designed for fine-tuning state-of-the-art large language models (LLMs). You can also use H2O LLM Studio with the command line interface (CLI) and specify the configuration file that contains all the experiment parameters. To finetune using H2O LLM Studio with CLI, activate the pipenv environment by running make shell. With H2O LLM Studio, training your large language model is easy and intuitive. First, upload your dataset and then start training your model. Start by creating an experiment. You can then monitor and manage your experiment, compare experiments, or push the model to Hugging Face to share it with the community.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    PyBoy

    PyBoy

    Game Boy emulator written in Python

    It is highly recommended to read the report to get a light introduction to Game Boy emulation. But do be aware, that the Python implementation has changed a lot. The report is relevant, even though you want to contribute to another emulator or create your own. If you are looking to make a bot or AI, you can find all the external components in the PyBoy Documentation. There is also a short example on our Wiki page Scripts, AI and Bots as well as in the examples directory. If more features are needed, or if you find a bug, don't hesitate to make an issue here on GitHub, or write on our Discord channel. If you need more details, or if you need to compile from source, check out the detailed installation instructions. We support: macOS, Raspberry Pi (Raspbian), Linux (Ubuntu), and Windows 10.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    Trax

    Trax

    Deep learning with clear code and speed

    Trax is an end-to-end library for deep learning that focuses on clear code and speed. It is actively used and maintained in the Google Brain team. Run a pre-trained Transformer, create a translator in a few lines of code. Features and resources, API docs, where to talk to us, how to open an issue and more. Walkthrough, how Trax works, how to make new models and train on your own data. Trax includes basic models (like ResNet, LSTM, Transformer) and RL algorithms (like REINFORCE, A2C, PPO). It is also actively used for research and includes new models like the Reformer and new RL algorithms like AWR. Trax has bindings to a large number of deep learning datasets, including Tensor2Tensor and TensorFlow datasets. You can use Trax either as a library from your own python scripts and notebooks or as a binary from the shell, which can be more convenient for training large models. It runs without any changes on CPUs, GPUs and TPUs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    AnyTrading

    AnyTrading

    The most simple, flexible, and comprehensive OpenAI Gym trading

    gym-anytrading is an OpenAI Gym-compatible environment designed for developing and testing reinforcement learning algorithms on trading strategies. It simulates trading environments for financial markets, including stocks and forex.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    CCZero (中国象棋Zero)

    CCZero (中国象棋Zero)

    Implement AlphaZero/AlphaGo Zero methods on Chinese chess

    ChineseChess-AlphaZero is a project that implements the AlphaZero algorithm for the game of Chinese Chess (Xiangqi). It adapts DeepMind’s AlphaZero method—combining neural networks and Monte Carlo Tree Search (MCTS)—to learn and play Chinese Chess without prior human data. The system includes self-play, training, and evaluation pipelines tailored to Xiangqi's unique game mechanics.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    ConvNetJS

    ConvNetJS

    Deep learning in Javascript to train convolutional neural networks

    ConvNetJS is a Javascript library for training Deep Learning models (Neural Networks) entirely in your browser. Open a tab and you're training. No software requirements, no compilers, no installations, no GPUs, no sweat. ConvNetJS is an implementation of Neural networks, together with nice browser-based demos. It currently supports common Neural Network modules (fully connected layers, non-linearities), classification (SVM/Softmax) and Regression (L2) cost functions, ability to specify and train Convolutional Networks that process images, and experimental Reinforcement Learning modules, based on Deep Q Learning. The library allows you to formulate and solve Neural Networks in Javascript. If you would like to add features to the library, you will have to change the code in src/ and then compile the library into the build/ directory. The compilation script simply concatenates files in src/ and then minifies the result.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    EasyRL

    EasyRL

    Reinforcement learning (RL) tutorial series

    easy-rl is a beginner-friendly reinforcement learning (RL) tutorial series and framework developed by Datawhale China. It provides educational resources and implementations of various RL algorithms to help new researchers and practitioners learn RL concepts.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Google Research Football

    Google Research Football

    Check out the new game server

    Google Research Football is a reinforcement learning environment simulating soccer matches. It focuses on learning complex behaviors such as team collaboration and strategy formation in competitive settings.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Machine Learning PyTorch Scikit-Learn

    Machine Learning PyTorch Scikit-Learn

    Code Repository for Machine Learning with PyTorch and Scikit-Learn

    Initially, this project started as the 4th edition of Python Machine Learning. However, after putting so much passion and hard work into the changes and new topics, we thought it deserved a new title. So, what’s new? There are many contents and additions, including the switch from TensorFlow to PyTorch, new chapters on graph neural networks and transformers, a new section on gradient boosting, and many more that I will detail in a separate blog post. For those who are interested in knowing what this book covers in general, I’d describe it as a comprehensive resource on the fundamental concepts of machine learning and deep learning. The first half of the book introduces readers to machine learning using scikit-learn, the defacto approach for working with tabular datasets. Then, the second half of this book focuses on deep learning, including applications to natural language processing and computer vision.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    PaLM + RLHF - Pytorch

    PaLM + RLHF - Pytorch

    Implementation of RLHF (Reinforcement Learning with Human Feedback)

    PaLM-rlhf-pytorch is a PyTorch implementation of Pathways Language Model (PaLM) with Reinforcement Learning from Human Feedback (RLHF). It is designed for fine-tuning large-scale language models with human preference alignment, similar to OpenAI’s approach for training models like ChatGPT.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB