Python Neural Network Libraries

View 52 business solutions

Browse free open source Python Neural Network Libraries and projects below. Use the toggles on the left to filter open source Python Neural Network Libraries by OS, license, language, programming language, and project status.

  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    spaCy

    spaCy

    Industrial-strength Natural Language Processing (NLP)

    spaCy is a library built on the very latest research for advanced Natural Language Processing (NLP) in Python and Cython. Since its inception it was designed to be used for real world applications-- for building real products and gathering real insights. It comes with pretrained statistical models and word vectors, convolutional neural network models, easy deep learning integration and so much more. spaCy is the fastest syntactic parser in the world according to independent benchmarks, with an accuracy within 1% of the best available. It's blazing fast, easy to install and comes with a simple and productive API.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 2
    SVoice (Speech Voice Separation)

    SVoice (Speech Voice Separation)

    We provide a PyTorch implementation of the paper Voice Separation

    SVoice is a PyTorch-based implementation of Facebook Research’s study on speaker voice separation as described in the paper “Voice Separation with an Unknown Number of Multiple Speakers.” This project presents a deep learning framework capable of separating mixed audio sequences where several people speak simultaneously, without prior knowledge of how many speakers are present. The model employs gated neural networks with recurrent processing blocks that disentangle voices over multiple computational steps, while maintaining speaker consistency across output channels. Separate models are trained for different speaker counts, and the largest-capacity model dynamically determines the actual number of speakers in a mixture. The repository includes all necessary scripts for training, dataset preparation, distributed training, evaluation, and audio separation.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    FairChem

    FairChem

    FAIR Chemistry's library of machine learning methods for chemistry

    FAIRChem is a unified library for machine learning in chemistry and materials, consolidating data, pretrained models, demos, and application code into a single, versioned toolkit. Version 2 modernizes the stack with a cleaner core package and breaking changes relative to V1, focusing on simpler installs and a stable API surface for production and research. The centerpiece models (e.g., UMA variants) plug directly into the ASE ecosystem via a FAIRChem calculator, so users can run relaxations, molecular dynamics, spin-state energetics, and surface catalysis workflows with the same pretrained network by switching a task flag. Tasks span heterogeneous domains—catalysis (OC20-style), inorganic materials (OMat), molecules (OMol), MOFs (ODAC), and molecular crystals (OMC)—allowing one model family to serve many simulations. The README provides quick paths for pulling models (e.g., via Hugging Face access), then running energy/force predictions on GPU or CPU.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    AIMET

    AIMET

    AIMET is a library that provides advanced quantization and compression

    Qualcomm Innovation Center (QuIC) is at the forefront of enabling low-power inference at the edge through its pioneering model-efficiency research. QuIC has a mission to help migrate the ecosystem toward fixed-point inference. With this goal, QuIC presents the AI Model Efficiency Toolkit (AIMET) - a library that provides advanced quantization and compression techniques for trained neural network models. AIMET enables neural networks to run more efficiently on fixed-point AI hardware accelerators. Quantized inference is significantly faster than floating point inference. For example, models that we’ve run on the Qualcomm® Hexagon™ DSP rather than on the Qualcomm® Kryo™ CPU have resulted in a 5x to 15x speedup. Plus, an 8-bit model also has a 4x smaller memory footprint relative to a 32-bit model. However, often when quantizing a machine learning model (e.g., from 32-bit floating point to an 8-bit fixed point value), the model accuracy is sacrificed.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Fairseq

    Fairseq

    Facebook AI Research Sequence-to-Sequence Toolkit written in Python

    Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks. We provide reference implementations of various sequence modeling papers. Recent work by Microsoft and Google has shown that data parallel training can be made significantly more efficient by sharding the model parameters and optimizer state across data parallel workers. These ideas are encapsulated in the new FullyShardedDataParallel (FSDP) wrapper provided by fairscale. Fairseq can be extended through user-supplied plug-ins. Models define the neural network architecture and encapsulate all of the learnable parameters. Criterions compute the loss function given the model outputs and targets. Tasks store dictionaries and provide helpers for loading/iterating over Datasets, initializing the Model/Criterion and calculating the loss.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    Flax

    Flax

    Flax is a neural network library for JAX

    Flax is a flexible neural-network library for JAX that embraces functional programming while offering ergonomic module abstractions. Its design separates pure computation from state by threading parameter collections and RNGs explicitly, enabling reproducibility, transformation, and easy experimentation with JAX transforms like jit, pmap, and vmap. Modules define parameterized computations, but initialization and application remain side-effect free, which pairs naturally with JAX’s staging and compilation model. Flax emphasizes composability: optimizers, training loops, and checkpointing are provided as examples or utilities rather than monolithic frameworks, encouraging research-friendly customization. The library is widely used in vision, language, and reinforcement learning, often serving as a thin layer atop NumPy-like JAX primitives. Tutorials and examples show patterns for multi-host training, mixed precision, and advanced input pipelines that scale from laptops to TPUs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    I3D models trained on Kinetics

    I3D models trained on Kinetics

    Convolutional neural network model for video classification

    Kinetics-I3D, developed by Google DeepMind, provides trained models and implementation code for the Inflated 3D ConvNet (I3D) architecture introduced in the paper “Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset” (CVPR 2017). The I3D model extends the 2D convolutional structure of Inception-v1 into 3D, allowing it to capture spatial and temporal information from videos for action recognition. This repository includes pretrained I3D models on the Kinetics dataset, with both RGB and optical flow input streams. The models have achieved state-of-the-art results on benchmark datasets such as UCF101 and HMDB51, and also won first place in the CVPR 2017 Charades Challenge. The project provides TensorFlow and Sonnet-based implementations, pretrained checkpoints, and example scripts for evaluating or fine-tuning models. It also offers sample data, including preprocessed video frames and optical flow arrays, to demonstrate how to run inference and visualize outputs.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Neural Network Intelligence

    Neural Network Intelligence

    AutoML toolkit for automate machine learning lifecycle

    Neural Network Intelligence is an open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning. NNI (Neural Network Intelligence) is a lightweight but powerful toolkit to help users automate feature engineering, neural architecture search, hyperparameter tuning and model compression. The tool manages automated machine learning (AutoML) experiments, dispatches and runs experiments' trial jobs generated by tuning algorithms to search the best neural architecture and/or hyper-parameters in different training environments like Local Machine, Remote Servers, OpenPAI, Kubeflow, FrameworkController on K8S (AKS etc.) DLWorkspace (aka. DLTS) AML (Azure Machine Learning) and other cloud options. NNI provides CommandLine Tool as well as an user friendly WebUI to manage training experiements.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Fast Artificial Neural Network Library is a free open source neural network library, which implements multilayer artificial neural networks in C with support for both fully connected and sparsely connected networks. Cross-platform execution in both fixed and floating point are supported. It includes a framework for easy handling of training data sets. It is easy to use, versatile, well documented, and fast. Bindings to more than 15 programming languages are available. An easy to read introduction article and a reference manual accompanies the library with examples and recommendations on how to use the library. Several graphical user interfaces are also available for the library.
    Downloads: 11 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    CNN for Image Retrieval
    cnn-for-image-retrieval is a research-oriented project that demonstrates the use of convolutional neural networks (CNNs) for image retrieval tasks. The repository provides implementations of CNN-based methods to extract feature representations from images and use them for similarity-based retrieval. It focuses on applying deep learning techniques to improve upon traditional handcrafted descriptors by learning features directly from data. The code includes training and evaluation scripts that can be adapted for custom datasets, making it useful for experimenting with retrieval systems in computer vision. By leveraging CNN architectures, the project showcases how learned embeddings can capture semantic similarity across varied images. This resource serves as both an educational reference and a foundation for further exploration in image retrieval research.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Opacus

    Opacus

    Training PyTorch models with differential privacy

    Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment. Vectorized per-sample gradient computation that is 10x faster than micro batching. Supports most types of PyTorch models and can be used with minimal modification to the original neural network. Open source, modular API for differential privacy research. Everyone is welcome to contribute. ML practitioners will find this to be a gentle introduction to training a model with differential privacy as it requires minimal code changes. Differential Privacy researchers will find this easy to experiment and tinker with, allowing them to focus on what matters.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    PyG

    PyG

    Graph Neural Network Library for PyTorch

    PyG (PyTorch Geometric) is a library built upon PyTorch to easily write and train Graph Neural Networks (GNNs) for a wide range of applications related to structured data. It consists of various methods for deep learning on graphs and other irregular structures, also known as geometric deep learning, from a variety of published papers. In addition, it consists of easy-to-use mini-batch loaders for operating on many small and single giant graphs, multi GPU-support, DataPipe support, distributed graph learning via Quiver, a large number of common benchmark datasets (based on simple interfaces to create your own), the GraphGym experiment manager, and helpful transforms, both for learning on arbitrary graphs as well as on 3D meshes or point clouds. All it takes is 10-20 lines of code to get started with training a GNN model (see the next section for a quick tour).
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    SFD

    SFD

    S³FD: Single Shot Scale-invariant Face Detector, ICCV, 2017

    S³FD (Single Shot Scale-invariant Face Detector) is a real-time face detection framework designed to handle faces of various sizes with high accuracy using a single deep neural network. Developed by Shifeng Zhang, S³FD introduces a scale-compensation anchor matching strategy and enhanced detection architecture that makes it especially effective for detecting small faces—a long-standing challenge in face detection research. The project builds upon the SSD framework in Caffe, with modifications tailored for face detection tasks. It includes training scripts, evaluation code, and pre-trained models that achieve strong results on popular benchmarks such as AFW, PASCAL Face, FDDB, and WIDER FACE. The framework is optimized for speed and accuracy, making it suitable for both academic research and practical applications in computer vision.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    captcha_break

    captcha_break

    Identification codes

    This project will use Keras to build a deep convolutional neural network to identify the captcha verification code. It is recommended to use a graphics card to run the project. The following visualization codes are jupyter notebookall done in . If you want to write a python script, you can run it normally with a little modification. Of course, you can also remove these visualization codes. captcha is a library written in python to generate verification codes. It supports image verification codes and voice verification codes. We use its function of generating image verification codes. First, we set our verification code format to numbers and capital letters, and generate a string of verification codes. It is well known that tensorflow occupies all video memory by default, which is not conducive to us conducting multiple experiments at the same time, so we can use the following code when tensorflow uses the video memory it needs instead of directly occupying all video memory.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    fastNLP

    fastNLP

    fastNLP: A Modularized and Extensible NLP Framework

    fastNLP is a lightweight framework for natural language processing (NLP), the goal is to quickly implement NLP tasks and build complex models. A unified Tabular data container simplifies the data preprocessing process. Built-in Loader and Pipe for multiple datasets, eliminating the need for preprocessing code. Various convenient NLP tools, such as Embedding loading (including ELMo and BERT), intermediate data cache, etc.. Provide a variety of neural network components and recurrence models (covering tasks such as Chinese word segmentation, named entity recognition, syntactic analysis, text classification, text matching, metaphor resolution, summarization, etc.). Trainer provides a variety of built-in Callback functions to facilitate experiment recording, exception capture, etc. Automatic download of some datasets and pre-trained models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Feed-forward neural network for python
    ffnet is a fast and easy-to-use feed-forward neural network training solution for python. Many nice features are implemented: arbitrary network connectivity, automatic data normalization, very efficient training tools, network export to fortran code. Now ffnet has also a GUI called ffnetui.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Neural Libs

    Neural Libs

    Neural network library for developers

    This project includes the implementation of a neural network MLP, RBF, SOM and Hopfield networks in several popular programming languages. The project also includes examples of the use of neural networks as function approximation and time series prediction. Includes a special program makes it easy to test neural network based on training data and the optimization of the network.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    A neural net module written in python. The aim of the project is to provide a large set of neural network types accessed by an API that is easy to use and powerful.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    AdaNet

    AdaNet

    Fast and flexible AutoML with learning guarantees

    AdaNet is a TensorFlow framework for fast and flexible AutoML with learning guarantees. AdaNet is a lightweight TensorFlow-based framework for automatically learning high-quality models with minimal expert intervention. AdaNet builds on recent AutoML efforts to be fast and flexible while providing learning guarantees. Importantly, AdaNet provides a general framework for not only learning a neural network architecture but also for learning to the ensemble to obtain even better models. At each iteration, it measures the ensemble loss for each candidate, and selects the best one to move onto the next iteration. Adaptive neural architecture search and ensemble learning in a single train call. Regression, binary and multi-class classification, and multi-head task support. A tf.estimator.Estimator API for training, evaluation, prediction, and serving models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Alpa

    Alpa

    Training and serving large-scale neural networks

    Alpa is a system for training and serving large-scale neural networks. Scaling neural networks to hundreds of billions of parameters has enabled dramatic breakthroughs such as GPT-3, but training and serving these large-scale neural networks require complicated distributed system techniques. Alpa aims to automate large-scale distributed training and serving with just a few lines of code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Compare GAN

    Compare GAN

    Compare GAN code

    compare_gan is a research codebase that standardizes how Generative Adversarial Networks are trained and evaluated so results are comparable across papers and datasets. It offers reference implementations for popular GAN architectures and losses, plus a consistent training harness to remove confounding differences in optimization or preprocessing. The library’s evaluation suite includes widely used metrics and diagnostics that quantify sample quality, diversity, and mode coverage. With configuration-driven experiments, you can sweep hyperparameters, run ablations, and log results at scale. The goal is to turn GAN experimentation into a disciplined, repeatable process rather than a patchwork of scripts. It also provides baselines strong enough to serve as starting points for new ideas without re-implementing the world.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CoreNet

    CoreNet

    CoreNet: A library for training deep neural networks

    CoreNet is Apple’s internal deep learning framework for distributed neural network training, designed for high scalability, low-latency communication, and strong hardware efficiency. It focuses on enabling large-scale model training across clusters of GPUs and accelerators by optimizing data flow and parallelism strategies. CoreNet provides abstractions for data, tensor, and pipeline parallelism, allowing models to scale without code duplication or heavy manual configuration. Its distributed runtime manages synchronization, load balancing, and mixed-precision computation to maximize throughput while minimizing communication bottlenecks. CoreNet integrates tightly with Apple’s proprietary ML stack and hardware, serving as the foundation for research in computer vision, language models, and multimodal systems within Apple AI. The framework includes monitoring tools, fault tolerance mechanisms, and efficient checkpointing for massive training runs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    DeepXDE

    DeepXDE

    A library for scientific machine learning & physics-informed learning

    DeepXDE is a library for scientific machine learning and physics-informed learning. DeepXDE includes the following algorithms. Physics-informed neural network (PINN). Solving different problems. Solving forward/inverse ordinary/partial differential equations (ODEs/PDEs) [SIAM Rev.] Solving forward/inverse integro-differential equations (IDEs) [SIAM Rev.] fPINN: solving forward/inverse fractional PDEs (fPDEs) [SIAM J. Sci. Comput.] NN-arbitrary polynomial chaos (NN-aPC): solving forward/inverse stochastic PDEs (sPDEs) [J. Comput. Phys.] PINN with hard constraints (hPINN): solving inverse design/topology optimization [SIAM J. Sci. Comput.] Residual-based adaptive sampling [SIAM Rev., arXiv] Gradient-enhanced PINN (gPINN) [Comput. Methods Appl. Mech. Eng.] PINN with multi-scale Fourier features [Comput. Methods Appl. Mech. Eng.]
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Differentiable Neural Computer

    Differentiable Neural Computer

    A TensorFlow implementation of the Differentiable Neural Computer

    The Differentiable Neural Computer (DNC), developed by Google DeepMind, is a neural network architecture augmented with dynamic external memory, enabling it to learn algorithms and solve complex reasoning tasks. Published in Nature in 2016 under the paper “Hybrid computing using a neural network with dynamic external memory,” the DNC combines the pattern recognition power of neural networks with a memory module that can be written to and read from in a differentiable way. This allows the model to learn how to store and retrieve information across long time horizons, much like a traditional computer. The architecture consists of modular components including an access module for managing memory operations, a controller (often an LSTM or feedforward network) for issuing read/write commands, and submodules for temporal linkage and memory allocation tracking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Forecasting Best Practices

    Forecasting Best Practices

    Time Series Forecasting Best Practices & Examples

    Time series forecasting is one of the most important topics in data science. Almost every business needs to predict the future in order to make better decisions and allocate resources more effectively. This repository provides examples and best practice guidelines for building forecasting solutions. The goal of this repository is to build a comprehensive set of tools and examples that leverage recent advances in forecasting algorithms to build solutions and operationalize them. Rather than creating implementations from scratch, we draw from existing state-of-the-art libraries and build additional utilities around processing and featuring the data, optimizing and evaluating models, and scaling up to the cloud. The examples and best practices are provided as Python Jupyter notebooks and R markdown files and a library of utility functions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB