Open Source C Generative Adversarial Networks (GAN)

C Generative Adversarial Networks (GAN)

View 12360 business solutions

Browse free open source C Generative Adversarial Networks (GAN) and projects below. Use the toggles on the left to filter open source C Generative Adversarial Networks (GAN) by OS, license, language, programming language, and project status.

  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • Cloud tools for web scraping and data extraction Icon
    Cloud tools for web scraping and data extraction

    Deploy pre-built tools that crawl websites, extract structured data, and feed your applications. Reliable web data without maintaining scrapers.

    Automate web data collection with cloud tools that handle anti-bot measures, browser rendering, and data transformation out of the box. Extract content from any website, push to vector databases for RAG workflows, or pipe directly into your apps via API. Schedule runs, set up webhooks, and connect to your existing stack. Free tier available, then scale as you need to.
    Explore 10,000+ tools
  • 1
    Grenade

    Grenade

    Deep Learning in Haskell

    Grenade is a composable, dependently typed, practical, and fast recurrent neural network library for concise and precise specifications of complex networks in Haskell. Because the types are so rich, there's no specific term level code required to construct this network; although it is of course possible and easy to construct and deconstruct the networks and layers explicitly oneself. Networks in Grenade can be thought of as a heterogeneous list of layers, where their type includes not only the layers of the network but also the shapes of data that are passed between the layers. To perform back propagation, one can call the eponymous function which takes a network, appropriate input, and target data, and returns the back propagated gradients for the network. The shapes of the gradients are appropriate for each layer and may be trivial for layers like Relu which have no learnable parameters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next