OPT

OPT

Meta
PanGu-α

PanGu-α

Huawei
+
+

Related Products

  • Vertex AI
    827 Ratings
    Visit Website
  • Google AI Studio
    11 Ratings
    Visit Website
  • LM-Kit.NET
    24 Ratings
    Visit Website
  • RaimaDB
    10 Ratings
    Visit Website
  • ClickLearn
    66 Ratings
    Visit Website
  • Dragonfly
    16 Ratings
    Visit Website
  • B2i
    2 Ratings
    Visit Website
  • Ango Hub
    15 Ratings
    Visit Website
  • RunPod
    205 Ratings
    Visit Website
  • Qloo
    23 Ratings
    Visit Website

About

Large language models, which are often trained for hundreds of thousands of compute days, have shown remarkable capabilities for zero- and few-shot learning. Given their computational cost, these models are difficult to replicate without significant capital. For the few that are available through APIs, no access is granted to the full model weights, making them difficult to study. We present Open Pre-trained Transformers (OPT), a suite of decoder-only pre-trained transformers ranging from 125M to 175B parameters, which we aim to fully and responsibly share with interested researchers. We show that OPT-175B is comparable to GPT-3, while requiring only 1/7th the carbon footprint to develop. We are also releasing our logbook detailing the infrastructure challenges we faced, along with code for experimenting with all of the released models.

About

PanGu-α is developed under the MindSpore and trained on a cluster of 2048 Ascend 910 AI processors. The training parallelism strategy is implemented based on MindSpore Auto-parallel, which composes five parallelism dimensions to scale the training task to 2048 processors efficiently, including data parallelism, op-level model parallelism, pipeline model parallelism, optimizer model parallelism and rematerialization. To enhance the generalization ability of PanGu-α, we collect 1.1TB high-quality Chinese data from a wide range of domains to pretrain the model. We empirically test the generation ability of PanGu-α in various scenarios including text summarization, question answering, dialogue generation, etc. Moreover, we investigate the effect of model scales on the few-shot performances across a broad range of Chinese NLP tasks. The experimental results demonstrate the superior capabilities of PanGu-α in performing various tasks under few-shot or zero-shot settings.

Platforms Supported

Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook

Platforms Supported

Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook

Audience

AI developers interested in a large language model

Audience

AI developers interested in a powerful large language model

Support

Phone Support
24/7 Live Support
Online

Support

Phone Support
24/7 Live Support
Online

API

Offers API

API

Offers API

Screenshots and Videos

No images available

Screenshots and Videos

No images available

Pricing

No information available.
Free Version
Free Trial

Pricing

No information available.
Free Version
Free Trial

Reviews/Ratings

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Reviews/Ratings

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Training

Documentation
Webinars
Live Online
In Person

Training

Documentation
Webinars
Live Online
In Person

Company Information

Meta
Founded: 2004
United States
www.meta.com

Company Information

Huawei
Founded: 1987
China
arxiv.org/abs/2104.12369

Alternatives

Alternatives

PanGu-Σ

PanGu-Σ

Huawei
T5

T5

Google
OPT

OPT

Meta
CodeQwen

CodeQwen

Alibaba
PanGu-α

PanGu-α

Huawei
GPT-NeoX

GPT-NeoX

EleutherAI
Llama 2

Llama 2

Meta

Categories

Categories

Integrations

No info available.

Integrations

No info available.
Claim OPT and update features and information
Claim OPT and update features and information
Claim PanGu-α and update features and information
Claim PanGu-α and update features and information