Qwen2.5-Max

Qwen2.5-Max

Alibaba
+
+

Related Products

  • Vertex AI
    783 Ratings
    Visit Website
  • Google AI Studio
    11 Ratings
    Visit Website
  • LM-Kit.NET
    23 Ratings
    Visit Website
  • Ango Hub
    15 Ratings
    Visit Website
  • DataDome
    223 Ratings
    Visit Website
  • AthenaHQ
    18 Ratings
    Visit Website
  • Evertune
    1 Rating
    Visit Website
  • ONLYOFFICE Docs
    706 Ratings
    Visit Website
  • Applauz Recognition
    125 Ratings
    Visit Website
  • DAT
    321 Ratings
    Visit Website

About

Qwen2.5-Max is a large-scale Mixture-of-Experts (MoE) model developed by the Qwen team, pretrained on over 20 trillion tokens and further refined through Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF). In evaluations, it outperforms models like DeepSeek V3 in benchmarks such as Arena-Hard, LiveBench, LiveCodeBench, and GPQA-Diamond, while also demonstrating competitive results in other assessments, including MMLU-Pro. Qwen2.5-Max is accessible via API through Alibaba Cloud and can be explored interactively on Qwen Chat.

About

The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. With some proper optimization, we can achieve this within a span of "just" 90 days using 16 A100-40G GPUs. We adopted exactly the same architecture and tokenizer as Llama 2. This means TinyLlama can be plugged and played in many open-source projects built upon Llama. Besides, TinyLlama is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint.

Platforms Supported

Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook

Platforms Supported

Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook

Audience

AI researchers, developers, and enterprises seeking a high-performance Mixture-of-Experts model for advanced reasoning, coding, and language tasks, accessible via API and interactive chat

Audience

Developers interested in a small language model

Support

Phone Support
24/7 Live Support
Online

Support

Phone Support
24/7 Live Support
Online

API

Offers API

API

Offers API

Screenshots and Videos

Screenshots and Videos

No images available

Pricing

Free
Free Version
Free Trial

Pricing

Free
Free Version
Free Trial

Reviews/Ratings

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Reviews/Ratings

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Training

Documentation
Webinars
Live Online
In Person

Training

Documentation
Webinars
Live Online
In Person

Company Information

Alibaba
Founded: 1999
China
qwenlm.github.io/blog/qwen2.5-max/

Company Information

TinyLlama
github.com/jzhang38/TinyLlama

Alternatives

DeepSeek R2

DeepSeek R2

DeepSeek

Alternatives

Llama 2

Llama 2

Meta
ERNIE 4.5

ERNIE 4.5

Baidu
ERNIE X1

ERNIE X1

Baidu
Qwen2

Qwen2

Alibaba
Qwen-7B

Qwen-7B

Alibaba
Llama

Llama

Meta

Categories

Categories

Integrations

Alibaba Cloud
Hugging Face
ModelScope
Qwen Chat
RunPod

Integrations

Alibaba Cloud
Hugging Face
ModelScope
Qwen Chat
RunPod
Claim Qwen2.5-Max and update features and information
Claim Qwen2.5-Max and update features and information
Claim TinyLlama and update features and information
Claim TinyLlama and update features and information