Related Products
|
||||||
About
Axolotl is an open source tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures. It enables users to train models, supporting methods like full fine-tuning, LoRA, QLoRA, ReLoRA, and GPTQ. Users can customize configurations using simple YAML files or command-line interface overrides, and load different dataset formats, including custom or pre-tokenized datasets. Axolotl integrates with technologies like xFormers, Flash Attention, Liger kernel, RoPE scaling, and multipacking, and works with single or multiple GPUs via Fully Sharded Data Parallel (FSDP) or DeepSpeed. It can be run locally or on the cloud using Docker and supports logging results and checkpoints to several platforms. It is designed to make fine-tuning AI models friendly, fast, and fun, without sacrificing functionality or scale.
|
About
Unsloth is an open source platform designed to accelerate and optimize the fine-tuning and training of Large Language Models (LLMs). It enables users to train custom models, such as ChatGPT, in just 24 hours instead of the typical 30 days, achieving speeds up to 30 times faster than Flash Attention 2 (FA2) while using 90% less memory. Unsloth supports both LoRA and QLoRA fine-tuning techniques, allowing for efficient customization of models like Mistral, Gemma, and Llama versions 1, 2, and 3. Unsloth's efficiency stems from manually deriving computationally intensive mathematical steps and handwriting GPU kernels, resulting in significant performance gains without requiring hardware modifications. Unsloth delivers a 10x speed increase on a single GPU and up to 32x on multi-GPU systems compared to FA2, with compatibility across NVIDIA GPUs from Tesla T4 to H100, and portability to AMD and Intel GPUs.
|
|||||
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
Platforms Supported
Windows
Mac
Linux
Cloud
On-Premises
iPhone
iPad
Android
Chromebook
|
|||||
Audience
Data scientists seeking a solution to fine-tune large language models across various architectures
|
Audience
AI researchers and developers in need of a solution to optimize the fine-tuning and training of LLMs with reduced resource consumption
|
|||||
Support
Phone Support
24/7 Live Support
Online
|
Support
Phone Support
24/7 Live Support
Online
|
|||||
API
Offers API
|
API
Offers API
|
|||||
Screenshots and Videos |
Screenshots and Videos |
|||||
Pricing
Free
Free Version
Free Trial
|
Pricing
Free
Free Version
Free Trial
|
|||||
Reviews/
|
Reviews/
|
|||||
Training
Documentation
Webinars
Live Online
In Person
|
Training
Documentation
Webinars
Live Online
In Person
|
|||||
Company InformationAxolotl
United States
axolotl.ai/
|
Company InformationUnsloth
Founded: 2023
United States
unsloth.ai/
|
|||||
Alternatives |
Alternatives |
|||||
|
|
||||||
|
|
||||||
Categories |
Categories |
|||||
Integrations
Gemma
Llama
Mistral AI
ChatGPT
Comet
DeepSpeed
Devstral
Falcon
GPT-J
Google Colab
|
Integrations
Gemma
Llama
Mistral AI
ChatGPT
Comet
DeepSpeed
Devstral
Falcon
GPT-J
Google Colab
|
|||||
|
|
|