ERNIE-4.5-0.3B-PT is a 360 million parameter dense language model by Baidu, post-trained to enhance performance on general-purpose natural language tasks. As part of the ERNIE 4.5 series, it emphasizes compactness and accessibility while maintaining strong capabilities for both English and Chinese text generation. The model features 18 transformer layers, 16 attention heads, and a remarkably long context window of 131,072 tokens. Optimized for use with the Hugging Face Transformers library, it supports seamless inference and fine-tuning, including SFT, DPO, and LoRA methods via ERNIEKit. It is fully compatible with PyTorch and includes support for vLLM-based deployment. Though smaller in size, it benefits from ERNIE's large-scale training infrastructure and multimodal innovations. ERNIE-4.5-0.3B-PT is ideal for developers and researchers seeking a lightweight, open-access LLM for dialogue systems and general text generation tasks.
Features
- 360M parameters in a dense transformer architecture
- Post-trained for improved downstream accuracy
- 18 layers with 16 Q-attention heads
- Context length up to 131,072 tokens
- Fully compatible with Hugging Face Transformers
- Supports ERNIEKit workflows (SFT, DPO, LoRA)
- Apache 2.0 license with commercial use permissions
- Optimized for PyTorch and vLLM deployment workflows