Unitalk
Back to Discovery
Qwen

Qwen2 1.5B Instruct (Free)

bysiliconcloud
Qwen2-1.5B-Instruct is an instruction-tuned large language model in the Qwen2 series, with a parameter size of 1.5B. This model is based on the Transformer architecture and employs techniques such as the SwiGLU activation function, attention QKV bias, and group query attention. It excels in language understanding, generation, multilingual capabilities, coding, mathematics, and reasoning across multiple benchmark tests, surpassing most open-source models. Compared to Qwen1.5-1.8B-Chat, Qwen2-1.5B-Instruct shows significant performance improvements in tests such as MMLU, HumanEval, GSM8K, C-Eval, and IFEval, despite having slightly fewer parameters.

Providers Supporting This Model

SiliconCloud
QwenQwen/Qwen2-1.5B-Instruct
Maximum Context Length
32K
Maximum Output Length
--
Input Price
--
Output Price
--