Navigate:
~$QWEN0.4%

Qwen: Alibaba Cloud's multilingual language models

Alibaba Cloud's pretrained LLMs supporting Chinese/English with up to 32K context length.

LIVE RANKINGS • 10:20 AM • STEADY
OVERALL
#185
77
AI & ML
#66
17
30 DAY RANKING TREND
ovr#185
·AI#66
STARS
20.5K
FORKS
1.7K
7D STARS
+73
7D FORKS
+8
Tags:
See Repo:
Share:

Learn more about Qwen

Qwen is a family of transformer-based large language models ranging from 1.8B to 72B parameters, developed by Alibaba Cloud. The models are pretrained on up to 3 trillion tokens of multilingual data with focus on Chinese and English content across various domains. Each model variant includes both base versions for general language modeling and chat versions fine-tuned for conversational interactions using supervised fine-tuning and reinforcement learning from human feedback. The models support features like tool usage, code interpretation, and mathematical problem solving, with quantized versions available in Int4 and Int8 formats for reduced memory requirements.

Qwen

1

Multilingual Focus

Specifically trained on Chinese and English datasets with strong performance in both languages. Supports cross-lingual tasks and culturally relevant content generation.

2

Multiple Quantization

Provides Int4 and Int8 quantized versions alongside full-precision models. Enables deployment on resource-constrained hardware while maintaining performance.

3

Extended Context

Supports context lengths up to 32K tokens in newer model variants. Handles long-form documents and extended conversational contexts effectively.


See how people are using Qwen

Loading tweets...


[ EXPLORE MORE ]

Related Repositories

Discover similar tools and frameworks used by developers