Qwen: Alibaba Cloud's multilingual language models
Alibaba Cloud's pretrained LLMs supporting Chinese/English with up to 32K context length.
Learn more about Qwen
Qwen is a family of transformer-based large language models ranging from 1.8B to 72B parameters, developed by Alibaba Cloud. The models are pretrained on up to 3 trillion tokens of multilingual data with focus on Chinese and English content across various domains. Each model variant includes both base versions for general language modeling and chat versions fine-tuned for conversational interactions using supervised fine-tuning and reinforcement learning from human feedback. The models support features like tool usage, code interpretation, and mathematical problem solving, with quantized versions available in Int4 and Int8 formats for reduced memory requirements.
Multilingual Focus
Specifically trained on Chinese and English datasets with strong performance in both languages. Supports cross-lingual tasks and culturally relevant content generation.
Multiple Quantization
Provides Int4 and Int8 quantized versions alongside full-precision models. Enables deployment on resource-constrained hardware while maintaining performance.
Extended Context
Supports context lengths up to 32K tokens in newer model variants. Handles long-form documents and extended conversational contexts effectively.
See how people are using Qwen
Related Repositories
Discover similar tools and frameworks used by developers
StabilityMatrix
Multi-backend inference UI manager with embedded dependencies.
open_clip
PyTorch library for contrastive language-image pretraining.
tiktoken
Fast BPE tokenizer for OpenAI language models.
ONNX Runtime
Cross-platform engine for optimized ONNX model execution.
WiFi DensePose
System for real-time human pose tracking using WiFi Channel State Information without cameras or wearables.