Qwen: Alibaba Cloud's multilingual language models
Alibaba Cloud's pretrained LLMs supporting Chinese/English with up to 32K context length.
Learn more about Qwen
Qwen is a family of transformer-based large language models ranging from 1.8B to 72B parameters, developed by Alibaba Cloud. The models are pretrained on up to 3 trillion tokens of multilingual data with focus on Chinese and English content across various domains. Each model variant includes both base versions for general language modeling and chat versions fine-tuned for conversational interactions using supervised fine-tuning and reinforcement learning from human feedback. The models support features like tool usage, code interpretation, and mathematical problem solving, with quantized versions available in Int4 and Int8 formats for reduced memory requirements.
Multilingual Focus
Specifically trained on Chinese and English datasets with strong performance in both languages. Supports cross-lingual tasks and culturally relevant content generation.
Multiple Quantization
Provides Int4 and Int8 quantized versions alongside full-precision models. Enables deployment on resource-constrained hardware while maintaining performance.
Extended Context
Supports context lengths up to 32K tokens in newer model variants. Handles long-form documents and extended conversational contexts effectively.
See how people are using Qwen
Top in AI & ML
Related Repositories
Discover similar tools and frameworks used by developers
whisper.cpp
Lightweight C++ port of OpenAI Whisper for cross-platform speech recognition.
llama_index
Connect LLMs to external data via RAG workflows.
segment-anything
Transformer-based promptable segmentation with zero-shot generalization.
Open Lovable
AI tool that generates React applications from website clones or user descriptions.
context7
MCP server delivering version-specific library documentation to LLMs.