Qwen, also known as Tongyi Qianwen, is a family of large language models developed by Alibaba Cloud. These models are designed to handle a variety of tasks, including natural language understanding, text generation, coding, and mathematical problem-solving. The Qwen series includes models with different parameter sizes, ranging from 0.5 to 72 billion parameters, catering to diverse application needs.
The latest iteration, Qwen 2.5, has been trained on an extensive dataset comprising up to 18 trillion tokens. This extensive training has significantly improved its knowledge base and capabilities. Notably, Qwen 2.5 has shown big improvements in coding proficiency and mathematical reasoning. Additionally, it excels in following instructions, generating long-form text, understanding structured data, and producing structured outputs.
Alibaba has also expanded the Qwen series to include specialized models such as Qwen-VL for vision-language tasks, Qwen-Audio for audio analysis, Qwen-Coder for coding applications, and Qwen-Math for mathematical problem-solving. These specialized models are designed to address specific domains, improving the versatility and applicability of the Qwen family across various industries.
In its commitment to fostering community innovation and accessibility, Alibaba has made over 100 Qwen models open-source, allowing developers and researchers to leverage these models for various applications. This open-source approach aims to accelerate advancements in AI by providing the tools necessary for further research and development.