Alibaba launches new AI model targeting rival DeepSeek, China’s hottest start-up

In a statement posted on WeChat, the e-commerce giant’s cloud computing and AI arm Alibaba Cloud said its new Qwen 2.5-Max model also outperformed OpenAI’s GPT-4o and Meta Platforms’ Llama-3.1-405B in LLM performance benchmark platforms Arena-Hard and LiveBench. Alibaba owns the South China Morning Post.
The benchmark performance of Qwen 2.5-Max, part of Alibaba’s Tongyi Qianwen LLM family, was on par with Anthropic’s Claude-3.5-Sonnet model, according to Alibaba Cloud. LLMs are the technology underpinning generative AI services like ChatGPT.
Alibaba’s multimodal model is offered in various sizes, from 3 billion to 72 billion parameters, and includes both base and instruction-tuned versions. The flagship model, Qwen2.5-VL-72B-Instruct, is now accessible through the Qwen Chat platform, while the entire Qwen2.5-VL series is available on open-source platform Hugging Face and Alibaba’s own open-source community Model Scope.
Alibaba Cloud’s new Qwen 2.5-Max artificial intelligence model is touted to have outperformed rival large language models from DeepSeek and OpenAI. Photo: AFP

Parameter is a machine-learning term for variables present in an AI system during training, which helps establish how data prompts yield the desired output. Open source gives public access to a software program’s source code, allowing third-party developers to modify or share its design, fix broken links or scale up its capabilities.

Read original article here

Denial of responsibility! Pioneer Newz is an automatic aggregator of the all world’s media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials, please contact us by email – [email protected]. The content will be deleted within 24 hours.

Leave a Comment