2024T-EDGE文章顶部

Alibaba Open-Sources its Seven-Billion-Parameter AI Model Similar to Meta’s Llama 2

Both Meta and Baichuan Intelligence have launched their open source models with seven billion and 13 billion parameters respectively. The announcement of AliCloud will intensify the competition of open-sourced AI models.

Credit: Visual China

Credit: Visual China

BEIJING, August 4 (TMTPost) -- Alibaba’s cloud computing announced on Thursday that it will open-source its seven-billion-parameter large language model (LLM) Tongyi Qianwen and make it free for commercial use. The global ecological competition around open-sourced LLMs has further intensified.

Compared to the lively AI open source ecology in the West, the Chinese community lacks excellent fundamental models. The open source of Tongyi Qianwen is expected to provide more choices for the open source community and promote the construction of the Chinese AI open source ecosystem, according to AliCloud in its statement.

On April 7 this year, Alibaba’s own ChatGPT-like product Tongyi Qianwen began to invite testing. As an ultra-large language model, Tongyi Qianwen can complete multi-round conversations, write emails and novels, solve simple math problems and write codes.

AliCloud has never disclosed the size of Tongyi Qianwen's parameters before, and said the open-sourced model is only a miniaturized version. It added the move aims to help users simplify the process of model training and deployment. Therefore, instead of training models from scratch, the users can build high-quality models quickly by downloading pre-trained models and fine-tuning them.

Zhou Jingren, CTO of AliCloud Intelligence, said at the AliCloud Guangzhou Summit in June this year that they are very supportive of open-sourced models, which enable people to reduce the cost of learning and achieve breakthroughs of their own.

In February, Meta, the parent company of Facebook, made its LLM LLaMA available to research institutions, with four versions of seven billion, 13 billion, 33 billion and 65 billion parameters. On July 18, Meta introduced Llama 2 with 7 billion, 13 billion and 70 billion parameters, for free research and commercial use.

In China, Baichuan Intelligence, a large model startup founded by Wang Xiaochuan, the founder of Sogou, released Baichuan-7B, a seven-billion-parameter open-sourced model, in June this year, and Baichuan-13B, a 13 billion-parameter model, in July. According to Wang, Baichuan Intelligence will

release closed-source large models with tens of billions and hundreds of billions of parameters later.

The open-sourced LLMs are safer because developers and researchers in the community can stress-test it to quickly find and solve problems, and Meta can further improve its own models by fixing the holes, according to Meta in a statement on open-source Llama 2.

  However, Meta's intention to catch up with OpenAI and Google is also very clear. OpenAI made its model available from its inception to the release of GPT-2 in 2019, and since then it has closed the source of its models in order to make profits, including the latest GPT-4 released in March this year. Google's latest PaLM 2 is also a closed source model.

In May this year, a Google software engineer said in a post that the open source community will pose a potential threat to OpenAI and Google with lower-cost and faster-evolving AI models.

  An investor in the AI field also said that if OpenAI’s ChatGPT has brought the "iPhone moment" for AI, he is caring about the appearance of the "Android moment". The biggest difference between the Android operating system and Apple's iOS is the former's open source, which guarantees it more than 80% of the global smartphone market share.

本文系作者 neo_english 授权钛媒体发表,并经钛媒体编辑,转载请注明出处、作者和本文链接
本内容来源于钛媒体钛度号,文章内容仅供参考、交流、学习,不构成投资建议。
想和千万钛媒体用户分享你的新奇观点和发现,点击这里投稿 。创业或融资寻求报道,点击这里

敬原创,有钛度,得赞赏

赞赏支持
发表评论
0 / 300

根据《网络安全法》实名制要求,请绑定手机号后发表评论

登录后输入评论内容

文章详情页右侧banner

扫描下载App

Baidu
map