Unveiling the most advanced Llama2 Chinese large model in the country

CN
巴比特
Follow
1 year ago

Author: FlagAlpha

Source: Llama Chinese Community

Image

Image Source: Generated by Wujie AI

Preface

On July 31, the Llama Chinese Community was the first to complete the first truly meaningful Chinese version of the Llama2-13B large model in China, achieving a significant optimization and improvement of the Chinese capabilities of the Llama2 model from the bottom layer. There is no doubt that the release of the Chinese version of Llama2 will usher in a new era of large models in China!


| The World's Strongest, but with a Shortcoming in Chinese

Llama2 is currently the strongest open-source large model globally, but its Chinese capabilities urgently need improvement.

As the most powerful open-source large model in the field of AI, Llama2 is pre-trained based on 200 trillion token data and fine-tuned on 1 million human-labeled data for dialogue modeling. It significantly outperforms many benchmark tests including inference, programming, dialogue, and knowledge testing compared to other open-source large language models such as MPT, Falcon, and the first generation of LLaMA, and for the first time rivals the commercial GPT-3.5, standing out among many open-source models.

Image

Image

Although the pre-training data of Llama2 has doubled compared to the first generation, the proportion of Chinese pre-training data is still very small, accounting for only 0.13%, which has resulted in the weak Chinese capabilities of the original Llama2.

When we asked some Chinese questions, we found that in most cases, Llama2 could not answer in Chinese, or it answered questions in a mixed form of Chinese and English. Therefore, it is necessary to optimize Llama2 based on large-scale Chinese data to give Llama2 better Chinese capabilities.

Image

To this end, a top Chinese university's doctoral team of large models founded the Llama Chinese Community and embarked on the journey of training the Llama2 large model in Chinese.


| The Leading Llama Chinese Community

The Llama Chinese Community is the most advanced open-source large model Chinese community in China, with 4.7k stars on Github, led by doctoral teams from Tsinghua University, Jiao Tong University, and Zhejiang University, bringing together more than 60 senior engineers in the AI field and over 2000 top talents from various industries.

Image

Community Link:

https://github.com/FlagAlpha/Llama2-Chinese

Image

Community History:

Image


| The First Pre-trained Chinese Llama2 Large Model!

Not fine-tuning! But trained from scratch based on 200B Chinese corpus!

The Llama Chinese Community is the first in China to complete the truly meaningful Chinese version of the 13B Llama2 model: Llama2-Chinese-13B, achieving a significant optimization and improvement of the Chinese capabilities of the Llama2 model from the bottom layer.

There are roughly two routes for the Chinese transformation of Llama2:

  1. Fine-tuning the pre-trained model based on existing Chinese instruction datasets to align the Chinese question-answering capabilities of the base model. The advantage of this route is the low cost, small amount of fine-tuning data, and less computational resources required, enabling the rapid realization of a prototype of a Chinese Llama.

    However, the disadvantages are also obvious. Fine-tuning can only stimulate the existing Chinese capabilities of the base model, but due to the limited Chinese training data of Llama2 itself, the stimulated capabilities are also limited. To fundamentally enhance the Chinese capabilities of the Llama2 model, it is necessary to start from pre-training.

  2. Pre-training based on large-scale Chinese corpus. The downside of this route is the high cost! It requires not only large-scale high-quality Chinese data but also large-scale computational resources. However, the advantages are also obvious, as it can optimize the Chinese capabilities from the bottom layer of the model, truly achieving a fundamental effect, injecting powerful Chinese capabilities into the core of the large model!

To achieve a thorough Chinese large model from the core, we have chosen the second route! We have gathered a batch of high-quality Chinese corpus data to optimize the Llama2 large model from pre-training. Some of the pre-training data is as follows:

Image

The initial pre-training data of the Llama2-Chinese-13B model contains 200B tokens. In the future, we will continue to iteratively update Llama2-Chinese, gradually increasing the pre-training data to 1T token. In addition, we will gradually open up the Chinese pre-training version of the 70B model, so stay tuned!

Image

We have asked the large model questions from different aspects such as general knowledge, language understanding, creative ability, logical reasoning, code programming, and work skills, and have achieved satisfactory results!

Some of the results are as follows:

General Knowledge

Image

Language Understanding

Image

Creative Ability

Image

Logical Reasoning

Image

Code Programming

Image

Work Skills

Image


🏅Llama Chinese Community Navigation Plan🏅

With a focus on the long-term development and rapid iteration of the community, on the one hand, we provide professional technical services for every AI enthusiast who is passionate and willing to invest in the wave of large models, and on the other hand, we enable every participant in the community to take the lead in the rapidly developing AI era and obtain various resource connections. We are now launching the first phase of the Llama Chinese Community Navigation Plan for a limited time! Each member of the Navigation Plan will receive the following "7TOP" benefits:

  1. Model TOP

    Join to obtain the right to use the first pre-trained Chinese version of Llama2-Chinese-13B model in China (non-fine-tuned version), and in the future, we will continue to enhance the Chinese capabilities of the model's core based on larger-scale data, and will also prioritize providing the most advanced model versions for each member of the Navigation Plan.

  2. Technology TOP

    Led by the top doctoral team from domestic universities, the most professional large model technical team. Whether it's the most cutting-edge technical issues or in-depth theoretical analysis, we will provide you with the most advanced solutions.

  3. Service TOP

    In the Navigation Plan, you will receive personalized 1V1 guidance. Whenever and wherever you have questions, we will provide timely answers. We are committed to providing comprehensive support to help you quickly implement the application of the Llama2 large model and ensure that you achieve technological breakthroughs smoothly. We will also help analyze and solve problems related to large models for your company.

  4. Teaching TOP

    A teaching mode that combines theory with practice will allow you to explore the mysteries of large models. From the technical analysis of large models to the explanation of key algorithms and papers, from building a private large model from scratch to training industry large models, we will guide you step by step to achieve technological advancement. The course outline is as follows:

    Image

  5. Resource TOP

    We have the largest Llama Chinese community in China, with 4.7k stars on Github, bringing together over 2000 top talents. Here, you can communicate with AI investors, CEO entrepreneurs, and leading figures from various industries, seeking one-stop services for cooperation, investment, promotion, and recruitment. Whether you are looking for a job/partner/investment/sales of products, we can meet your needs. This is a golden platform for technical talent exchange, where you can find top experts from various industries to communicate and discuss together.

  6. Activity TOP

    We not only hold regular online events, but also provide technical presentations and exchanges at offline events, aiming to empower various industries based on the Llama2 large model. We provide you with the opportunity to interact directly with top experts, allowing you to move forward alongside industry leaders. Whether you are a technical novice or an experienced expert, we will provide you with excellent opportunities to collaborate with the world's top technical talents and jointly explore the future!

  7. Computing Power TOP

    The community provides channels for students to access computing resources at a lower cost than the market price. We understand the importance of computing power for technological development and provide you with efficient and stable computing power support to help you shine in the field of technology.

免责声明:本文章仅代表作者个人观点,不代表本平台的立场和观点。本文章仅供信息分享,不构成对任何人的任何投资建议。用户与作者之间的任何争议,与本平台无关。如网页中刊载的文章或图片涉及侵权,请提供相关的权利证明和身份证明发送邮件到support@aicoin.com,本平台相关工作人员将会进行核查。

Bitget:注册返10%, 送$100
Ad
Share To
APP

X

Telegram

Facebook

Reddit

CopyLink