📢 Gate Square #Creator Campaign Phase 1# is now live – support the launch of the PUMP token sale!
The viral Solana-based project Pump.Fun ($PUMP) is now live on Gate for public sale!
Join the Gate Square Creator Campaign, unleash your content power, and earn rewards!
📅 Campaign Period: July 11, 18:00 – July 15, 22:00 (UTC+8)
🎁 Total Prize Pool: $500 token rewards
✅ Event 1: Create & Post – Win Content Rewards
📅 Timeframe: July 12, 22:00 – July 15, 22:00 (UTC+8)
📌 How to Join:
Post original content about the PUMP project on Gate Square:
Minimum 100 words
Include hashtags: #Creator Campaign
Alpacas evolve into whales, Meta "automates" alignment, and Humpback defeats all existing LLaMa models
Editors: Xiaozhou, Chen Ping
**Source:**The Heart of the Machine
In the past year, the large language model (LLM) represented by ChatGPT and GPT-4 has developed rapidly, followed by Meta's open source LLaMa and Llama 2 series models, which have also caused quite a stir in the AI world. But what followed was constant controversy. Some people believed that LLM had some uncontrollable risks, posing some potential threats to human survival.
In order to deal with these challenges, the research on LLM alignment has become more and more important. Some researchers have proposed instruction following (instruction following), but this method requires a lot of manual annotation. However, annotating such high-quality instruction-following datasets is costly.
In this paper, researchers from Meta AI propose a scalable method called instruction backtranslation, which builds a high-quality instruction-following language model by automatically annotating corresponding instructions.
Specifically, the study starts with a language model as a seed model, which is fine-tuned on a small amount of seed data as well as web corpora. The role of the seed model is to build training samples, and then some high-quality samples from these samples will be screened out, and then these data are used to fine-tune a more powerful model.
After two rounds of iterative dataset fine-tuning LLaMa, the resulting model Humpback outperforms other existing non-distilled models such as LIMA, Claude, Guanaco, etc. on the Alpaca leaderboard.
Humpback originally meant a humpback whale, also known as a humpback whale. Meta named the model Humpback, so there is no deep meaning.
Turing Award winner Yann LeCun gave a high-level overview of the study's methodology and praised Meta's work as an important contribution to alignment research:
This paper proposes a new data augmentation paradigm that needs to be completed in two steps. First, it is necessary to have a set of seed (instruction, output) pairs and a corpus to generate more good instruction data.
Method Introduction
The study proposes a self-training approach that generally assumes access to a basic language model, a small amount of seed data, and an unlabeled sample set (such as a web corpus). Unlabeled data is often a large collection of documents of various shapes, written by humans, including content on various topics of human interest, but most importantly, it has not been paired with instructions.
There are two key assumptions here. The first assumption is that there are some subsets of this very large text set (unlabeled sample set) that are suitable as generated samples for some user instructions. The second hypothesis is that the instructions of these candidate answers can be predicted, which can be used to form high-quality sample pairs to train instruction-following models.
As shown in Figure 1 below, the study proposes that the instruction back-translation process includes two core steps:
Experiment and Results
The data set in this paper mainly includes seed data and enhanced data. The specific information is shown in Table 2 and Figure 2: