2024-07-27
한어Русский языкEnglishFrançaisIndonesianSanskrit日本語DeutschPortuguêsΕλληνικάespañolItalianoSuomalainenLatina
Baijiao from Aofei Temple
Quantum Bit | Public Account QbitAI
AI training AI, could make AI stupid?!
Researchers from Oxford, Cambridge and other institutions have recently found that large models may collapse when trained with synthetic data. Their research results have been selected as the latestNature cover。
Direct one:GARBAGE OUT!
You know, now most of the big models of technology companies are using synthetic data to alleviate the "data shortage". This is undoubtedly a cold water for the entire industry.
The research team gave such an example.
They tested Meta's OPT-125m model, asking it for information about medieval architecture.
Each fine-tuning is trained by the data generated in the previous round. The results were good in the first few rounds. But in the ninth round, it started to speak nonsense...
What the hell is this about rabbits?!
The lead author of the paper said they had considered that synthetic data might cause errors in large models, but did not expect the models to deteriorate so quickly.
First, the team defined what a model crash is.
Model collapse is a degradation process where the content generated by the model will pollute the next generation of training data sets. After training on the polluted data, the new generation of models will easily misunderstand reality.
This cycle repeats itself, with each generation getting worse than the previous one.
As time goes by, there are mainly two situations: early model collapse and late model collapse.
In early model collapse, the model starts to lose some tail information (like some low probability events in the probability distribution) while in late model collapse, the model will converge to a distribution that has almost no resemblance to the original distribution.
The occurrence of this process is related to the model design, learning process and the quality of the data used.
Specifically in theory, these three errors mainly lead to the deviation of the large model from the original model.
The researchers then evaluated the impact of model collapse on language models. Since training large models from scratch is very expensive, they chose to evaluate the most common settings for language models:Fine-tune settings。
Each training cycle starts with a pre-trained model with the latest data. The training data comes from another fine-tuned pre-trained model.
They used the Meta causal language model OPT-125m, fine-tuned on wikitext2.
To generate data from the trained model, the team used a five-way beam search. They set the training sequence to be 64 tokens long; then for each token sequence in the training set, the model was asked to predict the next 64 tokens.
They go through all the original training datasets and generate an artificial dataset of the same size.If the model's error is 0, it generates the original wikitext2 dataset.
To further feel the difference, they used two different settings: one group had no original training data except for the initial training; the other group retained 10% of the original data.
The results show that over time, the errors produced by the model increase. Before the model completely breaks down, it also causes the model to forget low-probability events in the data set, and their outputs become more homogeneous. Eventually, this phenomenon occurs.
In addition, similar model collapse phenomena were observed in VAE and GMM models.
Professor Emily Wenger from Duke University said that so far, it has not been easy to alleviate the problem.
Leading technology companies have deployed a technology that embeds "watermarks" -
The difficulty is that this requires coordination among technology companies and is therefore not commercially viable.
In this way, the models trained by companies that previously obtained data from the Internet are more representative of the real world. Therefore, the first wave of large models has a first-mover advantage.
What do you think about this view?
Reference Links:
[1]https://www.nature.com/articles/d41586-024-02420-7
[2]https://www.nature.com/articles/d41586-024-02355-z
[3]https://www.nature.com/articles/s41586-024-07566-y