All-powerful, ever-pervasive AI is running out of internet

Artificial intelligence (AI) has relied on high-quality language data to train its models, but supply is running low. That depletion is forcing companies to look elsewhere for data sourcing as well as to change their algorithms to use data more efficiently. 

What is the scope of AI’s data problem?

Artificial intelligence needs to be trained, and data and information is used to accomplish that. Trouble is, the data is running out. A paper by Epoch, an AI research organization, found that AI could exhaust all the current high-quality language data available on the internet as soon as 2026. This could pose a problem as AI continues to grow. “The issue stems from the fact that, as researchers build more powerful models with greater capabilities, they have to find ever more texts to train them on,” said the MIT Technology Review. The quality of the data used in training AI is important. “The [data shortage] issue stems partly from the fact that language AI researchers filter the data they use to train models into two categories: high-quality and low-quality,” said the Review. “The line between the two categories can be fuzzy,” but “text from [high-quality data] is viewed as better-written and is often produced by professional writers.” 

AI models require vast amounts of data to be functional. For example, “the algorithm powering ChatGPT was originally trained on 570 gigabytes of text data, or about 300 billion words,” said Singularity Hub. In addition, “low-quality data such as social media posts or blurry photographs are easy to source but aren’t sufficient to train high-performing AI models,” and could even be “biased or prejudiced or may include disinformation or illegal content which could be replicated by the model.” Much of the data on the internet is considered useless for AI modeling. Instead, “AI companies are hunting for untapped information sources and rethinking how they train these systems,” said The Wall Street Journal. “Companies also are experimenting with using AI-generated, or synthetic, data as training material — an approach many researchers say could actually cause crippling malfunctions.”

  The great beaver comeback

What are AI companies doing to combat the imminent data scarcity?

The ticking clock on high-quality data has forced AI developers to think more creatively. For instance, Google has considered using user data from Google Docs, Google Sheets and similar company products. Other companies are “searching for content outside the free online space, such as that held by large publishers and offline repositories,” like those published before the internet existed, said Singularity Hub. Meta has considered purchasing Simon & Schuster publishing house to gain access to all its literary works. More broadly, many companies have looked to synthetic data, which is generated by AI itself. “As long as you can get over the synthetic data event horizon, where the model is smart enough to make good synthetic data, everything will be fine,” OpenAI CEO Sam Altman said at a tech conference in 2023. However, using synthetic data can present other problems. “Feeding a model text that is itself generated by AI is considered the computer-science version of inbreeding,” said the Journal. “Such a model tends to produce nonsense, which some researchers call ‘model collapse.'”

The other option is to rework AI algorithms to better and more efficiently use the existing high-quality data. One strategy being explored is called curriculum learning, which is when “data is fed to language models in a specific order in hopes that the AI will form smarter connections between concepts,” said the Journal. If successful, the method could cut the data required to run an AI model by half. Companies may also diversify the data sets used in AI models to include some lower-quality sources or instead opt to create smaller models that require less data altogether. “We’ve seen how smaller models that are trained on higher-quality data can outperform larger models trained on lower-quality data,” Percy Liang, a computer science professor at Stanford University, said to the MIT Technology Review.

(Visited 1 times, 1 visits today)

Leave a Reply

Your email address will not be published. Required fields are marked *