27 September 2023

Meta’s new learning algorithm can teach AI to multi-task

Start the conversation

Will Douglas Heaven* says the single technique for teaching neural networks multiple skills is a step towards general-purpose AI.


If you can recognise a dog by sight, then you can probably recognise a dog when it is described to you in words.

Not so for today’s artificial intelligence.

Deep neural networks have become very good at identifying objects in photos and conversing in natural language, but not at the same time: there are AI models that excel at one or the other, but not both.

Part of the problem is that these models learn different skills using different techniques.

This is a major obstacle for the development of more general-purpose AI, machines that can multi-task and adapt.

It also means that advances in deep learning for one skill often do not transfer to others.

A team at Meta AI (previously Facebook AI Research) wants to change that.

The researchers have developed a single algorithm that can be used to train a neural network to recognise images, text, or speech.

The algorithm, called Data2vec, not only unifies the learning process but performs at least as well as existing techniques in all three skills.

“We hope it will change the way people think about doing this type of work,” says Michael Auli, a researcher at Meta AI.

The research builds on an approach known as self-supervised learning, in which neural networks learn to spot patterns in data sets by themselves, without being guided by labelled examples.

This is how large language models like GPT-3 learn from vast bodies of unlabelled text scraped from the internet, and it has driven many of the recent advances in deep learning.

Auli and his colleagues at Meta AI had been working on self-supervised learning for speech recognition.

But when they looked at what other researchers were doing with self-supervised learning for images and text, they realised that they were all using different techniques to chase the same goals.

Data2vec uses two neural networks, a student and a teacher.

First, the teacher network is trained on images, text, or speech in the usual way, learning an internal representation of this data that allows it to predict what it is seeing when shown new examples.

When it is shown a photo of a dog, it recognises it as a dog.

The twist is that the student network is then trained to predict the internal representations of the teacher.

In other words, it is trained not to guess that it is looking at a photo of a dog when shown a dog, but to guess what the teacher sees when shown that image.

Because the student does not try to guess the actual image or sentence but, rather, the teacher’s representation of that image or sentence, the algorithm does not need to be tailored to a particular type of input.

Data2vec is part of a big trend in AI toward models that can learn to understand the world in more than one way.

“It’s a clever idea,” says Ani Kembhavi at the Allen Institute for AI in Seattle, who works on vision and language.

“It’s a promising advance when it comes to generalized systems for learning.”

An important caveat is that although the same learning algorithm can be used for different skills, it can only learn one skill at a time.

Once it has learned to recognise images, it must start from scratch to learn to recognise speech.

Giving an AI multiple skills at once is hard, but that’s something the Meta AI team wants to look at next.

The researchers were surprised to find that their approach actually performed better than existing techniques at recognising images and speech, and performed as well as leading language models on text understanding.

Mark Zuckerberg is already dreaming up potential metaverse applications.

“This will all eventually get built into AR glasses with an AI assistant,” he posted to Facebook today.

“It could help you cook dinner, noticing if you miss an ingredient, prompting you to turn down the heat, or more complex tasks.”

For Auli, the main takeaway is that researchers should step out of their silos.

“Hey, you don’t need to focus on one thing,” he says.

“If you have a good idea, it might actually help across the board.”

*Will Douglas Heaven is senior editor for AI at MIT Technology Review, where he covers new research, emerging trends and the people behind them.

This article first appeared at technologyreview.com.

Start the conversation

Be among the first to get all the Public Sector and Defence news and views that matter.

Subscribe now and receive the latest news, delivered free to your inbox.

By submitting your email address you are agreeing to Region Group's terms and conditions and privacy policy.