Posted by Roger Mallett Posted on 3 November 2023

Can AI Mimic Human Compositional Thinking?

Researchers advanced the capabilities of neural networks to make compositional generalizations, similar to how humans grasp and expand on new concepts

This new technique, named Meta-learning for Compositionality (MLC), challenges decades-old skepticism about the capacity of artificial neural networks.

MLC involves training the network through episodic learning to enhance its generalization skills.

Remarkably, in various tasks, MLC matched or even surpassed human performance.

Key Facts:

  1. The MLC technique focuses on episodic training of neural networks, allowing them to better generalize new concepts compositionally.
  2. In tasks involving novel word combinations, MLC performed on par with or superior to human participants.
  3. Despite its advancements, popular models like ChatGPT and GPT-4 have challenges with this kind of compositional generalization, but MLC might be a solution to enhance their capabilities.

Humans have the ability to learn a new concept and then immediately use it to understand related uses of that concept—once children know how to “skip,” they understand what it means to “skip twice around the room” or “skip with your hands up.” 

But are machines capable of this type of thinking?

In the late 1980s, Jerry Fodor and Zenon Pylyshyn, philosophers and cognitive scientists, posited that artificial neural networks—the engines that drive artificial intelligence and machine learning— are not capable of making these connections, known as “compositional generalizations.”

However, in the decades since, scientists have been developing ways to instill this capacity in neural networks and related technologies, but with mixed success, thereby keeping alive this decades-old debate.

Researchers at New York University and Spain’s Pompeu Fabra University have now developed a technique—reported in the journal Nature—that advances the ability of these tools, such as ChatGPT, to make compositional generalizations.

This technique, Meta-learning for Compositionality (MLC), outperforms existing approaches and is on par with, and in some cases better than, human performance.

MLC centers on training neural networks—the engines driving ChatGPT and related technologies for speech recognition and natural language processing—to become better at compositional generalization through practice.

Developers of existing systems, including large language models, have hoped that compositional generalization will emerge from standard training methods, or have developed special-purpose architectures in order to achieve these abilities.

MLC, in contrast, shows how explicitly practicing these skills allow these systems to unlock new powers, the authors note.

“For 35 years, researchers in cognitive science, artificial intelligence, linguistics, and philosophy have been debating whether neural networks can achieve human-like systematic generalization,” says Brenden Lake, an assistant professor in NYU’s Center for Data Science and Department of Psychology and one of the authors of the paper.

Read More – Can AI Mimic Human Compositional Thinking?

From our advertisers