Science & Social Science

Thinking About Thinking: The Nexus of Neuroscience, Psychology and AI Research

July 27, 2022 2050
Human size device with one arm and two 'fingers' moves colored blocks on a table
This robot, powered by an AI called Rosie, learned how to solve this puzzle from a human who communicated to the robot using natural language. (Photo: James Kirk/ CC BY-ND)

Progress in artificial intelligence has enabled the creation of artificial intelligences, or AIs, that perform tasks previously thought only possible for humans, such as translating languages, driving cars, playing board games at world-champion level and extracting the structure of proteins. However, each of these AIs has been designed and exhaustively trained for a single task and has the ability to learn only what’s needed for that specific task.

Recent AIs that produce fluent text, including in conversation with humans, and generate impressive and unique art can give the false impression of a mind at work. But even these are specialized systems that carry out narrowly defined tasks and require massive amounts of training.

The Conversation logo
This article by Paul S. Rosenbloom, Christian Lebiere and John E. Laird originally appeared on The Conversation, a Social Science Space partner site, under the title “Cross-pollination among neuroscience, psychology and AI research yields a foundational understanding of thinking”

It still remains a daunting challenge to combine multiple AIs into one that can learn and perform many different tasks, much less pursue the full breadth of tasks performed by humans or leverage the range of experiences available to humans that reduce the amount of data otherwise required to learn how to perform these tasks. The best current AIs in this respect, such as AlphaZero and Gato, can handle a variety of tasks that fit a single mold, like game-playing. Artificial general intelligence (AGI) that is capable of a breadth of tasks remains elusive.

Ultimately, AGIs need to be able to interact effectively with each other and people in various physical environments and social contexts, integrate the wide varieties of skill and knowledge needed to do so, and learn flexibly and efficiently from these interactions.

Building AGIs comes down to building artificial minds, albeit greatly simplified compared to human minds. And to build an artificial mind, you need to start with a model of cognition.

From human to Artificial General Intelligence

Humans have an almost unbounded set of skills and knowledge, and quickly learn new information without needing to be re-engineered to do so. It is conceivable that an AGI can be built using an approach that is fundamentally different from human intelligence. However, as three longtime researchers in AI and cognitive science, our approach is to draw inspiration and insights from the structure of the human mind. We are working toward AGI by trying to better understand the human mind, and better understand the human mind by working toward AGI.

From research in neuroscience, cognitive science and psychology, we know that the human brain is neither a huge homogeneous set of neurons nor a massive set of task-specific programs that each solves a single problem. Instead, it is a set of regions with different properties that support the basic cognitive capabilities that together form the human mind.

These capabilities include perception and action; short-term memory for what is relevant in the current situation; long-term memories for skills, experience and knowledge; reasoning and decision making; emotion and motivation; and learning new skills and knowledge from the full range of what a person perceives and experiences.

Instead of focusing on specific capabilities in isolation, AI pioneer Allen Newell in 1990 suggested developing Unified Theories of Cognition that integrate all aspects of human thought. Researchers have been able to build software programs called cognitive architectures that embody such theories, making it possible to test and refine them.

Cognitive architectures are grounded in multiple scientific fields with distinct perspectives. Neuroscience focuses on the organization of the human brain, cognitive psychology on human behavior in controlled experiments, and artificial intelligence on useful capabilities.

The Common Model of Cognition

We have been involved in the development of three cognitive architectures: ACT-R, Soar and Sigma. Other researchers have also been busy on alternative approaches. One paper identified nearly 50 active cognitive architectures. This proliferation of architectures is partly a direct reflection of the multiple perspectives involved, and partly an exploration of a wide array of potential solutions. Yet, whatever the cause, it raises awkward questions both scientifically and with respect to finding a coherent path to AGI.

Fortunately, this proliferation has brought the field to a major inflection point. The three of us have identified a striking convergence among architectures, reflecting a combination of neural, behavioral and computational studies. In response, we initiated a communitywide effort to capture this convergence in a manner akin to the Standard Model of Particle Physics that emerged in the second half of the 20th century.

a graphic showing a human head and brain on the left, a robot head with circuits on the right, and a chart with five colored blocks and arrows connecting the blocks
This basic model of cognition both explains human thinking and provides a blueprint for true artificial intelligence. (Image: Andrea Stocco/CC BY-ND)

This Common Model of Cognition divides humanlike thought into multiple modules, with a short-term memory module at the center of the model. The other modules – perception, action, skills and knowledge – interact through it.

Learning, rather than occurring intentionally, happens automatically as a side effect of processing. In other words, you don’t decide what is stored in long-term memory. Instead, the architecture determines what is learned based on whatever you do think about. This can yield learning of new facts you are exposed to or new skills that you attempt. It can also yield refinements to existing facts and skills.

The modules themselves operate in parallel; for example, allowing you to remember something while listening and looking around your environment. Each module’s computations are massively parallel, meaning many small computational steps happening at the same time. For example, in retrieving a relevant fact from a vast trove of prior experiences, the long-term memory module can determine the relevance of all known facts simultaneously, in a single step.

Guiding the way to Artificial General Intelligence

The Common Model is based on the current consensus in research in cognitive architectures and has the potential to guide research on both natural and artificial general intelligence. When used to model communication patterns in the brain, the Common Model yields more accurate results than leading models from neuroscience. This extends its ability to model humans – the one system proven capable of general intelligence – beyond cognitive considerations to include the organization of the brain itself.

We are starting to see efforts to relate existing cognitive architectures to the Common Model and to use it as a baseline for new work – for example, an interactive AI designed to coach people toward better health behavior. One of us was involved in developing an AI based on Soar, dubbed Rosie, that learns new tasks via instructions in English from human teachers. It learns 60 different puzzles and games and can transfer what it learns from one game to another. It also learns to control a mobile robot for tasks such as fetching and delivering packages and patrolling buildings.

Rosie is just one example of how to build an AI that approaches AGI via a cognitive architecture that is well characterized by the Common Model. In this case, the AI automatically learns new skills and knowledge during general reasoning that combines natural language instruction from humans and a minimal amount of experience – in other words, an AI that functions more like a human mind than today’s AIs, which learn via brute computing force and massive amounts of data.

From a broader AGI perspective, we look to the Common Model both as a guide in developing such architectures and AIs, and as a means for integrating the insights derived from those attempts into a consensus that ultimately leads to AGI.


Critical thinking bootcamp logo. Skills, tools and resources for librarians and faculty to combat misinformation amid rapidly changing technology. August 9 2022 9 a.m. ET

As artificial intelligence, or AI, becomes smarter and more deeply embedded in how we access information and algorithms increasingly dictate what information we consume, how can we train students to spot and respond to misinformation? And what ethical considerations do we need to think through along the way?  

To address these vital questions, SAGE Publishing — the parent of Social Science Space – will hold its Third Annual Critical Thinking Bootcamp on August 9. The free three-hour virtual event, starting at 9 a.m. ET, will offer insights, guidance, and resources to help librarians, professors, and other staff to engage in critical thinking in and out of the classroom. The sessions will spotlight trends in personal technology affecting our media ecosystem and offer educational tactics to help students combat misinformation. Read more or sign up now.

Paul S. Rosenbloom is a professor emeritus of computer science at the University of Southern California. Christian Lebiere is a research psychologist at Carnegie Mellon University. John E. Laird is the John L. Tishman Professor of Engineering at the University of Michigan.

View all posts by Paul S. Rosenbloom, Christian Lebiere and John E. Laird

Related Articles

Canada’s Storytellers Challenge Seeks Compelling Narratives About Student Research
Communication
November 21, 2024

Canada’s Storytellers Challenge Seeks Compelling Narratives About Student Research

Read Now
Our Open-Source Tool Allows AI-Assisted Qualitative Research at Scale
Innovation
November 13, 2024

Our Open-Source Tool Allows AI-Assisted Qualitative Research at Scale

Read Now
This Anthropology Course Looks at Built Environment From Animal Perspective
Industry
September 10, 2024

This Anthropology Course Looks at Built Environment From Animal Perspective

Read Now
2024 Henry and Bryna David Lecture: K-12 Education in the Age of AI
Event
September 5, 2024

2024 Henry and Bryna David Lecture: K-12 Education in the Age of AI

Read Now
Philosophy Has Been – and Should Be – Integral to AI

Philosophy Has Been – and Should Be – Integral to AI

Philosophy has been instrumental to AI since its inception, and should still be an important contributor as artificial intelligence evolves..

Read Now
The Decameron Revisited – Pandemic as Farce

The Decameron Revisited – Pandemic as Farce

After viewing the the televised version of the The Decameron, our Robert Dingwall asks what the farce set during the Black Death says about a more recent pandemic.

Read Now
New SSRC Project Aims to Develop AI Principles for Private Sector

New SSRC Project Aims to Develop AI Principles for Private Sector

The new AI Disclosures Project seeks to create structures that both recognize the commercial enticements of AI while ensuring that issues of safety and equity are front and center in the decisions private actors make about AI deployment.

Read Now
0 0 votes
Article Rating
Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Learn how your comment data is processed.

0 Comments
Newest
Oldest Most Voted
Inline Feedbacks
View all comments