Google’s powerful AI spotlights human cognitive glitch: Mistaking fluent speech for fluent thought
Kyle Mahowald, Assistant Professor of Linguistics, University of Texas at Austin College of Liberal Arts; Anna A. Ivanova, PhD Candidate in Brain and Cognitive Sciences, Massachusetts Institute of Technology
Words can have a powerful effect on people, even when they’re generated by an unthinking machine. iStock via Getty Images
When you read a sentence like this one, your past experience tells you that it’s written by a thinking, feeling human. And, in this case, there is indeed a human typing these words: [Hi, there!] But these days, some sentences that appear remarkably humanlike are actually generated by artificial intelligence systems trained on massive amounts of human text.
People are so accustomed to assuming that fluent language comes from a thinking, feeling human that evidence to the contrary can be difficult to wrap your head around. How are people likely to navigate this relatively uncharted territory? Because of a persistent tendency to associate fluent expression with fluent thought, it is natural — but potentially misleading — to think that if an AI model can express itself fluently, that means it thinks and feels just like humans do.
Thus, it is perhaps unsurprising that a former Google engineer recently claimed that Google’s AI system LaMDA has a sense of self because it can eloquently generate text about its purported feelings. This event and the subsequent media coverage led to a number of rightly skeptical articles and posts about the claim that computational models of human language are sentient, meaning capable of thinking and feeling and experiencing.
The question of what it would mean for an AI model to be sentient is complicated (see, for instance, our colleague’s take), and our goal here is not to settle it. But as languageresearchers, we can use our work in cognitive science and linguistics to explain why it is all too easy for humans to fall into the cognitive trap of thinking that an entity that can use language fluently is sentient, conscious or intelligent.
Using AI to generate humanlike language
Text generated by models like Google’s LaMDA can be hard to distinguish from text written by humans. This impressive achievement is a result of a decadeslong program to build models that generate grammatical, meaningful language.
The first computer system to engage people in dialogue was psychotherapy software called Eliza, built more than half a century ago. Rosenfeld Media/Flickr, CC BY
Early versions dating back to at least the 1950s, known as n-gram models, simply counted up occurrences of specific phrases and used them to guess what words were likely to occur in particular contexts. For instance, it’s easy to know that “peanut butter and jelly” is a more likely phrase than “peanut butter and pineapples.” If you have enough English text, you will see the phrase “peanut butter and jelly” again and again but might never see the phrase “peanut butter and pineapples.”
Today’s models, sets of data and rules that approximate human language, differ from these early attempts in several important ways. First, they are trained on essentially the entire internet. Second, they can learn relationships between words that are far apart, not just words that are neighbors. Third, they are tuned by a huge number of internal “knobs” – so many that it is hard for even the engineers who design them to understand why they generate one sequence of words rather than another.
The models’ task, however, remains the same as in the 1950s: determine which word is likely to come next. Today, they are so good at this task that almost all sentences they generate seem fluid and grammatical.
Peanut butter and pineapples?
We asked a large language model, GPT-3, to complete the sentence “Peanut butter and pineapples___”. It said: “Peanut butter and pineapples are a great combination. The sweet and savory flavors of peanut butter and pineapple complement each other perfectly.” If a person said this, one might infer that they had tried peanut butter and pineapple together, formed an opinion and shared it with the reader.
But how did GPT-3 come up with this paragraph? By generating a word that fit the context we provided. And then another one. And then another one. The model never saw, touched or tasted pineapples — it just processed all the texts on the internet that mention them. And yet reading this paragraph can lead the human mind — even that of a Google engineer — to imagine GPT-3 as an intelligent being that can reason about peanut butter and pineapple dishes.
Large AI language models can engage in fluent conversation. However, they have no overall message to communicate, so their phrases often follow common literary tropes, extracted from the texts they were trained on. For instance, if prompted with the topic “the nature of love,” the model might generate sentences about believing that love conquers all. The human brain primes the viewer to interpret these words as the model’s opinion on the topic, but they are simply a plausible sequence of words.
The human brain is hardwired to infer intentions behind words. Every time you engage in conversation, your mind automatically constructs a mental model of your conversation partner. You then use the words they say to fill in the model with that person’s goals, feelings and beliefs.
The process of jumping from words to the mental model is seamless, getting triggered every time you receive a fully fledged sentence. This cognitive process saves you a lot of time and effort in everyday life, greatly facilitating your social interactions.
However, in the case of AI systems, it misfires — building a mental model out of thin air.
A little more probing can reveal the severity of this misfire. Consider the following prompt: “Peanut butter and feathers taste great together because___”. GPT-3 continued: “Peanut butter and feathers taste great together because they both have a nutty flavor. Peanut butter is also smooth and creamy, which helps to offset the feather’s texture.”
The text in this case is as fluent as our example with pineapples, but this time the model is saying something decidedly less sensible. One begins to suspect that GPT-3 has never actually tried peanut butter and feathers.
Ascribing intelligence to machines, denying it to humans
A sad irony is that the same cognitive bias that makes people ascribe humanity to GPT-3 can cause them to treat actual humans in inhumane ways. Sociocultural linguistics — the study of language in its social and cultural context — shows that assuming an overly tight link between fluent expression and fluent thinking can lead to bias against people who speak differently.
These biases are deeply harmful, often lead to racist and sexist assumptions, and have been shown again and again to be unfounded.
Fluent language alone does not imply humanity
Will AI ever become sentient? This question requires deep consideration, and indeed philosophers have pondered it for decades. What researchers have determined, however, is that you cannot simply trust a language model when it tells you how it feels. Words can be misleading, and it is all too easy to mistake fluent speech for fluent thought.
Contributors to this article were Evelina Fedorenko, Associate Professor of Neuroscience, Massachusetts Institute of Technology; Idan Asher Blank, Assistant Professor of Psychology and Linguistics, UCLA Luskin School of Public Affairs; Joshua B. Tenenbaum, Professor of Computational Cognitive Science, Massachusetts Institute of Technology; and Nancy Kanwisher, Professor of Cognitive Neuroscience, Massachusetts Institute of Technology.
Kyle Mahowald receives funding from NSF.
Evelina Fedorenko receives funding from NIH.
Joshua B. Tenenbaum receives relevant funding from NSF, the US Department of Defense, IBM, Google, and Microsoft.
Nancy Kanwisher receives funding from NIH and NSF.
Anna A. Ivanova and Idan Asher Blank do not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and have disclosed no relevant affiliations beyond their academic appointment.
This article is republished from The Conversation under a Creative Commons license.
Amazon’s Alexa could soon mimic voices of dead relatives
Stanisic Vladimir // Shutterstock
It’s a jungle out there—few places so much so as the world of “smart” advertising.
There, marketing geniuses have developed increasingly sophisticated algorithms that take all the information gathered about you online or from your phone and piece together a customer profile that may include everything from your favorite pair of socks to your children’s names.
Analyzing current market practices, Wicked Reports explored how artificial intelligence, or AI, can be wielded to gather data and make sales predictions across the internet. Some techniques you may know, such as persistent cookies that turn your computer into a ping hub for the websites you visit. Others are much more sophisticated, compiling all of your characteristics by analyzing what you’ve bought in the past, what you’ve put in your cart and abandoned, and what you’ve searched for. From there, advertisers can even make a version of similar customers to market to them as well.
The digital advertising industry is expected to crest $20 billion in 2022. That’s far from enough to crack the top 10 biggest industries in the U.S., but it’s a substantial amount of money—particularly when compared to the big-ticket ad buys of the past in splashy magazine spreads. Companies today are more eager than ever to spend what it takes to bring in ideal customers.
Continue reading to discover some of the tactics AI uses to predict buying behaviors.
Stanisic Vladimir // Shutterstock
It’s a jungle out there—few places so much so as the world of “smart” advertising.
There, marketing geniuses have developed increasingly sophisticated algorithms that take all the information gathered about you online or from your phone and piece together a customer profile that may include everything from your favorite pair of socks to your children’s names.
Analyzing current market practices, Wicked Reports explored how artificial intelligence, or AI, can be wielded to gather data and make sales predictions across the internet. Some techniques you may know, such as persistent cookies that turn your computer into a ping hub for the websites you visit. Others are much more sophisticated, compiling all of your characteristics by analyzing what you’ve bought in the past, what you’ve put in your cart and abandoned, and what you’ve searched for. From there, advertisers can even make a version of similar customers to market to them as well.
The digital advertising industry is expected to crest $20 billion in 2022. That’s far from enough to crack the top 10 biggest industries in the U.S., but it’s a substantial amount of money—particularly when compared to the big-ticket ad buys of the past in splashy magazine spreads. Companies today are more eager than ever to spend what it takes to bring in ideal customers.
Continue reading to discover some of the tactics AI uses to predict buying behaviors.
Amazon’s Alexa could soon mimic voices of dead relatives
picture alliance // Getty Images
You may know about cookies: tiny text files that websites deposit on your computer as a way to track online behavior.
When you visit websites from Europe, for example, a law there mandates that you click through a cookie agreement that’s much more transparent than in the U.S. There are session cookies lasting one browsing “session” (until you restart your computer or browser) and persistent cookies that stay until you delete them. Think of a cookie as a waving arm each time you visit the same website. Together, they form a heat map of how often and when you visit every website in your browsing history. They can even flag your presence to other websites as a way to combine your data.
picture alliance // Getty Images
You may know about cookies: tiny text files that websites deposit on your computer as a way to track online behavior.
When you visit websites from Europe, for example, a law there mandates that you click through a cookie agreement that’s much more transparent than in the U.S. There are session cookies lasting one browsing “session” (until you restart your computer or browser) and persistent cookies that stay until you delete them. Think of a cookie as a waving arm each time you visit the same website. Together, they form a heat map of how often and when you visit every website in your browsing history. They can even flag your presence to other websites as a way to combine your data.
Amazon’s Alexa could soon mimic voices of dead relatives
Anna Hoychuk // Shutterstock
User characteristics, and something called demographic segmentation, is a key way online advertising targets you. User characteristics are any of your qualities, from your gender and age to what car you drive and the pets you own. These user characteristics lead to the advertising concept of demographic segmentation, in which companies can buy lists of really specific people.
Are you a 25-year-old white man with one dog, a full-time job as an auto tech, and an apartment rental in a “transitional” neighborhood? We have just the plaid shirt for you.
Anna Hoychuk // Shutterstock
User characteristics, and something called demographic segmentation, is a key way online advertising targets you. User characteristics are any of your qualities, from your gender and age to what car you drive and the pets you own. These user characteristics lead to the advertising concept of demographic segmentation, in which companies can buy lists of really specific people.
Are you a 25-year-old white man with one dog, a full-time job as an auto tech, and an apartment rental in a “transitional” neighborhood? We have just the plaid shirt for you.
Amazon’s Alexa could soon mimic voices of dead relatives
mhong84 // Shutterstock
If you’ve used GPS in your smartphone or any of the hyperlocal dating apps, you’ve leveraged location data to your advantage—at least for now.
How does your phone know where you are? Cellphone towers ping your phone when you’re nearby. In your home, your Wi-Fi network is likely hardcoded with your location. That’s also true of any Wi-Fi network you hop into or onto during your errands, at school, at work, and so forth. After that, GPS can pinpoint your phone to an alarmingly small area as you carry it around, so not just in your home but in one corner of one room.
mhong84 // Shutterstock
If you’ve used GPS in your smartphone or any of the hyperlocal dating apps, you’ve leveraged location data to your advantage—at least for now.
How does your phone know where you are? Cellphone towers ping your phone when you’re nearby. In your home, your Wi-Fi network is likely hardcoded with your location. That’s also true of any Wi-Fi network you hop into or onto during your errands, at school, at work, and so forth. After that, GPS can pinpoint your phone to an alarmingly small area as you carry it around, so not just in your home but in one corner of one room.
Amazon’s Alexa could soon mimic voices of dead relatives
Gorodenkoff // Shutterstock
Some items on this list are not very surprising, or we’re used to being told about them so they don’t seem as insidious and scary as they once did. But people are likely still surprised by the depths that companies will go to in order to better advertise to you. Your favorite clothing store, for example, might put together a complete data “picture” of you: what you’ve purchased from them, what size you shop for, where your address is, and more. Then they can reverse engineer someone just like you and buy a demographically matching list.
Anything can be filtered until just the exact desired customer base remains, and then they buy the ads.
Gorodenkoff // Shutterstock
Some items on this list are not very surprising, or we’re used to being told about them so they don’t seem as insidious and scary as they once did. But people are likely still surprised by the depths that companies will go to in order to better advertise to you. Your favorite clothing store, for example, might put together a complete data “picture” of you: what you’ve purchased from them, what size you shop for, where your address is, and more. Then they can reverse engineer someone just like you and buy a demographically matching list.
Anything can be filtered until just the exact desired customer base remains, and then they buy the ads.
Amazon’s Alexa could soon mimic voices of dead relatives
Habichtland // Shutterstock
How much do you know about your IP address? Many of us are old enough to remember a time when connecting to the internet required knowing a specific IP address and typing it into our PC settings.
Today, the router you likely have in your home has a hard-coded IP address whose number values reflect where you are as well as which “node” you have on your local network. That information may be for sale to different companies because, with the right technology, they can use some IP addresses in order to infer the rest—and guess where you live. Apple is among the tech companies pushing back on IP targeting of this nature by masking IP addresses in its proprietary browser Safari.
This story originally appeared on Wicked Reports and was produced and distributed in partnership with Stacker Studio.
Habichtland // Shutterstock
How much do you know about your IP address? Many of us are old enough to remember a time when connecting to the internet required knowing a specific IP address and typing it into our PC settings.
Today, the router you likely have in your home has a hard-coded IP address whose number values reflect where you are as well as which “node” you have on your local network. That information may be for sale to different companies because, with the right technology, they can use some IP addresses in order to infer the rest—and guess where you live. Apple is among the tech companies pushing back on IP targeting of this nature by masking IP addresses in its proprietary browser Safari.
This story originally appeared on Wicked Reports and was produced and distributed in partnership with Stacker Studio.