Google's AI answers keep telling me 'You can't lick a badger twice' is a real saying

4 hours ago 3

Magnifying glass with "meaning" highlighted in search bar

Getty Images; Alyssa Powell/BI
  • Google's AI answers will give you a definition of any made-up saying. I tried: "You can't lick a badger twice."
  • This is exactly the kind of thing AI should be really good at — explaining language use. But something's off.
  • Is it a hallucination, or AI just being too eager to please?

What does "You can't lick a badger twice" mean?

Like many English sayings — "A bird in the hand is worth two in the bush," "A watched pot never boils" — it isn't even true. Frankly, nothing stops you from licking a badger as often as you'd like, although I don't recommend it.

(I'm sure Business Insider's lawyers would like me to insist you exercise caution when encountering wildlife, and that we cannot be held liable for any rabies infections.)

If the phrase doesn't ring a bell to you, it's because, unlike "rings a bell," it is not actually a genuine saying — or idiom — in the English language.

But Google's AI Overview sure thinks it's real, and will happily give you a detailed answer of what the phrase means.

Greg Jenner, a British historian and podcaster, saw people talking about this phenomenon on Threads and wanted to try it himself with a made-up idiom. The badger phrase "just popped into my head," he told Business Insider. His Google search spit out an answer that seemed reasonable.

I wanted to try this myself, so I made up a few fake phrases — like "You can't fit a duck in a pencil" — and added "meaning" onto my search query.

Google took me seriously and explained:

you can't fit a fuck in a pencil google search

"You can't fit a duck in a pencil." Business Insider

So I tried some others, like "The Road is full of salsa." (This one I'd like to see being used in real life, personally.)

A Google spokeswoman told me, basically, that its AI systems are trying their best to give you what you want — but that when people purposely try to play games, sometimes the AI can't exactly keep up.

"When people do nonsensical or 'false premise' searches, our systems will try to find the most relevant results based on the limited web content available," spokeswoman Meghann Farnsworth said.

"This is true of Search overall — and in some cases, AI Overviews will also trigger in an effort to provide helpful context."

the road is full of salsa meaning

"The road is full of salsa." Business Insider

Basically, AI Overviews aren't perfect (duh), and these fake idioms are "false premise" searches that are purposely intended to trip it up (fair enough).

Google does try to limit the AI Overviews from answering things that are "data voids," i.e., when there are no good web results to a question.

But clearly, it doesn't always work.

I have some ideas about what's going on here — some of it is good and useful, some of it isn't. As one might even say, it's a mixed bag.

But first, one more made-up phrase that Google tried hard to find meaning for: "Don't kiss the doorknob." Says Google's AI Overview:

don't kiss the doorknob google search

"Don't kiss the doorknob." Business Insider

So what's going on here?

The Good:

English is full of idioms like "kick the bucket" or "piece of cake." These can be confusing if English isn't your first language (and frankly, they're often confusing for native speakers, too). My case in point is that the phrase is commonly misstated as "case and point."

So it makes lots of sense that people would often be Googling to understand the meaning of a phrase they came across that they don't understand. And in theory, this is a great use for the AI Overview answers: You want to see the simply-stated answer right away, not click on a link.

The Bad:

AI should be really good at this particular thing. LLMs are trained on vast amounts of the English written language — reams of books, websites, YouTube transcriptions, etc., so being able to recognize idioms is something they should be very good at doing.

The fact that it's making mistakes here is not ideal. What's going wrong that Google's AI Overview isn't giving the real answer, which is "That isn't a phrase, you idiot"? Is it just a classic AI hallucination?

The ugly:

Comparatively, ChatGPT gave a better answer when I asked it about the badger phrase. It told me that it was not a standard English idiom, even though it had the vaguely folksy sound of one. Then it offered, "If we treat it like a real idiom (for fun)," and gave a possible definition.

So this isn't a problem across all AI — it seems to be a Google problem.

badger

You can't lick a badger twice? REUTERS/Russell Cheyne

This is somewhat different from last year's Google AI Overview answers fiasco where the results pulled in information from places like Reddit without considering sarcasm — remember when it suggested people should eat rocks for minerals or put glue in their pizza (someone on Reddit had once joked about glue in pizza, which seems to be where it drew from).

This is all very low-stakes and silly fun, making up fake phrases, but it speaks to the bigger, uglier problems with AI becoming more and more enmeshed in how we use the internet. It means Google searches are somehow worse, and since people start to rely on these more and more, that bad information is just getting out there into the world and taken as fact.

Sure, AI search will get better and more accurate, but what growing pains will we endure while we're in this middle phase of a kinda wonky, kinda garbage-y, slop-filled AI internet?

AI is here, it's already changing our lives. There's no going back, the horse has left the barn. Or as they say, you can't lick a badger twice.

Read Entire Article
| Opini Rakyat Politico | | |