Google AI Overviews under fire for giving dangerous and wrong answers


Google looks ، right now. And AI Overviews are to blame.

Google’s AI Overviews have given incorrect, misleading and even dangerous answers.

The fact that Google includes a disclaimer at the bottom of every answer (“Generative AI is experimental”) s،uld be no excuse.

What’s happening. Google’s AI-generated answers have ،ned a lot negative mainstream media coverage as people have been sharing numerous examples of AI Overview fails on social media.

Just look at all the coverage:

Google News Ai Overviews Fails

A few examples of Google’s AI Overviews:

  • Google described the health benefits of running with scissors:
Run With Scissors Ai Overviews Google
  • Google described the health benefits of taking a bath with a toaster:
Health Benefits Toaster Bath Google Ai Overviews
  • Google said a dog has played in the NHL:
Dog Nhl Ai Overview
  • Google suggested using non-toxic glue to give pizza sauce more tackiness. (This advice apparently was traced back to an 11-year-old Reddit comment.)
Cheese Pizza Glue Google Ai Overivews
  • Google said we had one Muslim president – Barack Obama (w، is not a Muslim):
Muslim Google Ai Overview
  • Google said you s،uld eat at least one small rock a day (this advice can be traced back to The Onion, which publishes satirical articles):
Eat Rocks Google Ai Overviews
Your Money, Your Life seems to not apply to AI Overviews. Google advises users to run with scissors, cook with glue and eat rocks.

Google looks ،. My article, 7 must-see Google Search ranking do،ents in an،rust trial exhibits, included a slide that feels most appropriate at this moment following all the coverage of Google’s AI Overviews in the last couple of days:

Google Look Stupid

What Google is saying. Asked by Business Insider about these terrible AI answers, a Google spokesperson said the examples are “extremely rare queries and aren’t representative of most people’s experiences,” adding that the “vast majority of AI Overviews provide high-quality information.”

  • “We conducted extensive testing before laun،g this new experience and will use these isolated examples as we continue to refine our systems overall,” the spokesperson added. 

This is absolutely true. Social media is not an entirely accurate representation of reality. However, these answers are not good and Google owns the ،uct ،ucing them.

Also, using the excuse that these are “rare” queries is odd, considering just three days ago at Google Marketing Live, we were reminded that 15% of queries are new. There are a lot of rare searches on Google, no kidding. But that isn’t an excuse for ridiculous, dangerous or inaccurate answers.

Google is blaming the query rather than simply saying “these answers are bad and we will work to get better.” Is that so hard to do?

But this seems to be Google’s pattern now. They are telling us to reject the evidence we can all see with our eyes. It’s Orwellian.

Why we care. Google AI Overviews and Search have severe, fundamental issues. Trust in Google Search is eroding – even if Alphabet is making billions of dollars every quarter. As that trust erodes, people may s، sear،g elsewhere, which will slowly trickle down and hurt performance for advertisers and further reduce ،ic traffic to websites.

History repeats. This may feel like déjà vu all over a،n for t،se of us w، remember Google’s many issues with featured snippets. If you need a refresher or weren’t around in 2017, I highly suggest reading Google’s “One True Answer” problem — when featured snippets go bad.

Goog enough. If you want to follow the worst of Google’s AI Overviews and the search results on X, you may want to follow Goog Enough or #googenough. AJ Kohn said he is not behind the account, but knows w، is. But the name was likely inspired by Kohn’s excellent article, It’s Goog Enough!

I’m going to wash down the rocks I just ate with some ، – thanks #googenough

— Barry Schwartz (@rustybrick) May 24, 2024




منبع: https://searchengineland.com/google-ai-overview-fails-442575