• Email Us: [email protected]
  • Contact Us: +1 718 874 1545
  • Skip to main content
  • Skip to primary sidebar

Medical Market Report

  • Home
  • All Reports
  • About Us
  • Contact Us

Google And Bing’s AI Chatbots Appear To Be Citing Each Other’s Lies

March 24, 2023 by Deborah Bloomfield

Bing’s artificial intelligence (AI) chatbot appears to have cited lies originally made by Google’s rival chatbot Bard, in a worrying example of how misinformation could be spread by the new large language models.

It’s been a tough week for AI chatbots, so it’s probably a good thing they don’t have feelings. Google launched its new chatbot Bard to the public, with a fairly bumpy few first days.

Advertisement

“The launch of Bard has been met with mixed reactions,” Bard told IFLScience, pooling as it does information from around the web. “Some people are excited about the potential of Bard to be a powerful tool for communication and creativity, while others are concerned about the potential for Bard to be used for misinformation and abuse.”

It’s funny that Bard should mention that aspect, as it has already performed a few notable hallucinations – delivering a confident and coherent response that has no correlation with reality. 

One user, freelance UX writer and content designer Juan Buis, discovered that the chatbot believed that it had already been shut down due to a lack of interest.

Buis discovered that the sole source that the bot had used for this information was a 6-hour-old joke comment on Hacker News. Bard went on to list more reasons that Bard (which has not been shut down) had been shut down, including that it didn’t offer anything new or innovative.

Advertisement

“Whatever the reason, it is clear that Google Bard was not a successful product,” Google Bard added. “It was shut down after less than six months since its launch, and it is unlikely that it will ever be revived.”

Now, as embarrassing as the error was, it was fixed fairly quickly and could have ended there. However, a number of websites wrote stories about the mistake, which Bing’s AI chatbot then wildly misinterpreted.

As spotted by The Verge, Bing then picked up on one of the articles, misinterpreted it, and started telling users that Bard had been discontinued.

The error has now been fixed, but it’s a reasonable look at how these new chatbots can and do go wrong, as people begin to rely on them more for information. As well as hallucinating information, chatbots may now end up sourcing information based on the hallucinations and mistakes of other chatbots. It could get really messy.

Deborah Bloomfield
Deborah Bloomfield

Related posts:

  1. Rallying – Toyota’s Rovanpera leads in Greece after day one
  2. Binter and Iberia cancel flights to La Palma due to volcanic ash
  3. Trudeau’s return to power with big spending plans could fuel Canada’s hot inflation
  4. Tennis-Murray makes online plea to help find stolen wedding ring

Source Link: Google And Bing's AI Chatbots Appear To Be Citing Each Other's Lies

Filed Under: News

Primary Sidebar

  • Musk Outlines The Questionable Reason He Wants To Get To Mars So Badly, NASA Astronaut Responds
  • In 1972 The Soviets Launched A Spacecraft Bound For Venus. In The Next Few Days, It Will Return To Earth
  • Sounds From Inside A Star Reveal Unexpected Properties Of An Aging Orange Dwarf
  • Hear An Elephant Reunion Spark Sounds Even Keepers Had Not Heard Before
  • Why Do Elevators Have Mirrors Inside Them?
  • Cuttlefish Communicate With Arm Waving And Can Sense The Ripples With Their Bodies
  • First Ever Fatal Bear Attack In Florida Leads To The Deaths Of 3 Black Bears
  • Pathogenic Fungal Spores Found Surviving Miles Above Our Heads In Earth’s Stratosphere
  • “Alchemy” In Action As CERN Detects Lead Atoms Turning Into Gold
  • When Did The Earth’s Magnetic Field Form?
  • Who Were The Mysterious “Sea Peoples”, Destroyers Of The Ancient Empires?
  • Galaxy’s Extreme Core Might Have A Whole New Source Of Ghostly Particles
  • 20 Years Of “Very Concerning” Data Concludes Cats Can Catch Bird Flu And Could Pass It To Humans
  • The Ancient Pythagorean “Cup Of Justice” Pranks Users If They Fill It With Too Much Wine
  • When It Comes To Pain, The Nocebo Effect Beats The Placebo Effect
  • English Speakers Obey This Quirky Grammar Rule, Even If They Don’t Know It
  • How Is The Black, White, And Secret Third Smoke Made During The Conclave?
  • Can Children Help Each Other Pass The Famous Marshmallow Test?
  • California’s Highest-Altitude Tree Found By Happy Accident At 12,657 Feet
  • Is The Spiny Devil Katydid The Strangest Insect In The World? You Tell Us
  • Business
  • Health
  • News
  • Science
  • Technology
  • +1 718 874 1545
  • +91 78878 22626
  • [email protected]
Office Address
Prudour Pvt. Ltd. 420 Lexington Avenue Suite 300 New York City, NY 10170.

Powered by Prudour Network

Copyrights © 2025 · Medical Market Report. All Rights Reserved.

Go to mobile version