Why does ChatGPT provide inaccurate citations?
Josh Sackman
President and Cofounder at AppliedVR | Advisor and Board Member | Digital health innovation
I wanted to see how well ChatGPT understands public figures and can report biographical information.
For my article on breathwork, I asked the AI to summarize the biography of Wim Hof and to include his Guinness World Records. I was happy to see an accurate description.
I was glad to see it also cited some of the media outlets that have covered Hof, so I wanted to find some of those sources.
It seems ChatGPT has fairly poor short-term memory and didn't full grasp the request of the "him" I was referring to. It shared a number of Guinness World Records related to cold exposure that didn't seem to match the records it provided above.
Those weren't quite the news articles I was looking for, so I improved my prompt to be more specific. The response appeared to be successful, finding articles in Telegraph, The Guardian, BBC and CNN.
However, as I clicked on the links...none of them worked!
领英推荐
Interestingly, each of those publications have actually covered Wim Hof's achievements. And the titles of the articles are factually correct. But, similar to my prior experience with ChatGPT as a research assistant, none of the articles reported are real!
In searching for the articles' titles the AI provided, none of those headlines have ever been printed within those publications. Rather, it provided article names and links for plausible publications that could exist...but each was a complete fabrication.
Why Does ChatGPT provide inaccurate citations?
This led me to explore more about this limitation of ChatGPT. Despite some interesting, but highly exaggerated and sensational headlines such as "ChatCPT Is Dumber Than You Think" in the Atlantic and "AI Platforms like ChatGPT Are Easy to Use but Also Potentially Dangerous" in Scientific American, they didn't really reveal the source of the problem. It wasn't until I dove into a YCombinator Hacker News thread until I started understanding this issue better.
They key idea is that ChatGPT is a language model, and not a knowledge model. One user explains, "I've seen it referred to as 'stochastic parroting' elsewhere, and that probably gives more insight into what is happening. These large language models are trained to predict the next word for a given input. And they don't have a choice about this; they must predict the next word, even if it means that they have to make something up."
Another user expands, "These model are extremely good when they have creative freedom and are used to produce some kind of art. Poems, text in various styles, images without important details, and so on. But they fail miserably when you give them tasks that require world knowledge or precision."
So it's not that ChatGPT is dumb or dangerous, but that our expectations are misguided about how ChatGPT works and how it can currently be used.
Key Takeaways
US Marine Veteran Turned Entrepreneur: Leading unique Web Design, Event Hosting, and Non-Profit Transformations with Vibrant Visions Consulting and DDR Sanctuary.
1 年Thanks for sharing
Now complete the circle. Assign a machine reader to assess your machine written piece.
AI Innovation Leader | Chief Innovation Officer @ Infinity Nurse Case Management | Generative AI & Digital Transformation | Microsoft Certified
1 年Great insights, Josh! I totally agree that this train has left the station and will continually improve over a relatively-short time. What some thought would be impossible until decades from now is actually happening.
Developer, Innovator and AI Researcher
1 年Most peer review and documentation is behind paywalls. LLMs are trained mainly on open crawl. It has nothing of value to learn from. The same is not for GPT-3.5 as there was an inclusion of arXiv in its training data, but the sources are all outdated. That is my theory :).?
?? Startup founder innovating #healthcare and pain management with #XR tech at Softcare Studios | ?? Digital Safety & Privacy at XRSI | ?? Startup mentor & Lecturer on Emerging Technologies | PhD in Neuropharmacology
1 年In the meantime, https://www.perplexity.ai/ looks like a useful tool fixing the gap around the references.