Digital Digging with Henk van Ess

Digital Digging with Henk van Ess

Share this post

Digital Digging with Henk van Ess
Digital Digging with Henk van Ess
#ChatGPT : 2. Fighting misinformation
Copy link
Facebook
Email
Notes
More

#ChatGPT : 2. Fighting misinformation

AI thinks my goldfish can play the triangle, but it CAN be a good debunker too

Henk van Ess's avatar
Henk van Ess
Apr 25, 2023
∙ Paid
6

Share this post

Digital Digging with Henk van Ess
Digital Digging with Henk van Ess
#ChatGPT : 2. Fighting misinformation
Copy link
Facebook
Email
Notes
More
1
Share
Original artwork digitally made by Henk van Ess with MidJourney

Experts claim that ChatGPT can hallucinate and create false information. It fools scientists, writes fake articles, produces medical data, and lies about people

I know about the AI’s bad behavior. I claimed that my goldfish can play the triangle better than a human, ChatGPT agrees: playing the triangle is definitely within their physical capabilities.

My brilliant idea is that ChatGPT can be a trustworthy debunker. That sounds foolish and naive. But hear me out.

ChatGPT is an LLM (large language model) that has been trained on a massive dataset of text from the internet, which includes information from reputable sources such as news articles, academic papers, and government reports. This means that when information is already in the public domain, there is a higher likelihood that ChatGPT has already encountered and processed it, making it easier for the model to identify false or misleading claims and provide accurate information.

Again, made by MidJourney, by Henk van Ess, typing in keywords

My claim “My GoldFish Plays Triangle!” isn’t part of any theory, yet, hence the AI starts hallucinating. Yes, it is possible to train a goldfish to follow your finger. Goldfish are curious and will often investigate new objects in their environment. But the moment #chatgpt has to process something completely new, it can behave like a clown. Only when it has already some data about your topic, it can shine.

ChatGPT can quickly spot the BS and set the record straight with precision and clarity, especially when the outrageous claims have already been called out by someone outside the echo chamber of true believers. Here are a few examples I tested and posted today.

Scientists can't understand why Antarctica hasn't warmed in 70 years despite CO2 rise. Wow, who would have thought that a group with "Constructive" in their name would spread misinformation? And yet, someone still thought it was a good idea to share their climate change conspiracy theories. Thank goodness for fact-checkers & .... #chatgpt!

Image

Has Kristen a point, do you think?

No she hasn't, or at least it’s not proven yet.

Image

And can I trust thedogisland.com?

No you can’t.

This looks like a neutral and rational site, right?

No it doesn't.

This Instagram posting, is it true?

No it isn’t.

Especially when the outlandish claims have already been debunked by someone who's not drinking the same Kool-Aid as the true believers, #Chatgpt has a role to play.

Swiss students did an interesting test. Their study evaluating ChatGPT's fact-checking performance found that it accurately categorized statements in 72% of cases, with greater accuracy at identifying true claims than false claims. ChatGPT's potential to help label misinformation is demonstrated, but it should never replace human fact-checking experts' crucial work in upholding information accuracy.

Keep reading with a 7-day free trial

Subscribe to Digital Digging with Henk van Ess to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 𝚑𝚎𝚗𝚔 𝚟𝚊𝚗 𝚎𝚜𝚜
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More