Stack Overflow Banned ChatGPT, Will Others follow?
Michael Spencer
A.I. Writer, researcher and curator - full-time Newsletter publication manager.
Why Stack Overflow Banned ChatGPT
If you enjoy articles about A.I. at the intersection of breaking news join AiSupremacy?here. I cannot continue to write without community support. (follow the link below). For the price of a cup of coffee, Join 130 other paying subscribers.
https://aisupremacy.substack.com/subscribe
The Q&A site has been flooded with ChatGPT coding answers that look correct but often aren't
Hey Everyone,
In my article the?Dark Side of Generative A.I.?I argue (among other things) that unmitigated spam could be a problem with tools like ChatGPT. For A.I. at the intersection of coding, the lack of fact and objective verification of ChatGPT is seriously problematic.
Many sites are having to temporarily ban ChatGPT outright. The Q&A site has been flooded with ChatGPT coding answers that look correct but often aren't, with moderators calling for a halt. This is what happens when you let the public test a demo product that hasn’t been tested enough internally.
One week ago, on December 7th, 2022 Stack Overflow said it was withholding a permanent decision on AI-generated answers until after a larger staff discussion, but was taking action now due to fears that ChatGPT could be "substantially harmful" to both the org and its users.
After the honeymoon phase with ChatGPT’s not so open trial ends, we’ll have to take some time to take a hard look at if our A.I. ethics and legal system are ready for Generative A.I. conversational agents like ChatGPT. From sites like DeviantArt having to deal with text-to-image “creations” to GitHub Copilot running amuck with copyright, clearly we seem to be rushing things.
Is it quality content? Is it safe? Clearly the mods at Stack Overflow had to do something. So as the mods explained, ChatGPT simply makes it too easy for users to generate responses and flood the site with answers that seem correct at first glance but are often wrong on close examination. For OpenAI’s demo, this is a misinformation stain. Microsoft in the end will have to be accountable for the fledgling A.I. lab, which is basically its foundation model generator.
"The average rate of getting correct answers from ChatGPT is too low," Stack Overflow?said?in a policy statement entitled "ChatGPT is banned."?
ChatGPT is Breaking Trust & Safety
Stack Overflow is a community built upon trust. The community trusts that users are submitting answers that reflect what they actually know to be accurate and that they and their peers have the knowledge and skill set to verify and validate those answers.
I actually think the Stack Overflow example best demonstrates the relationship of ChatGPT and misinformation at scale. And this is important guys, “the primary problem?is that while the answers which ChatGPT produces have a high rate of being incorrect, they typically?look like?they?might?be good?and the answers are?very?easy to produce,” wrote the mods (emphasis theirs).
We know that on the censored and moderated internet, the misinformation that slips though does tend to go viral much easier, and as has been the case with ChatGPT hype. Even when it was basically fooling users into feeling like it was telling the truth. I cannot verify how many times I witnessed this in tweets and use cases, but with over 1 million users of the demo, you can imagine how much misinformation care of OpenAI has been flying around without any legal responsibility.
The mods at Stack Overflow continued:
Overall, because the average rate of?getting?correct?answers from ChatGPT is too low, the posting of answers created by ChatGPT is?substantially harmful?to the site and to users who are asking or looking for?correct?answers.
ChatGPT was released about two weeks ago and touted by OpenAI in a?blog post?as a conversational AI that can provide detailed answers to questions, as well as "answer follow-up questions, admit its mistakes, challenge inappropriate premises and reject inappropriate requests."?The problem is it doesn’t seem like OpenAI did their due diligence on how the tool would be used, misused and create spam and confusion. OpenAI has seemed very proud of its number of users (hint: free testers).
It’s all about the RLHF in GPT-3.5?
Nathan Lambert?and the team at Hugging Face wrote a nice piece about the importance of?reinforcement learning from human feedback?(RLHF) recently.However OpenAI is also getting clever at trying to protect themselves legally: they made Sam say this:
If you create a flawed demo and promote it and let Tweets (that are basically misinformation proliferate), yes Sam, it will indeed be misused in more serious contexts. But hang on a second, the owner of Twitter was also a co-founder of OpenAI. Major conflict of interest! Oh well, not legally binding, fair use right?!
Remember guys even before Generative A.I. hits its stride Web 3.0 is basically unable to moderate itself even with A.I. doing the heavy lifting. Between January 1 and June 30, more than 21 million fake accounts were detected and removed from LinkedIn, according to the company’s community report. That doesn’t stop me from encounting dozens each week. We simply won’t be able to moderate, regulate or enforce the spam of tools like ChatGPT. There’s simply no way! And it’s a problem.
While we hype Generative A.I. on one hand, we won’t be able to even moderate it properly on the other. So what will be the result? The Western internet will continue to degrade. Free speech will turn into an A.I. playground. The dials of censorship and spam won’t really work any longer in way that doesn’t feel like total chaos, and even in 2022 we’ve had a dose of what that might feel like moving forwards.
For Silicon Valley tycoons, it’s their job to promote A.I. at scale. Whether you are related to?a PayPal mafia?is beside the point.
The mods at Stack Overflow a serious site for coders, realized quickly that ChatGPT was a spam problem. They?quickly needed to reduce the volume of these posts?and needed to be able to deal with the ones which are posted quickly, which means dealing with users, rather than individual posts. So, for now, the use of ChatGPT to create posts there on Stack Overflow is not permitted. If a user is believed to have used ChatGPT after this temporary policy is posted,?sanctions will be imposed to prevent users?from continuing to post such content, even if the posts would otherwise be acceptable.
Can you realize the nightmare this is going to create from everything from Canva to Adobe? Who can integrate text-to-image the best and the greatest? The Generative A.I. use of tools, coding and text-to-video on the mobile short-form internet becomes even more vague and gets even more dystopian with misinformation as we head deeper down the rabbit-hole of deep fakes.
Curiously that Generative A.I. is a spam problem is mostly totally unaddressed in the mainstream media or by the actors themselves. But my Twitter, Reddit, Discord and LinkedIn feeds are full of very bad examples of how ChatGPT is used even as a flawed demo today that highlights the scale of the issues created.
In early November, 2022 DeviantArt, a Wix-owned artist community, announced a new protection for creators to prevent their work from being scraped for AI development. It’s not just artists that are being ripped off and who gets to benefit? Microsoft recently announced that GitHub Copilot for Business would go live at $19. I was surprised to see the price so much higher than an individual account and?what that actually included. The legal grey areas here are significant, and BigTech can bulldoze A.I. adoption at scale, before regulation and moderation is even remotely ready.
Stack Overflow really put it well when it said:
The objective nature of the content on Stack Overflow means that if any part of an answer is wrong, then the answer is objectively wrong. In order for Stack Overflow to maintain a strong standard as a reliable source for correct and verified information, such answers must be edited or replaced.
The Internet is About to Enter a Deepfakes and Misinformation Era
Increasingly as users of the internet we have to trust our sources less and do more fact-checking ourselves. But with Generative A.I. and tools like chatGPT it’s the illusion of truth-sounding content that begins to blur, just as deepfakes get better at producing believable fake accounts that even Microsoft’s LinkedIn A.I. cannot catch them upon account creation.
Between January 1 and June 30, more than 21 million fake accounts were detected and removed from LinkedIn, according to the company’s community report. But there’s always about 4 percent that get though, and that turns out to be a lot. While 95.3% of those fake accounts were stopped at registration by automated defenses, according to the company, there was a nearly 28% increase in fake accounts caught compared to the previous six-month period. LinkedIn says it currently has more than 875 million members on its platform. The deepfake internet is clearly just beginning.
Even the internet doesn’t understand what’s coming. Generative A.I. tools are already becoming controversial like Lensa A.I. that?can sexualize users without their consent. But what happens when you believe the source to be accurate as is the case so often with ChatGPT? Who is liable? OpenAI doesn’t seem to think it is liable for much of anything, and that’s a problem.
You can read the rest of the article first published here.
What do you think, will Generative AI make a more fair and truthful place for the internet?
If you enjoy articles about A.I. at the intersection of breaking news join AiSupremacy?here. I cannot continue to write without community support. (follow the link below). For the price of a cup of coffee, Join 130 other paying subscribers.
https://aisupremacy.substack.com/subscribe
Data scientist || GenAI Engineer || Data Artist || Data Science || Machine Learning || GenAI || Azure
1 年#chatgpt3 #openai
Without data you are just another person with an opinion!
1 年Stack Overflow always bans something and if there is nothing left to ban they probably ban themselves. Should be ?Stack Overmoderated“
Senior Software Developer
1 年We have build a fascination for AGI, so what ever comes in the market got hype, #chatgpt has similar fate, people are expecting alot from this new baby where its in its infancy stage and too far from AGI. As of now, nothing got replaced by any AI thing, only we can get the assistance.
Director of Product Management at Iguazio | Data Science Platform | MLOps
1 年It's ridiculous to ban, the discussion should be about understanding the boundaries of this technology and agreeing on the ethics. I also doubt this is any threat to Google. They have deepmind and they are the ones who invented the transformer NN architecture (https://ai.googleblog.com/2017/08/transformer-novel-neural-network.html). It's actually an opportunity for them. Now that other teams are breaking new grounds, the world is getting familiarized with the capabilities and Google can evaluate the overall reaction and sentiment. Hence they can build a model that will augment their search service and other services.
A.I. Writer, researcher and curator - full-time Newsletter publication manager.
1 年I wonder if Stack Overflow could give us more context.