Or I can spend a much shorter amount of time formulating a question for Chat-GPT and generally get a helpful, focused answer without any pedantic digressions.
It seems likely that the AI benefits from the information in SO. If Open AI can help improve the SO experience that would be fantastic.
You don't go to SO to crowdsource creative ideas. It's for very specific one-off questions that many people will likely find themselves asking at some point.
For some reason, they don't. Honestly, I don't understand why, but there is a cohort of people out there who are ok with it.
(Well, you can stop using ChatGPT, and that's what I ended up doing. General idea or inspiration? Sure, I can ask it. Specific technical question? Nope, google it is)
I think both SO and OpenAI see the writing on the wall (unfortunately). The real "partnership" is OpenAI gets to say "look, we're working together!" to avoid accusations of destroying SO, and SO gets to save a little bit of face (and hopefully make a little money) on the way down.
https://www.reddit.com/r/programming/comments/1592s82/the_fa...
I think it boils down to more of "Hey, we can criticize StackOverflow since we're on the inside... but if someone attacks from the outside, we have its back."
> Docker for Windows won't run if you have the Razer Synapse driver management tool running.
https://twitter.com/Foone/status/1229641258370355200
:)
Edit: The reason was that both software directly copied something from stackoverflow.
This already has meant SO dropped out of relevance for anything that's long-lived but evolving. I assume it still works for brand-new stuff where there are no apparent duplicates. It works for unchanging old stuff (and the absolute basics of programming), because the old answers are still relevant. But take anything like Java, C#, Python, or Javascript that have evolved radically since SO's inception and the answers are often garbage.
IMHO, SO needs to solve this to not die... if it isn't already too late.
I can't tell from the article, but a logical use of AI on SO would be to answer questions, tailored to each user, just like people do with ChatGPT etc today. However this means there's now no new questions even feeding in, let alone new/updated answers. So the training data for the AI becomes increasingly out-of-date/wrong. I don't see how this solves the existential problem SO has, but maybe it will delay their demise a bit.
There are also answers that "work" and aren't insecure but will near certainly cause other issues.
I'm sure some people upvote because they had the same question, tried the solution, and it seemingly worked (even if it's not secure, performant, etc.), so they upvote. But you'd think they'd at least check the comments and see what people are saying before trying (let alone upvoting) a solution.
Even worse is the outdated information
Microsoft, `Open`AI, Github, LinkedIn, Stackoverflow .. Feels like it will end badly.
This will be interesting
Laundering human responses via a large language model not only makes it impossible to acknowledge SO contributors: it encourages people to think GPT figured these things out solely because it's simply so darn clever.
It doesn't help that SO's marketing is encouraging developers to not care about integrity or professionalism:
> provide OpenAI users and customers with the accurate and vetted data foundation that AI tools need to quickly find a solution to a problem so that technologists can stay focused on priority tasks.
Hey buddy, you got priority tasks to focus on. Just let the plagiarism robot do its thing.
Stackoverflow.com is one (most popular/biggest) of them.
While at this, here is the list of all communities (they are quite cool! do browse a few): https://stackexchange.com/sites
In this day and age of phishing using domains like that is not really the smartest thing to do I would say...
edit: Actually I've gone ahead and just started deleting everything. I realize they're already part of the dataset, but my goal is to hurt Stack Overflow (ever so slightly) for this decision.
I find it hard to imagine that AI will need humans to teach it technologies like programming languages and APIs for long.
We don't need humans to teach computers how to play chess anymore.
They did at one point turn off the data dumps, early in the AI in fact and likely because they wanted to sell the data. But they were reinstated after massive backlash [3]. They could do this again and make future content exclusive. But haven’t done so yet, and if they do, it will be very public.
[1] https://meta.stackexchange.com/questions/344491/an-update-on....
[2] https://data.stackexchange.com
[3] https://meta.stackexchange.com/questions/389922/june-2023-da...
F SO
And no, buying the rights after you've already stolen all the data to make billions is not acceptable.
They broke the law on a grand scale, used this to make shitloads of money, and are now trying to use that money to pay off anyone that might give them trouble.
Classic mob mentality.
And I'd be remiss if I didn't point out that their trade dress is MIT licensed. https://stackoverflow.design
Have fun.
In the past, the state of the community has already made me to use Stack Exchange as the last resort, and this move completely closes the doors.
This hellscape is forming way too fast.
It doesn't indicate that generative AI is going to be shoehorned into StackOverflow's websites. It would seem counterproductive, in fact, to do that, since the gist of this seems to be that StackOverflow provides a large wealth of organized, validated human-generated knowledge, which is exactly the sort of thing you want to train LLMs on. Feeding AI-generated data back into that would diminish the value of the data SO hosts for that purpose.
> provide attribution to the Stack Overflow community within ChatGPT
...and that didn't seem important enough for OpenAI to bother to mention it on any of their media channels that I've seen.
so, who knows?
It feels like it's a whole lot of nothing to me, and exchange they're letting OpenAI having all of their Q/A data.
I doubt it will make any significant difference to S/O for most people; and anyone who thinks putting S/O links in a chatGPT response is going to drive traffic back to S/O is kiddddddddddding themselves.
On top of this, you could say the same about any disrupting technology.
I use OpenAI because StackOverflow answers are just the absolute wrong answer. A combination of gaslighting (you shouldn't be having this problem), dogmatic enforcement of good ideas that started as guidelines and problematic example code that should not be trusted. You are better of with a reddit thread or a blogpost and much better of with actual documentation. StackOverflow is the thing that causes the bugs and the tech debt in the first place.
At least now OpenAI's competition has a fighting chance, because their models won't be poisoned by SO
I would have thought that OpenAI had already trained off of SO data. Does anybody know if this is the case?
If they did, then they broke (or, I guess charitably, dodged the question of) copyright law in their training, got first mover advantage with the results, and now they can go back to the copyright holders to "partner" with them after the fact to prevent others from doing the same thing?
An AI being able to consistently outperform us in recalling the syntax for switch statements, is a world away from "all of our basic needs being taken care of by automation". The former is going to take a few more weeks/months, while the latter is going to take a few more decades/centuries.
In the interim, there will be some winners, and many losers from this innovation. Wealth will concentrate significantly towards the winners, while the losers will be out of work with a valueless skillset, and their basic needs going unmet. While this may be true for most high-skill professions in the coming decades, there's a unique irony for programmers - who will be the losers, having invented and then fueled the engine of their own demise on behalf of the winners.
It's not necessarily a value-judgement based comment. It's just noting the irony, and highlighting that it's a specific genre of irony that economists absolutely salivate over.
Haven't we been promised this for literally a century? We don't even have a four-day workweek.
Before that happens, so many other professions shall then have been rendered totally obsolete. So many it'd have profound societal consequences. I understand the "me, myself and I" and the fear but programmers coding themselves into irrelevance is really the least of our concerns.