I understand Google and Microsoft getting into it as it makes sense as a “better” Google search but for StackOverflow that sounds like they have just given up on their current platform.
I’ve found it great for asking documentation questions. It saves me a ton of time having to search through documentation myself. The problem is when it encounters something it doesn’t have information on, it’ll just confidently make shit up, and if you’re not enough of an expert to recognize when that happens, you can be mislead. It still saves me time, but I use it as a recall tool to get me started when I’m learning to do something new, I’d never use the code it puts out without reading through it line by line. I’m also experienced enough to know when it’s wrong and how to refactor its examples. People new to programming could get set down the wrong path by over relying on gpt to teach them.
Yeah works well, as long as the code is rather simple and it occurred rather often in the training set. But I seldom use it currently (got a little bit more complex stuff going on). It’s good though to find new stuff (as it often introduces a new library I haven’t known yet). But actual code… I’m writing myself (tried it often, and the quality just isn’t there… and I think it even got worse over the last couple of months as also studies suggest)
I’ve gotten really good results asking chat gpt for programming help. Problem is that it’s wrong like 10% of the time, and when it’s wrong it’s very confidently incorrect. That wasn’t a problem for me because I knew when it was wrong and could course correct it and get the correct solution and it still saved me time and helped me eventually get to the right solution. But if someone who’s still getting started is trying to use chat gpt to learn, they could easily be mislead because they won’t know when its output is wrong.
Definitely depends on the type of question. I find for documentation type questions I get the 90% good answers, like how do I do something with this library, it’s good, which makes sense because that libraries documentation is probably in the training data. But for more open ended questions, like how do I solve this problem, I see similar performance to what you’re saying. I think it’s a good retrieval and synthesises tool which can really save a ton of time if you already have a high level plan of action and just use it to fill in some specific details.
I get the whole community resource and all that hoorah, but what bothers me the most is that C*O somewhere that’s padding his bonus and CV, waiting for the ship to sink so he can move on to the next thing where he can sing praises to the AI revolution.
Hah, good to know that even on [email protected] there are people who agree that stack overflow moderation is too draconian to ask questions in anymore. It’s a good resource, though, so an LLM will probably be the answer to make the knowledge base more usable without angering its elder gods.
I’m not liking the announced changes to search. That sounds like we will be losing the lexical search and in exchange we will be getting the same technology that allows google to answer questions different to the one we asked.
How many minutes between starting to use OverflowAI until we get something like “As a large language model trained by the Stack Exchange Network i can not answer duplicated questions”.
This artificial pseudointelligence exists because there’s the “gee whiz, that’s cool” of a computer talking like a person, and a bunch of hype chasers looking to cash in. Much like cryptocurrency before it, and the dot-com boom before that, there is little substance to it, and most of it will be commercially irrelevant a decade from now.
Can someone tell me what their angle is? Are user’s supposed to curate and help train the model for free? Is it just a model trained on stackoverflow data?
All their data is open so what edge do they over the already established competition.
This type of Q&A interface is very popular and stealing traffic away from sites like Google and Stack Overflow. Stack Overflow can train it on their data and has a feature where it links to every answer it pulled from. I think that’s a nice feature and like that I can troubleshoot further on my own, as AI can often hallucinate an answer or lose a piece of context I need.
Add comment