Study Finds That 52 Percent of ChatGPT Answers to Programming Questions Are Wrong

ekZepp@lemmy.world to Technology@lemmy.ml – 600 points –
Study Finds That 52 Percent of ChatGPT Answers to Programming Questions Are Wrong
futurism.com
128

You are viewing a single comment

Sure, but by randomly guessing code you'd get 0%. Getting 48% right is actually very impressive for an LLM compared to just a few years ago.

Just useful enough to become incredibly dangerous to anyone who doesn't know what they're doing. Isn't it great?

Now non-coders can finally wield the foot-gun once reserved only for coders! /s

Truth be told, computer engineering should really be something that one needs a licence to do commercially, just like regular engineering. In this modern era where software can be ruinous to someone's life just like shoddy engineering, why is it not like this already.

Look, nothing will blow up if I mess up my proxy setup on my machine. I just won't have internet until I revert my change. Why would that be different if I were getting paid for it?

Nothing happens if you fuck up your proxy, but if you develop an app that gets very popular and don't care about safety, so hackers are able to take control over your whole Server they can do a lot of damage. If you develop software for critical infrastructure it can actually cost human lives if you fuck up your security systems.

Yes, but people with master's degrees also fuck this up, so it's not like some accreditation system will solve the issue of people making mistakes

Yeah, but its probably more likely that the untaught might fuck up some stuff.

Is it, though? A lot of self-taught programmers do great work. I'm not sure this is true

Setting up proxy is not engineering.

I have to actually modify the code to properly package it for my distro, so it's engineering because I have to make decisions for how things work

I don't see how this supports your point then. If "setting up proxy" means "packaging it to run on thousands user machines" then isn't there obvious and huge potential for a disastrous fuckup?

No, because it either runs the program successfully, or it fails to launch. I don't mess with the protocol. It runs as root because it needs to set the iptables when turned on to be a "global" proxy

It’s pretty fun, interesting times ahead. I wonder what kind of bullshit will take place and can’t wait to see that lol. Between all the climate, ai, warmongering future won’t be boring guys that is certain. Unpack your popcorn

Exactly, I also find that it tends to do a pretty good job pointing you in the right direction. It's way faster than googling or going through sites like stackoverflow because the answers are contextual. You can ask about a specific thing you want to do, and and an answer that gives you a general idea of what to do. For example, I've found it to be great for crafting complex sql queries. I don't really care if the answer is perfect, as long as it gives me an idea of what I need to do.

You can also play with it to try and get closer to correct. I had problems with getting an Excel macro working and getting unattended-updates working on my pihole. GPT was wrong at first, but got me partly there and I could massage the question and Google and get closer to the right answer. Without it, I wouldn't have been able to get any of it, especially with the macro.