this post was submitted on 28 Jan 2024
1045 points (97.6% liked)

Programmer Humor

19480 readers
202 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 113 points 9 months ago (6 children)

lol, I'd love to see the fucking ruin of the world we'd live in if current LLMs replaced senior developers. Maybe it'll happen some day, but in the meantime it's job security! I get to fix all of the bugfuck crazy issues generated by my juniors using Copilot and ChatGPT.

[–] [email protected] 65 points 9 months ago (1 children)

So much hallucinated crap and shoddy answers. Just because it was AI generated doesn't mean it was a good solution

[–] [email protected] 28 points 9 months ago (2 children)

There is a reason Microsoft has branded it copilot...

[–] [email protected] 14 points 9 months ago

Headline: airline fires every second pilot; says the copilot is good enough to fly the machine.

[–] [email protected] 4 points 9 months ago

...because Tesla had already made the mistake of over promising with the "autopilot" name?

[–] [email protected] 31 points 9 months ago* (last edited 9 months ago) (1 children)

One of my uni lecturers does the whole "You are out of a job" thing. He's a smart guy but he's barley written a line of code in his life. This comes up frequently and everytime I ask him "Get CHATGPT to write fizz buzz in X86 ASM." Without fail it will crash when trying to build everytime. This technology is very advanced but I find people get it to the the simplest tasks and then expect it to solve the most complex ones.

[–] [email protected] 19 points 9 months ago (1 children)

I tried using AI tools to do some cleanup and refactoring of some legacy embedded C code and was curious if it could do any optimization or knew any clever algorithms.

It's pretty good at figuring out the function of the code and adding comments, it did some decent refactoring of some sections to make them more readable.

It has no clue about how to work in a resource constrained environment or about the main concepts that separate embedded from everything else. Namely that it has to be able to run "forever", operate in realtime on a constant flow of sensor data, and that nobody else is taking care of your memory management.

It even explained to me that we could do input filtering by using big arrays to do simple averaging on a device with only 1kB RAM, or use a long long for a never-reset accumulator without worrying about what will happen because "it will be years before it overflows".

AI buddy, some of these units have run for decades without a power cycle. If lazy coders start dumping AI output into embedded systems the whole world is going to get a lot more glitchy.

[–] [email protected] 13 points 9 months ago (2 children)

This is how AI is a threat to humanity. Not because it will choose to act against us, but because people will trust what it says without question and base huge decisions on faulty information.

[–] [email protected] 8 points 9 months ago

A million tiny decisions can be just as damaging. In my limited experience with several different local and cloud models you have to review basically all output as it can confidently introduce small errors. Often code will compile and run, but it has small errors that can cause output to drift, or the aforementioned long-run overflow type errors.

Those are the errors that junior or lazy coders will never notice and walk away from, causing hard to diagnose failure down the road. And the code "looks fine" so reviewers would need to really go over it with a fine toothed comb, which only happens in critical industries.

I will only use AI to write comments and documentation blocks and to get jumping off points for algorithms I don't keep in my head. ("Write a function to sort this array") It's better than stack exchange for that IMO.

[–] [email protected] 1 points 9 months ago* (last edited 9 months ago)

Maybe the real "AI nuking the world" scenario was that it ie caused by the faulty information the AI hallucinated into existence

[–] [email protected] 27 points 9 months ago

I was helping someone with their programming homework, every time copilot suggested anything he just blindly added it, and every time i had to ask him "and why do you need those lines? What do they do?", and he could never answer...

Sometimes those lines made sense, other times they were completely irrelevant to the problem, but he just add the suggestions on reflex without even reading them

[–] [email protected] 25 points 9 months ago (2 children)

It'll be like when we were all supposed to lose our jobs to outsourcing

[–] [email protected] 8 points 9 months ago

And when "web frameworks means we don't need web developers anymore" and when "COBOL is basically plain English, so anyone can code, so we don't need specialists anymore".

[–] [email protected] 5 points 9 months ago

Millions did. It's just that after a while the advantages stopped being convincing and the trend reversed. If the same thing happens here, expect to go jobless for a while until you're needed again.

[–] [email protected] 18 points 9 months ago

And those juniors don't realize they've set themselves up to be forever-juniors since they aren't learning how to do the basics themselves.

[–] [email protected] 16 points 9 months ago

I had to pull aside a developer to inform him that he "would be" violating our national security by pasting code online to an AI and that there were potentially repercussions far beyond his job.

He's a lot slower now, but the code is better.