Amazon cloud boss echoes NVIDIA CEO on coding being dead in the water: "If you go forward 24 months from now, it's possible that most developers are not coding"

floofloof@lemmy.ca to Technology@lemmy.world – 319 points –
Amazon cloud boss echoes NVIDIA CEO's of coding being dead in the water with the rapid prevalence of AI sentiments: "If you go forward 24 months from now, it's possible that most developers are not coding"
windowscentral.com
209

The only people who would say this are people that don’t know programming.

LLMs are not going to replace software devs.

Wrong, this is also exactly what people selling LLMs to people who can't code would say.

It's this. When boards and non-tech savvy managers start making decisions based on a slick slide deck and a few visuals, enough will bite that people will be laid off. It's already happening.

There may be a reckoning after, but wall street likes it when you cut too deep and then bounce back to the "right" (lower) headcount. Even if you've broken the company and they just don't see the glide path.

It's gonna happen. I hope it's rare. I'd argue it's already happening, but I doubt enough people see it underpinning recent lay offs (yet).

AI as a general concept probably will at some point. But LLMs have all but reached the end of the line and they're not nearly smart enough.

LLMs have already reached the end of the line 🤔

I don’t believe that. At least from an implementation perspective we’re extremely early on, and I don’t see why the tech itself can’t be improved either.

Maybe it’s current iteration has hit a wall, but I don’t think anyone can really say what the future holds for it.

LLMs have been around since roughly 2016 2017 (comment below corrected me that Attention paper was 2017). While scaling the up has improved their performance/capabilities, there are fundamental limitations on the actual approach. Behind the scenes, LLMs (even multimodal ones like gpt4) are trying to predict what is most expected, while that can be powerful it means they can never innovate or be truth systems.

For years we used things like tf-idf to vectorize words, then embeddings, now transformers (supped up embeddings). Each approach has it limits, LLMs are no different. The results we see now are surprisingly good, but don't overcome the baseline limitations in the underlying model.

The "Attention Is All You Need" paper that birthed modern AI came out in 2017. Before Transformers, "LLMs" were pretty much just Markov chains and statistical language models.

I'm not trained in formal computer science, so I'm unable to evaluate the quality of this paper's argument, but there's a preprint out that claims to prove that current computing architectures will never be able to advance to AGI, and that rather than accelerating, improvements are only going to slow down due to the exponential increase in resources necessary for any incremental advancements (because it's an NP-hard problem). That doesn't prove LLMs are end of the line, but it does suggest that additional improvements are likely to be marginal.

Reclaiming AI as a theoretical tool for cognitive science

we're extremely early on

Oh really! The analysis has been established since the 80's. Its so far from early on that statement is comical

Transformers, the foundation of modern "AI", was proposed in 2017. Whatever we called "AI" and "Machine Learning" before that was mostly convolutional networks inspired by the 80's "Neocognitron", which is nowhere near as impressive.

The most advanced thing a Convolutional network ever accomplished was DeepDream, and visual Generative AI has skyrocketed in the 10 years since then. Anyone looking at this situation who believes that we have hit bedrock is delusional.

From DeepDream to Midjourney in 10 years is incredible. The next 10 years are going to be very weird.

"at some point" being like 400 years in the future? Sure.

Ok that's probably a little bit of an exaggeration. 250 years.

I can see the statement in the same way word processing displaced secretaries.

There used to be two tiers in business. Those who wrote ideas/solutions and those who typed out those ideas into documents to be photocopied and faxed. Now the people who work on problems type their own words and email/slack/teams the information.

In the same way there are programmers who design and solve the problems, and then the coders who take those outlines and make it actually compile.

LLM will disrupt the programmers leaving the problem solvers.

There are still secretaries today. But there aren't vast secretary pools in every business like 50 years ago.

It'll have to improve a magnitude for that effect. Right now it's basically an improved stack overflow.

...and only sometimes improved. And it'll stop improving if people stop using Stack Overflow, since that's one of the main places it's mined for data.

Nah, it's built into the editors and repos these days.

?

If no one uses Stack Overflow anymore, then no one posts new answers. So AI has no new info to mine.

They are mining the IDE and GitHub.

You seem to be missing what I'm saying, and missing my point. But I'm not going to try to rephrase it again.

There is no reason to believe that LLM will disrupt anyone any time soon. As it stands now the level of workmanship is absolutely terrible and there are more things to be done than anyone has enough labor to do. Making it so skilled professionals can do more literally just makes it so more companies can produce quality of work that is not complete garbage.

Juniors produce progressively more directly usable work with reason and autonomy and are the only way you develop seniors. As it stands LLM do nothing with autonomy and do much of the work they do wrong. Even with improvements they will in near term actually be a coworker. They remain something you a skilled person actually use like a wrench. In the hands of someone who knows nothing they are worth nothing. Thinking this will replace a segment of workers of any stripe is just wrong.

I wrote a comment about this several months ago on my old kbin.social account. That site is gone and I can't seem to get a link to it, so I'm just going to repost it here since I feel it's relevant. My kbin client doesn't let me copy text posts directly, so I've had to use the Select feature of the android app switcher. Unfortunately, the comment didn't emerge unscathed, and I lack the mental energy to fix it due to covid brain fog (EDIT: it appears that many uses of I were not preserved). The context of the old post was about layoffs, and it can be found here: https://kbin.earth/m/asklemmy@lemmy.ml/t/12147

I want to offer my perspective on the Al thing from the point of view of a senior individual contributor at a larger company. Management loves the idea, but there will be a lot of developers fixing auto-generated code full of bad practices and mysterious bugs at any company that tries to lean on it instead of good devs. A large language model has no concept of good or bad, and it has no logic. happily generate string- templated SQL queries that are ripe for SQL injection. I've had to fix this myself. Things get even worse when you have to deal with a shit language like Bash that is absolutely full of God awful footguns. Sometimes you have to use that wretched piece of trash language, and the scripts generated are horrific. Remember that time when Steam on Linux was effectively running rm -rf /* on people's systems? I've had to fix that same type of issue multiple times at my workplace.

I think LLMs will genuinely transform parts of the software industry, but I absolutely do not think they're going to stand in for competent developers in the near future. Maybe they can help junior developers who don't have a good grasp on syntax and patterns and such. I've personally felt no need to use them, since spend about 95% of my time on architecture, testing, and documentation.

Now, do the higher-ups think the way that do? Absolutely not. I've had senior management ask me about how I'm using Al tooling, and they always seem so disappointed when I explain why I personally don't feel the need for it and what feel its weaknesses are. Bossman sees it as a way to magically multiply IC efficiency for nothing, so absolutely agree that it's likely playing a part in at least some of these layoffs.

Basically, I think LLMs can be helpful for some folks, but my experience is that the use of LLMs by junior developers absolutely increases the workload of senior developers. Senior developers using LLMs can experience a productivity bump, but only if they're very critical of the output generated by the model. I am personally much faster just relying on traditional IDE auto complete, since I don't have to change from "I'm writing code" mode to "I'm reviewing code mode."

The one colleague using AI at my company produced (CUDA) code with lots of memory leaks that required two expert developers to fix. LLMs produce code based on vibes instead of following language syntax and proper coding practices. Maybe that would be ok in a more forgiving high level language, but I don't trust them at all for low level languages.

I was trying to use it to write a program in python for this macropad I bought and I have yet to get anything usable out of it. It got me closer than I would have been by myself and I don't have a ton of coding experience so it's problems are probably partially on me but everything it's given me has required me to correct it to work.

Will there even be a path for junior level developers?

The same one they have now, perhaps with a steeper learning curve. The market for software developers is already saturated with disillusioned junior devs who attended a boot camp with promises of 6 figure salaries. Some of them did really well, but many others ran headlong into the fact that it takes a lot more passion than a boot camp to stand out as a junior dev.

From what I understand, it's rough out there for junior devs in certain sectors.

The problem with this take is the assertion that LLMs are going to take the place of secretaries in your analogy. The reality is that replacing junior devs with LLMs is like replacing secretaries with a network of typewriter monkeys who throw sheets of paper at a drunk MBA who decides what gets faxed.

I'm saying that devs will use LLM's in the same way they currently use word processing to send emails instead of handing hand written notes to a secretary to format, grammar/spell check, and type.

I thought by this point everyone would know how computers work.

That, uh, did not happen.

I don't know if you noticed but most of the people making decisions in the industry aren't programmers, they're MBAs.

Irrelevant, anyone who tries to replace their devs with LLMs will crash and burn. The lessons will be learned. But yes, many executives will make stupid ass decisions around this tech.

It's really sad how even techheads ignore how rapidly LLM coding has come in the last 3 years and what that means in the long run.

Just look how rapidly voice recognition developed once Google started exploiting all of its users' voice to text data. There was a point that industry experts stated 'There will never be a general voice recognition system that is 90%+ across all languages and dialects.' And google made one within 4 years.

The natural bounty of a no-salary programmer in a box is too great for this to ever stop being developed, and the people with the money only want more money, and not paying devs is something they've wanted since the coding industry literally started.

Yes its terrible now, but it is also in its infancy, like voice recognition in the late 90s it is a novelty with many hiccoughs. That won't be the case for long and anyone who confidently thinks it can't ever happen will be left without recourse when it does.

But that's not even the worst part about all of this but I'm not going into black box code because all of you just argue stupid points when I do but just so you know, human programming will be a thing of the past outside of hobbyists and ultra secure systems within 20 years.

Maybe sooner

Maybe in 20 years. Maybe. But this article is quoting CEOs saying 2 years, which is bullshit.

I think it’s just as likely that in 20 years they’ll be crying because they scared enough people away from the career that there aren’t enough developers, when the magic GenAI that can write all code still doesn’t exist.

The one thing that LLMs have done for me is to make summarizing and correlating data in documents really easy. Take 20 docs of notes about a project and have it summarize where they are at so I can get up to speed quickly. Works surprisingly well. I haven’t had luck with code requests.

That’s not what was said. He specifically said coding.

It’ll replace brain dead CEOs before it replaces programmers.

I'm pretty sure I could write a bot right now that just regurgitates pop science bullshit and how it relates to Line Go Up business philosophy.

Edit: did it, thanks ChatJippity

def main():
    # Check if the correct number of arguments are provided
    if len(sys.argv) != 2:
        print("Usage: python script.py ")
        sys.exit(1)
    # Get the input from the command line
    PopScienceBS = sys.argv[1]
    # Assign the input variable to the output variable
    LineGoUp = PopScienceBS
    # Print the output
    print(f"Line Go Up if we do: {LineGoUp}")
if __name__ == "__main__":
    main()
if lineGoUp {

    CollectUnearnedBonus()

} else {

   FireSomePeople()
   CollectUnearnedBonus()

}

I think we need to start a company and commence enshittification, pronto.

This company - employee owned, right?

I'm just going to need you to sign this Contributor License Agreement assigning me all your contributions and we'll see about shares, maybe.

I love how even here there's line metric coding going on

I know just enough about this to confirm that this statement is absolute horseshit

Sounds like the no-ops of a decade ago and cloud will remove the need for infrastructure engineers. 😂🤣😂🤣😂🤣😂😂😂🤣

SHUT UP AND GO BACK TO OUR SHITTY YAML BASED INFRASTRUCTURE!

It isn't that AI will have replaced us in 24 months, it's that we will be enslaved in 24 months. Or in the matrix. Etc.

Will the matrix it puts us in be in 1999? Because I'd take that deal.

Matrix lookin pretty good rn - 1999, stable climate, free apartment, 90s gf (she loves u) etc

I'll take "things business people dont understand" for 100$.

No one hires software engineers to code. You're hired to solve problems. All of this AI bullshit has 0 capability to solve your problems, because it can only spit out what it's already stolen from seen somewhere else

It can also throw things against the wall with no concern for fitness-to=purpose. See "None pizza, left beef".

I’ve worked with a few PMs over my 12 year career that think devs are really only there to code like trained monkeys.

I'm at the point where what I work on requires such a depth of knowledge that I just manage my own projects. Doesn't help that my work's PM team consistently brings in new hires only to toss them on the difficult projects no one else is willing to take. They see a project is doomed to fail so they put their least skilled and newest person on it so the seniors don't suffer any failures.

Simplifying things to a level that is understandable for the PMs just leads to overlooked footguns. Trying to explain a small subset of the footguns just leads to them wildly misinterpreting what is going on, causing more work for me to sort out what terrible misconceptions they've blasted out to everyone else.

If you can't actually be a reliable force multiplier, or even someone I can rely on to get accurate information from other teams, just get out of my way please.

Guys that are putting billions of dollars into their AI companies making grand claims about AI replacing everyone in two years. Whoda thunk it

He who knows, does not speak. He who speaks, does not know.

--Lao Tzu...

What does the person "who knows" do when they have to give a presentation?

But coding never was the difficult part. It's understanding a concept, identify a problem and solve it with the possible methods. An AI just makes the coding part faster and gives me options to quicker identify a possible solution. Thankfully there's a never ending pile of projects, issues, todos and stackholder wants, that I don't see how we need less programmers. Maybe we need more to deal with AI, as now people can do a lot more in house instead of outsourcing, but as soon as that threshold is reached, companies will again contact large software companies. If people want to put AI into everything, you need people feeding the AI with company specific data and instruct people to use this AI.

All I see is middle management getting replaced, because instead of a boring meeting, I could just ask an AI.

I dread meetings and I can't wait for AIs to replace those managers. Or perhaps we'll have even more meetings because the management wants to know why we're so late despite the AI happily churning out meaningless codes that look so awesome like all that CSI VB GUI crap.

That's when you write an AI auto reply cron. Let the snake eat its tail. Hehe

It's been said before but the whiter your collar the more likely you are to be replaced by AI simply because the grunts tend to do more varied less pleibeon things.

Middle managers tend to write a lot of documents and emails which is something AI excels at. The programmers meanwhile have to come up with creative solutions to problems, and AI is less good at being creative, it basically just copy pastes known solutions from the web.

Realises devs have always joked about their jobs just being about copy-pasting solutions from StackOverflow 80% of the time

Oh God...

CEOs without a clue how things work think they know how things work.

I swear if we had no CEOs from today on the only impact would be that we wouldve less gibberish being spoken

If AI could replace anyone... it's those dingbats. I mean, what would you say, in this given example, the CEO does... exactly? Make up random bullshit? AI does that. Write a speech? AI does that. I love how these overpaid people think they can replace the talent but they... they are absolutely required and couldn't possibly be replaced! Talent and AI can't buy and enjoy the extra big yacht, or private jets, or over priced cars, or a giant over sized mansion... no you need people for that.

This will be used as an excuse to try to drive down wages while demanding more responsibilities from developers, even though this is absolute bullshit. However, if they actually follow through with their delusions and push to build platforms on AI-generated trash code, then soon after they'll have to hire people to fix such messes.

If, 24 months from now, most people aren't coding, it'll be because people like him cut jobs to make a quicker buck. Or nickel.

Well if it works, means that job wasn't that important, and the people doing that job should improve themselves to stay relevant.

Edit: wow what a bunch of hypersensitive babies. I swear, y'all just allergic to learning or something. I just said people need to improve themselves to stay relevant, and people freak out and send me death threats. What a joke.

job wasn’t that important

I keep telling you that changing out the battery in the smoke alarm isn't worth the effort and you keep telling me that the house is currently on fire, we need to get out of here immediately, and I just roll my eyes because you're only proving my point.

Sure, believe what you want to believe. You can either adapt to what's happening, or just get phased out. AI is happening whether you like it or not. You may as well learn to use it.

I get why you're enthusiastic about AI. This whole comment reads like it was AI generated.

You can adapt, but how you adapt matters.

AI in tech companies is like a hammer or drill. You can either get rid of your entire construction staff and replace them with a few hammers, or you can keep your staff and give each worker a hammer. In the first scenario, nothing gets done, yet jobs are replaced. In the second scenario, people keep their jobs, their jobs are easier, and the house gets built.

Yup. Most of us aren't CEOs, so we don't have a lot of say about how most companies are run. All we can do is improve ourselves.

For some reason, a lot of people seem to be against that. They prefer to whine.

Define "works"?

If you're a CEO, cutting all your talent, enshittifying your product, and pocketing the difference in new, lower costs vs standard profits might be considered as "working".

Hmmm maybe you're misunderstanding me.

What I mean is "coding" is basically the grunt work of development. The real skill is understanding the requirements and building something efficiently. Tbh, I hate coding.

What tools like Gemini or ChatGPT brings to the table is the ability to create small, efficient snippets of code that works. We can then just modify it to meet our more specific requirements.

This makes things much faster, for me at least. If the time comes when the AI can generate more efficient code, making my job easier, I'd count that as "works" for me.

Oh perhaps the CEOs are the ones that need to be replaced?

Yup, notice nowhere did I say they shouldnt. People read and infer what they want

Define "works."

Because the goals of a money-hungry CEO don't always align with those of the workers in the company itself (or often, even the consumer). I imagine this guy will think it worked just fine as he's enjoying his golden parachute.

If that's true, how come there isn't a single serious project written exclusively or mostly by an LLM? There isn't a single library or remotely original application made with Claude or Gemini. Not one.

Lets wait for any LLM do a single sucessful MR on Github first before starting a project on its own. Not aware of any.

there isn’t a single serious project written exclusively or mostly by an LLM? There isn’t a single library or remotely original application

IMHO "original" here is the key. Finding yet another clone of a Web framework ported from one language to another in order to push online a basic CMS slightly faster, I can imagine this. In fact I even bet that LLM, because they manipulate words in languages and that code can be safely (even thought not cheaply) tested within containers, could be an interesting solution for that.

... but that is NOT really creating value for anyone, unless that person is technically very savvy and thus able to leverage why a framework in a language over another creates new opportunities (say safety, performances, etc). So... for somebody who is not that savvy, "just" relying on the numerous existing already existing open-source providing exactly the value they expect, there is no incentive to re-invent.

For anything that is genuinely original, i.e something that is not a port to another architecture, a translation to another language, a slight optimization, but rather something that need just a bit of reasoning and evaluating against the value created, I'm very skeptical, even less so while pouring less resources EVEN with a radical drop in costs.

How many times does the public have to learn if the CEO says it, he probably doesn't know what he's talking about. If the devs say it, listen

Todays news: Rich assholes in suits are idiots and don’t know how their own companies are working. Make sure to share what they’re saying.

Yeah, that's not going to happen.

Yeah writing the code isn't really the hard part. It's knowing what code to write and how to structure it to work with your existing code or potential future code. Knowing where things might break so you can add the correct tests or alerts. Giving time estimates on how long it will take to build the parts of the system and building in phases to meet your teams needs.

I've always thought that design and maintenance are the difficult and gruelling parts, and writing code is when you get to relax for a bit. Most of the time you're in maintenance mode, and it's harder than writing new code.

This. I’m learning a new skill right now & hardly any of it is actual writing— it’s how to arrange the pieces someone else wrote (& which sometimes AI can decently reproduce.)

When you use a computer you don’t start by mining iron, because the thing is already built

Good luck debugging AI-generated code...

They think it will be easier than having people write the code from scratch. I don't know shit about coding but I know that's definitely not right.

AI is quite good at writing small sections of code. Usually because it's more or less just copying something off the internet that it's found, maybe changing a few bits around, but essentially just regurgitating something that's in its data set. I could of course just have done that but it saves time since the AI can find the relevant piece of code to copy and modify more or less instantly.

But it falls apart if you ask it to build entire applications. You can barely even get it to write pong without a lot of tinkering around after the fact which rather defeats the point really.

It also doesn't deal well if the thing you're trying to program for is not very well documented, which would include things like drivers, which presumably is their bread and butter.

That might actually be a good test for managers who think coders can be replaced by this. Have them try to make a working version of Pong using AI prompts.

Really simple. Just ask it to point out the error. Also maybe tell it how the code is wrong. And then hope that the new code didn't introduce new errors in formerly working sections. And that it understood what you meant. In a language that is inherently vague.

A company I used to work for outsourced most of their coding to a company in India. I say most because when the code came back the internal teams anways had to put a bunch of work in to fix it and integrate it with existing systems. I imagine that, if anything, LLMs will just take the place of that overseas coding farm. The code they spit out will still need to be fixed and modified so it works with your existing systems and that work is going to require programmers.

So instead of spending 1 day writing good code, we'll be spending a week debugging shitty code. Great.

It's the same claim when tools like Integromat, WayScript, PureData, vvvv and other VPLs (Visual Programming Languages) started to get some hype. I once worked for a company that strongly believed they'd "retire the need for coding", and my ex-boss was so confident and happy about that... Although VPLs were a practical thing, time is the ruler of truth, and for every dev-related job vacancy I see, they ask some programming language, the written ones (JS, PHP, Python, Ruby, Lua, and so on).

Because if you look closely, deep inside, voila, there's code in anything that is claimed to be no-code! Wow, could anyone imagine that? 🤯 /sarcasm

I made this meme a while back and I think it's relevant

Looking at your examples, and I have to object at putting scratch in there.

My kids use it in clubs, and it's great for getting algorithmic basics down before the keyboard proficiency is there for real coding.

It's still code. What makes scratch special is that it structurally rules out syntax errors while still looking quite like ordinary code. Node editors -- I have a love and hate relationship with them. When you're in e.g. Blender throwing together a shader it's very very nice to have easy visualisation of literally everything, but then you know you want to compute abs(a) + sin(b) + c^2 and yep that's five nodes right there because apparently even the possibility to type in a formula is too confusing for artists. Never mind that Blender allows you to input formulas (without variables though) into any field that accepts a number.

that's just how the code is rendered. There's still all the usual constructs

Everybody talks about AI killing programming jobs, but any developer who has had to use it knows it can’t do anything complex in programming. What it’s really going to replace is program managers, customer reps, makes most of HR obsolete, finance analysts, legal teams, and middle management. This people have very structured, rule based day to days. Getting an AI to write a very customized queuing system in Rust to suit your very specific business needs is nearly impossible. Getting AI to summarize Jira boards, analyze candidates experience, highlight key points of meetings (and obsolete most of them altogether), and gather data on outstanding patents is more in its wheelhouse.

I am starting to see a major uptick in recruiters reaching out to me because companies are starting to realize it was a mistake to stop hiring Software Engineers in the hopes that AI would replace them, but now my skills are going to come at a premium just like everyone else in Software Engineering with skills beyond “put a react app together”

Copilot can't even suggest a single Ansible or Terraform task without suggesting invalid/unsupported options. I can't imagine how bad it is at doing anything actually complex with an actual programming language.

It also doesn’t know what’s going on a couple line before it, so say I am in a language that has options for functional styling using maps and I want to keep that flow going, it will start throwing for loops at you, so you end up having to rewrite it all anyway. I have find I end up spending more time writing the prompts then validating it did what I want correctly (normally not) than just looking at the docs and doing it myself, the bonus being I don’t have to reprompt it again later because now I know how to do it

Trouble is, you're basing all that on now, not a year from now, or 6 months from now. It's too easy to look at it's weaknesses today and extrapolate. I think people need to get real about coding and AI. Coding is language and rules. Machines can learn that enormously faster and more accurately than humans. The ones who survive will be those who can wield it as a tool for creativity. But if you think it won't be capable of all the things it's currently weak at you're just kidding yourself unfortunately. It'll be like anything else - a tool for an operator. Middlemen will be wiped out of the process, of course, but those with money remain those without time or expertise, and there will always be a place for people willing to step in at that point. But they won't be coding. They'll be designing and solving problems.

We are 18 months into AI replacing me in 6 months. I mean… the CEO of OpenAI as well as many researchers have already said LLMs have mostly reached their limit. They are “generalizers” and if you ask them to do anything new they hallucinate quite frequently. Trying to get AI to replace developers when it hasn’t even replaced other menial office jobs is like saying “we taught AI to drive, it will replace all F1 drivers in 6 months”.

McDonald's tried to get AI to take over order taking. And gave up.

Yeah, it's not going to be coming for programmer jobs anytime soon. Well, except maybe a certain class of folks that are mostly warming seats that at most get asked to prep a file for compatibility with a new Java version, mostly there to feed management ego about 'number of developers' and serve as a bragging point to clients.

It's based on the last few years of messaging. They've consistently said AI will do X, Y, and Z, and it ends up doing each of those so poorly that you need pretty much the same staff to babysit the AI. I think it's actually a net-negative in terms of productivity for technical work because you end up having to go over the output extremely carefully to make sure its correct, whereas you'd have some level of trust with a human employee.

AI certainly has a place in a technical workflow, but it's nowhere close to replacing human workers, at least not right now. It'll keep eating at the fringes for the next 5 years minimum, if not indefinitely, and I think the net result will be making human workers more productive, not replacing human workers. And the more productive we are per person, the more valuable that person is, and the more work gets generated.

An inherent flaw in transformer architecture (what all LLMs use under the hood) is the quadratic memory cost to context. The model needs 4 times as much memory to remember its last 1000 output tokens as it needed to remember the last 500. When coding anything complex, the amount of code one has to consider quickly grows beyond these limits. At least, if you want it to work.

This is a fundamental flaw with transformer - based LLMs, an inherent limit on the complexity of task they can 'understand'. It isn't feasible to just keep throwing memory at the problem, a fundamental change in the underlying model structure is required. This is a subject of intense research, but nothing has emerged yet.

Transformers themselves were old hat and well studied long before these models broke into the mainstream with DallE and ChatGPT.

The real work of software engineering isn't the coding. That is like saying that being a doctor is all about reading health charts. Planning, designing, testing and maintaining software is the hard part, and it is often much more political than it is a technical challenge. I'm not worried about getting replaced by AI. In fact, LLMs ability to generate high volumes of code only makes the skills to understand it to be more in demand.

It's tons easier to repkace CEOs, HR, managers and so on than coders. Coders needs to be creative, an HR or manager not so much. Are they leaving three months from now you think?

I'll start worrying when they are all gone.

I don't understand how you could understand how LLMs work, and then write this.

Machines can learn that...

Ah, nevermind.

If you'll excuse me saying, I feel that you are the one who is looking at something and extrapolating.

While I highly doubt that becoming true for at least a decade, we can already replace CEOs by AI, you know? (:

https://www.independent.co.uk/tech/ai-ceo-artificial-intelligence-b2302091.html

Most middle managers could be replaced by a simple script already.

while(True):

staffNumbers-=1

staffWorkload*=1.1

staffWages*=0.95

executiveWages*=1.2

That's probably because they hit all the VC keywords of 2023.

  • AI
  • metaverse

I feel sorry for all those people in AWS that now have him as a leader...

24 months from now? Unlikely lol

I'm sure they'll hold strong to that prediction in 24 mo. It's just 24 more months away

We'll have full self driving next year.

I remember a little over a decade ago while I was still in public school hearing about super advanced cars that had self driving were coming soon, yet we're hardly anywhere closer to that goal (if you don't count the Tesla vehicles running red lights incidents).

A subscription to Popular Science magazine through most of my teen years did wonders for my skepticism.

We should all be switched to hydrogen fuel by now, for our public transport lines with per person carriages that can split off from the main line seamlessly at speed to go off on side routes to your individual destination, that automatically rejoin the main line when you're done with it. They were talking about all of that pre-2010.

I think I remember the hydrogen fuel thing.

Also, fuck Popular Science for making me think there was gonna be a zombie apocalypse due to some drug that turns you into a zombie.

In Phoenix you can take a Waymo (self driving taxi) just like an Uber. They have tons of ads and they’re everywhere on the roads.

I am in Phoenix and just took one to the airport. First time riding in a Waymo. It was uncannily good and much more confident than the FSD Tesla I’ve ridden in a few times.

I haven’t taken one yet but have several friends who have. Besides being generally good, one of the best parts is unlike Uber, there’s no chance that you have a weird driver that wants to talk to you the whole ride

That’s probably the amount of time remaining before they move on to selling the next tech buzz word to some suckers.

And just like that, they'll forget about these previous statements as well.

I fear Elon Musk's broken promises method is being admired and copied.

15 years at least. probably more like 30. and it will be questionable, because it will use a lot of energy for every query and a lot of resources for cooling

it will use a lot of energy for every query and a lot of resources for cooling

Well, so do coders. Coffee can be quite taxing on the environment, as can air conditioning!

If generative AI hasn't replaced artists, it won't replaced programmers.

Generative AI is much better at art than coding.

It will never replace artists anyway.

Art isn't just about what it looks, like it's also about an emotional connection. Inherently we think that you cannot have an emotional connection with something created by a computer. Humans will always prefer art created by humans, even if objectively there isn't a lot of difference.

The problem is that not everyone looks for that human-to-human emotional connection in art. For some, it's just a part of a much bigger whole.

For example, if you're an indie game dev with a small budget and no artistic skills, you may not be that scrupulous about getting an AI to generate some sprites or 3D models for you, if the alternative is to commission the art assets with money you don't have.

Similar idea applies to companies building a website. Why pay for a licence to download some stock images or design assets if you can just get a GenAI to pump out hundreds for you that are very convincing (and probably even better) for a couple bucks?

Sure, but those jobs are often pretty low-paid, like on fiverr or something. But for anything with a broader impact, like AAA games, large corporations, or public art, you'll commission a professional artist. AI works fine for low-budget projects and as a stand-in for works in progress, but it's not replacing human artists anytime soon, though it may assist artists (e.g. in producing mockups and whatnot quickly).

Generative AI is much better at art than coding.

Mostly because humans invented this convenient thing called abstract art - and since then tolerates pretty much everything that looks "strange" as art. Must have been a deep learning advocate with a time machine who came up with abstract art.

Don't need to be abstract art, it manages to make many kinds of art.

The difference between art and coding is that if you pick a slightly different color or make a line with slightly the wrong angle, it doesn't change much. In code, however, slight mistakes usually result in bugs.

I meant that thanks to abstract art we're willing to forgive "image glitches" in art by deep learning models.

I don't think that's the case though. Obvious glitches like 6-fingered hands can be avoided by generating a bunch of samples and picking the best one, and less obvious glitches tend to be overlooked, not considered a "feature" due to an appreciation of abstract art.

AI art works best for pieces that need to fade into the background, like stock images and whatnot to accompany more important copy. If it's taking center stage, it needs a lot more hand-holding that probably makes it about as costly as just having a human create it.

Seriously how can these CEOs of a GPU company not talk to a developer. You have loads of them to interview

Sure, Microsoft is happy to let their AIs scan everyone else’s code., but is anyone aware of any software houses letting AIs scan their in-house code?

Any lawyer worth their salt won’t let AIs anywhere near their company’s proprietary code intil they are positive that AI isn’t going to be blabbing the code out to every one of their competitors.

But of course, IANAL.

The LLMs they train on their code will only be accessible internally. They won’t leak their own intellectual property.

Will that not be more experiensive than having developers?

Yeah which is why this is a dumb statement from Amazon. But then again I don't expect C-suite managers to really understand the intricacies of their own companies.

Of course not. It will be more expensive and they'll still have to pay developers to figure out what's wrong with their AI code.

Possibly. It’s hard to know without seeing the numbers and assessing output quality and volume.

Also it’s not unheard of that some bigwig wastes millions of company €€ for some project they fancy. (Billions if they happen to be Elon)

If only we had an overarching structure that everyone in society has agreed exists for the purposes of enforcing laws and regulating things. Something that governs people living in a region... Maybe then they could be compelled to show exactly what they're using, and what those models are being trained with.

Oh well.

Let's assume this is true, just for discussion's sake. Who's going to be writing the prompts to get the code then? Surely someone who can understand the requirements, make sure the code functions, and then test it afterwards. That's a developer.

I don't believe for a single instance that what he says is going to happen, this is just a play for funding... But if it were to happen I'm pretty sure most companies would hire anything that moves for those jobs. You have many examples of companies offloading essential parts of their products externally.

I've also seen companies hiring tourism graduates (et al non engineering related) giving them a 3/4 week programming course, slapping a "software engineer" sticker on them and off they are to work on products they have no experience to work on. Then it's up to senior engineers to handle all that crap.

I think that's the point? They're saying that those coders will turn into prompt engineers. They didn't say they wouldn't have a job, just that they wouldn't be "coding".

Which I don't believe for a minute. I could see it eventually, but it's not "2 years" away by any stretch of the imagination.

Possibly. But... Here's the thing. I've dealt with "business rules" engines before at a job. I used a few different ones. The idea is always to make coding simpler so non technical people can do it. Unless you couldn't tell from context, I'm a software engineer lol. I was the one writing and troubleshooting those tools. And it was harder than if it was just in a "normal" language like Java or whatever.

I have a soft spot for this area and there's a non zero chance this comment makes me obsess over them again for a bit lol. But the point I'm making is that "normal" coding was always better and more useful.

It's not a perfect comparison because LLMs output "real" code and not code that is "Scratch-like", but I just don't see it happening.

I could see using LLMs exclusively over search engines (as a first place to look that is) in 2 years. But we'll see.

Definitely be coding less I think. Coding or programming is basically the "grunt work". The real skill is understanding requirements and translating that into some product.

No, going by them, they just talk to an AI voice and it will pop out a finished product.

I seem to recall about 13 years ago when "the cloud" was going to put everyone in IT Ops out of a job. At least according to people who have no idea what the IT department actually does.

"The cloud" certainly had an impact but the one thing it definitely did NOT do was send every system and network admin to the unemployment office. If anything it increased the demand for those kinds of jobs.

I remain unconcerned about my future career prospects.

Yes... because there will be users who will always refuse to fix their own computer issues. Even if there's an easy solution at their fingertips. Many don't even try to reboot. They just tell IT to fix it... then go get coffee for a half hour.

I'm going to call BS on that unless they are hiding some new models with huge context windows...

For anything that's not boilerplate, you have to type more as a prompt to the AI than just writing it yourself.

Also, if you have a behaviour/variable that is similar to something common, it will stubbornly refuse to do what you want.

Have you ever attempted to fill up one of those monster context windows up with useful context and then let the model try to do some useful task with all the information in it?

I have. Sometimes it works, but often it’s not pretty. Context window size is the new MHz, in terms of misleading performance measurements.

I find there comes a point where, even with a lot of context, the AI just hasn't been trained to solve the problem. At that point it will cycle you round and round the same few wrong answers until you give up and work it out yourself.

I think that giving the LLM an API to access additional context and then making it more of an agent style process will give the most improvement.

Let it request the interface for the class your using, let it request the code for that extension method you call. I think that would solve a lot, but I still see a LOT of instances where it calls wrong class/method names randomly.

This would also require a lot more in depth (and language specific!) IDE integration though, so I forsee a lot of price hikes for IDEs in the near future!

To actually answer your question - yes, but the only times I actually find it useful is for tests, for everything else it's usually iffy and takes longer.

Intelligently loading the window could be the next useful trick

I'm curious about what the "upskilling" is supposed to look like, and what's meant by the statement that most execs won't hire a developer without AI skills. Is the idea that everyone needs to know how to put ML models together and train them? Or is it just that everyone employable will need to be able to work with them? There's a big difference.

1 more...

!remindme in 24 months

I will put down a solid grand that this exact same article will be printed by the exact same website 24 months from now and it will receive the exact same reception.

Nah, if it doesn't pan out then all these folks will pretend they never said this, but in 24 months programming will be obsoleted by

If you go forward 12 months the AI bubble will have burst. If not sooner.

Most companies who bought into the hype are now (or will be soon) realizing it's nowhere near the ROI they hoped for, that the projects they've been financing are not working out, that forcing their people to use Copilot did not bring significant efficiency gains, and more and more are realizing they've been exchanging private and/or confidential data with Microsoft and boy there's a shitstorm gathering on that front.

If you have the ability to build an AI app in house - holy shit shit that can improve productivity. Copilot itself for office use.... Meh so far.

The most successful ML in-house projects I've seen took at least 3 times as long than initially projected to become usable, and the results were underwhelming.

You have to keep in mind that most of the corporate ML undertakings are fundamentally flawed because they don't use ML specialists. They use eager beavers who are enthusiastic about ML and entirely self-taught and will move on in 1 year and want to have "AI" on their resume when they leave.

Meanwhile, any software architect worth their salt will diplomatically avoid to give you any clear estimate for anything having to do with ML – because it's basically a black box full of hopes and dreams. They'll happily give you estimates and build infrastructure around the box but refuse to touch the actual thing with a ten foot pole.

There aren't enough AI specialists. More are being created by picking up these projects.

The problem is that AI is too hyped and people are trying to solve things it probably can't solve. The projects I have seen work are basically fancy data ingress/parsing/summarisation apps. That's where the current AI tech can really shine.

Says the person who is primarily paid with Amazon stock, wants to see that stock price rise for their own benefit, and won’t be in that job two years from now to be held accountable. Also, who has never written a kind of code. Yeah…. Ok. 🤮

Translation: "We're going to make the suite for building, testing, and deploying so obnoxiously difficult to integrate with your work environment that in two years nobody in your DevOps team will be able to get anything to a release state."

Me, fiddling with a config file for a legacy Perl script that's been holding up the ass-end of the business since 1996: "Uh, yeah that's great."

All the manufacturers of mechanical keyboards just cried 🥺

I wonder how they think that's possible, the attempts I've made at having an "AI" produce working code have failed spectacularly.

It will be interesting to find out if these words will come back and haunt them.

  • “I think there is a world market for maybe five computers”.
  • “640K ought to be enough for anybody.”

Nonsense. But then CEOs rarely know what the hell they’re talking about.

Can I join anyone's band of AI server farm raiders 24 months from now? Anyone forming a group? I will bring my meat bicycle.

Guys selling something claim it will make you taller and thinner, your dick bigger, your mother in law stop calling, and work as advertised.

That'd be an exciting world, since it'd massively increase access to software.

I am also very dubious about that claim.

In the long run, I do think that AI can legitimately handle a great deal of what humans do today. It's something to think about, plan for, sure.

I do not think that anything we have today is remotely near being on the brink of the kind of technical threshold required to do that, and I think that even in a world where that was true, that it'd probably take more than 2 years to transition most of the industry.

I am enthusiastic about AI's potential. I think that there is also -- partly because we have a fair number of unknowns unknowns, and partly because people have a strong incentive to oversell the particular AI thing that they personally are involved with to investors and the like -- a tendency to be overly-optimistic about the near-term potential.

I have another comment a while back talking about why I'm skeptical that the process of translating human-language requirements to machine-language instructions is going to be as amenable as translating human-language to human-consumable output. The gist, though, is that:

  • Humans rely on stuff that "looks to us like" what's going on in the real world to cue our brain to construct something. That's something where the kind of synthesis that people are doing with latent diffusion software works well. An image that's about 80% "accurate" works well enough for us; the lighting being a little odd or maybe an extra toe or something is something that we can miss. Ditto for natural-language stuff. But machine language doesn't work like that. A CPU requires a very specific set of instructions. If 1% is "off", a software package isn't going to work at all.

  • The process of programming involves incorporating knowledge about the real world with a set of requirements, because those requirements are in-and-of-themselves usually incomplete. I don't think that there's a great way to fill in those holes without having that deep knowledge of the world. This "deep knowledge and understanding of the world" is the hard stuff to do for AI. If we could do that, that's the kind of stuff that would let us create a general artificial intelligence that could do what a human does in general. Stable Diffusion's "understanding" of the world is limited to statistical properties of a set of 2D images; for that application, I think that we can create a very limited AI that can still produce useful output in a number of areas, which is why, in 2024, without producing an AI capable of performing generalized human tasks, we can still get some useful output from the thing. I don't think that there's likely a similar shortcut for much by way of programming. And hell, even for graphic arts, there's a lot of things that this approach just doesn't work for. I gave an example earlier in a discussion where I said "try and produce a page out of a comic book using stuff like Stable Diffusion". It's not really practical today; Stable Diffusion isn't building up a 3D mental model of the world, designing an entity that stably persists from image to image, and then rendering that. It doesn't know how it's reasonable for objects and the like to interact. I think that to reach that point, you're going to have to have a much-more-sophisticated understanding of the world, something that looks a lot more like what a human's looks like.

    The kind of stuff that we have today may be a component of such an AI system. But I don't think that the answer here is going to be "take existing latent diffusion software and throw a lot of hardware at it". I think that there's going to have to be some significant technical breakthroughs that have not happened yet, and that we're probably going to spend some time heading down dead-end approaches before we get to that. There's probably going to be a lot of hard R&D before we get there, and that's going to take time.

I guess the programmers should start learning how to mine coal...

We will all be given old school Casio calculators a d sent to crunch numbers in the bitcoin mines.

As software developer I am not scared that A.I will take away our jobs. What I am scared is that at that point A.I good enough to do most jobs out there.

All it really needs to do is replace large chunk of the service industry to wreck massive havock in our society.

Connecting human existence to their labour has a designed defect

While I do understand all of the scepticism in this thread, I have to say that I am personally amazed by GitHub Copilot.

I am just ramping up in a new company working on web development with Angular and Spring Boot. Even though I have 0 experience with this and have a background in python and C++, I got productive extremely quickly thanks to Copilot. Of course it does not work without flaws and you still need programming knowledge to wirte proper prompts and fix smaller issues in the resulting code. But without it I would be much further behind. It was even able to fix some issues in the html just based on a description of the issue I am observing in the webpage.

I do not think it will replace all programmers, but I do think it will replace some low level programmers who did repetitive tasks as the good programmers are extremely accelerated by only having to type subsets of what was needed before.

The thing about co-pilot is if you don't do anything with it, it just sits there.

You can't give it a prompt, you actually have to code stuff. It is a more advanced version of autocomplete. Now admittedly, it can write very large chunks of boilerplate code, which is extremely helpful, but it can't code the entire app, and it can't work with natural language prompts, at all.

Technically that makes chatGPT, (It really needs a better name) a more capable coder than copilot.

GitHub Copilot does chat like ChatGPT, and writes code based on a prompt. I have decades of experience programming and I use it a lot. It gives me starting points, boilerplate and examples. It won't build a whole application with no coding work from the human.

You can definitely 'chat' with copilot, like other llms as well as the inline editor auto complete.

If we replace all the low level coders with this fancy, expensive, environmentally destructive autocomplete - where are all the high level coders going to come from? Just spring from clone vats?

There will still be people who are able to rapidly learn with AI on their side and people who fail to do it. The fast ones are able to focus much more on the underlying problems and less on language specifics. The definition of low level and high level programers will change in the context of AI. Nobody today is implementing a linked list outside of university and in the future nobody will be needed to write the repeated code which still is needed in a lot of frameworks.

But of course there might the point of a critical collapse if AI only learns from its own code and inefficient code gets repeated constantly.

They aren't wrong, just late. Coding is already dead. Most coders I know spend very little time writing new code. Meeting/discussions about requirements, debugging, fighting with pipelines or tests. I once read that a good programmer writes 10 to 100 lines of fully functional, tested, working, and meeting the actual need code a day. I believe it.

Okay but you literally said they're still writing the code though lmao.

Coding is already dead. Most coders I know spend very little time writing new code.

Oh no, I should probably tell this my whole company and all of their partners. We're just sitting around getting paid for nothing apparently. I've never realised that. /s