44 pointsby delichon3 hours ago6 comments
  • jdw642 hours ago
    Personally, I prefer vibe coding in the sense of stitching things together at the function-to-method level.

    Unlike people who take the extreme position that vibe coders are useless, I do think LLMs often write individual functions or methods better than I do. But in a way, that does not fundamentally change the nature of the work. Even before LLMs, many functions and methods were effectively assembled from libraries, Stack Overflow snippets, documentation examples, and copied patterns.

    The real limitation comes from the nature of transformer-based LLMs and their context windows. Agentic coding has a ceiling. Once the codebase reaches a scale where the agent can no longer hold the relevant structure in context, you need a programmer again.

    At that point, software engineering becomes necessary: knowing how to split things according to cohesion and coupling, using patterns to constrain degrees of freedom, and designing boundaries that keep the system understandable.

    In my experience, agentic coding is useful for building skeletons. But if you let the agent write everything by itself, the codebase tends to degrade. The human role is to divide the work into task units that the agent can handle well.

    Eventually, a person is still needed.

    If you make an agent do everything, it tends to create god objects, or it strangely glues things together even when the structure could have been separated with a simpler pattern. Thinking about it now, this may be exactly why I was drawn to books like EIB: they teach how to constrain freedom in software design so the system does not collapse under its own flexibility.

    • wombat-man2 hours ago
      The models are improving. The software that harnesses them is also improving. It wasn't that long ago that the models were quite bad at a lot of the tasks that they are excelling at today. I do agree there's probably a ceiling to what we can get out of these, but I also don't think we have quite hit that point yet.
      • jdw642 hours ago
        I agree with what you said. And perhaps my belief that “people like me are still needed” is just a desperate form of self-persuasion.

        If AI replaces everything, then I become unnecessary. So maybe I am simply trying to convince myself that developers like me are still needed.

        That said, realistically, I still think there are limits unless the essence of architecture itself changes. I also acknowledge part of your perspective.

        Those of us who are not in the AI field tend to experience AI progress not as a linear or continuous process, but as a series of discrete events, such as major model releases. Because of that, there is inevitably a gap in perspective.

        People inside the industry, at least those who are not just promoting hype, often seem to feel that technological progress is exponential. But since we are not part of that industry, we experience it more episodically, as separate events.

        At the same time, capital has a self-fulfilling quality. If enough capital concentrates in one direction, what looked like linear progress may suddenly accelerate in an almost exponential way.

        However, even that kind of model can eventually hit a specific limit. I do not know when that limit will arrive, because I am not an AI industry insider. More precisely, I am closer to someone who uses Hugging Face models, builds around them, and serves them, rather than someone working on AI R&D itself.

        • tharkun__an hour ago

              “people like me are still needed” is just a desperate form of self-persuasion.
          
          No, no it's not. I've seen what "PM armed with an LLM" will do. Trust me, if you're a decent enough Full Stack software engineer that can take an idea and run with it to implement it, you'll have a leg up over the PM with the idea that has no idea how to "do computers".

          Most of what these PMs can produce nowadays turns boardroom heads, sure. But it's just that: visuals and just enough prototype functionality that it fools the people you're demoing to. Seen enough of these in the recent past.

          Will there be some PMs that can become "software developers" while armed with an LLM? Sure!

          But that's not the majority. On the other hand, yes there are going to be "software developers" that will be out of a job because of LLMs, because the devs that were FS and could take an idea from 0-1 with very little overhead even in the past can now do so much faster and further without handing off to the intermediates and juniors. They mentor their LLM intern rather than their intermediates and juniors. The perpetual intermediate devs with 20 years of experience are the ones that are gonna have a larger and larger problem I'd say.

          The Staff engineer that was able to run circles around others all along? They'll teach their LLM intern into an intermediate rather than having to "10 times" a bunch of perpetual intermediates with 20 years of experience.

    • 8note2 hours ago
      its not necessarily better, but its certainly good enough, if youre already used to distributing work to different people

      the scale of code doesnt really matter that much, as long as a programmer can point it at the right places.

      i think actually you want to be really involved in the skeleton, since from what ive seen the agent is quite bad at making skeletons that it can do a good job extending.

      if you get the base right though, the agent can make precise changes in large code bases

      • jdw64an hour ago
        Thinking about it, I think what is interesting about the output of agentic coding is this:

        I mostly agree with the general tendency that it starts to break down as the context grows. But there is also a difference in how people evaluate it. Some people say agents are good at building the skeleton, while others say they are better at extending an existing structure.

        I think this depends on the setup, and it is ultimately a trade-off.

        In my case, I usually work on codebases around 60,000 LoC. The programs I deliver are generally between 60,000 and 80,000 lines of code. I think I can fairly call myself a specialist at that scale, since I have personally delivered close to 40 projects of that size.

        At that scale, I felt that agentic coding was actually very good at building the initial skeleton.

        I do not know what kind of work you usually do, but if your work involves highly precise, low-level tasks, then I can understand why you might feel differently.

        In my case, I mostly assemble high-level libraries and frameworks into working systems, so that may be why I experience it this way.

        • sroussey3 minutes ago
          The coding agents are good at growing code.

          Like a child growing up!

          Also, like a cancer.

          Similar process, different outcomes.

  • 2 hours ago
    undefined
  • dontgetfired2 hours ago
    90% of the job ads I see have the word "AI" in them. It can be a startup hoping for a get-rick-quick opportunity from the AI hype, or an established company.

    Both types expect you to spend as many tokens as possible so that the AI bubble doesn't burst (presumably because leadership has a financial interest in this).

    Your actual productivity isn't important. If you point out that you're much faster writing code on your own in 90% of cases, you will be told you're not good at AI, you're not prompting it correctly and that generally you're not AI-native and that you'll be left behind. To be precise, token usage is a performance metric, so you'll be let go if Claude is not running continuously 8 hours a day.

    I'd like to know how many places have mandates to write 100% of your code using AI, as well as to max out your AI agent's plan. For some reason nobody talks about it even though I know several companies around the world that are forcing this on their employees.

    If you're looking for a job then you don't have a choice, it's better to have an income. But if you're looking to change jobs to get away from AI to actually be productive and gain experience then it's a very bad job market.

    • 2 hours ago
      undefined
  • 2 hours ago
    undefined
  • enraged_camel2 hours ago
    Title is editorialized and the report is from two months ago.
  • trhway2 hours ago
    our labor market is cyclic, relatively short busts and long initially-slow-and-faster-and-faster booms. We had busts of 2000-2003, 2008-2010(11?), 2022- i guess 2026. I wasn't in US in 199x, yet i guess beginning of the 199x also was a bit tough.

    Unavoidable AI-based productivity growth, in software and in all the other industries, will lead to the software, specifically AI in this case, not just eating the wold, it would be devouring it. Such AI revolution will mean even more need for software engineers, just like the Personal Computer revolution and the Internet revolution did in their times. Of course the software engineering will get changed like it did in those previous revolutions.