49 pointsby CrankyBear7 hours ago9 comments
  • s_ting7654 hours ago
    Coding agents are like asking a genie for code. They will give you the code you ask for alright but you never know what kind of curse has been crontabbed for you.
  • beastman825 hours ago
    gotta love a site that hijacks your back button and makes you hit it 3 times.
    • rtkwe5 hours ago
      Doesn't for me until I scroll past the end of the article to read the next one. To get 3 you'd have to scroll through multiple articles.
    • kjkjadksj5 hours ago
      On mobile you get a little 1.5” strip to read. Rest of the screen is autoplaying ads. No, I didn’t suffer through that to read the article.
      • klibertp4 hours ago
        I learned (from the second paragraph) that 7 out of 12 is "vast majority". I'm a bit reluctant to read on after that...
      • hagbard_c2 hours ago
        > Rest of the screen is autoplaying ads.

        Only if you don't have a working content blocker on that device which you should have. Using web devices without content blockers is just as bad for your health as sunbathing on the local garbage dump. Just get a content blocker and protect yourself from the filth. If you happen to use a device which does not allow effective content blocking - something from the fruit factory or something totally googly - you might want to consider getting something which is more aligned with your needs and less with those of the aforementioned companies.

  • MithrilTuxedo5 hours ago
    I'm thinking of Debian and how much effort it takes to maintain stability and security over time.

    I can't imagine we'll really be able to trust AI without it's use in open source software where we can see how reliable it is.

    • ghywertelling4 hours ago
      If AI works, imagine 10 years of security updates and possibly 10 years of full OS upgrades for Android phones.
      • soperj4 hours ago
        why would they do that when they want to sell you a new phone?
        • nout4 hours ago
          With AI you can do that, or smaller companies can do that. It levels the field.
  • ozlikethewizard6 hours ago
    How many year's end have to pass?
  • Hamuko5 hours ago
    When people suggest to use AI for open-source projects, what exactly are they advocating for given that the median open-source project budget is pretty much $0/month? Maybe $1/month if the maintainer likes to have a website for the project.
    • hparadiz4 hours ago
      The $20 / month subscription is less than I pay for electricity already and the local models are also capable enough. A lot of the top open source projects have paid devs working on them already.
    • ivaivanova4 hours ago
      Ha, I just shipped an open source library using the Anthropic API. The way I solved the cost concern was by having users bring their own API key. Zero infra cost on my end and they pay for what they use.
  • fsflover5 hours ago
    See also: https://news.ycombinator.com/item?id=47547849

    AI bug reports went from junk to legit overnight, says Linux kernel czar (theregister.com)

    58 points by amarant 4 days ago

  • supernes6 hours ago
    TLDR: Greg Kroah-Hartman says that last month something magical happened and AI output is no longer "slop".
    • abletonlive4 hours ago
      For months, people on hackernews and reddit with a bad model of reality and lack of observation and understanding have been telling people that LLMs are useless and they are just toys, that they can't program etc. I have a whole graveyard of replies in my comments section from users on HN saying exactly this.

      Nothing has changed this month, it's been good for a while and a small minority could see that it was already decided that it was paradigm shifting.

      This is a notice to all of you that are just now changing your mind and crossing over: Your cognition of reality is flawed and you are not as good as you think you are at observing technological progress. The only thing that has changed is that you just now noticed how capable LLMs are. There are many that have been telling you before 2026 that it was here and you all tried to paint us as charlatans.

      Reddit is still very far behind. Browsing /r/programming and other software dev releated subs like /r/cscareerquestions is like being at a dinosaur museum

      • saltcured3 hours ago
        Based on your last sentences, I am pretty sure you will dismiss me. But, I have a null hypothesis to consider...

        Like you implied, I think a personal threshold crossing gives this false impression that "everything changed" this month or last month or last year. Like you said, the main thing that changed in one particular month was the observer.

        But, perhaps the AI epiphany is not waking up to recognize how good AI already was. Instead, it could be when an individual's standards degrade such that the same AI usage is seen as a benefit instead of a liability. Both interpretations yield the same basic pattern of adoption and commentary that we see right now.

        The difference will be in the long-term outcome. Some years from now, will we see that this mass adoption yielded a renaissance of productivity and quality, or a cataclysm of slop-induced liability and loss?

        • keeda2 hours ago
          I know appeal to authority can be a fallacy, but there is something to be said for appeal to a preponderence of concurring authorities. Multiple notable personalities known for their technical chops have been endorsing AI assisted coding, so it's hard to argue that every one of them lowered their standards.

          It's been fun seeing the cognitive dissonance in anti-LLM tech circles as technical giants that they idolized, from Torvalds through Carmack all the way up to Knuth, say something positive about AI, let alone sing praises of it!

        • jason_oster2 hours ago
          I have to point out that having "high personal standards" is its own fatal flaw. The worst quality code I've seen comes from developers with little self awareness or humility. They call themselves artisans and take no responsibility for the minefield of bugs and security vulnerabilities left in their wake. The Internet is held together with bubblegum and baling wire [1] [2] because artisans reject self improvement.

          These same artisans complain about how bad AI generated code is. The AI is trained on your bad artisan code. It's like they are looking in the mirror for the first time and being disgusted by what they see.

          [1]: https://techcrunch.com/2014/03/29/the-internet-is-held-toget...

          [2]: https://krebsonsecurity.com/2021/11/the-internet-is-held-tog...

    • throwaway20275 hours ago
      I noticed it last December.
      • BatFastard2 hours ago
        December for me too, went from hallucinating a lot, to doing really well. I guess it was opus 4.5?
      • no_shadowban_64 hours ago
        paradigm shift, bro.
    • dragonelite4 hours ago
      Let me guess the big tech cheque finally cleared?
    • georgemcbay5 hours ago
      IMO around December of last year LLM output (for coding at least, not for everything) went from "almost 100% certainly slop" to "probably not slop, if you asked for the right thing while being aware of context limitations".

      A lot of people seem stuck with their older (correct at the time) views of them still always producing slop.

      FWIW I am more of an AI doomer (in the sense that I think the economic results from them will be disastrous for knowledge workers given our political realities) than booster, but in terms of utility to get work done they did pass a clear inflection point quite recently.

      • bluefirebrand5 hours ago
        > if you asked for the right thing while being aware of context limitations

        So, still pretty likely to produce slop in a large majority of cases

        If the most useful place for them is where you've already specced things out to that degree of precision then they aren't that useful?

        Speccing things to that precision is the time consuming and difficult work anyways, after all.

        • georgemcbay5 hours ago
          I think LLMs currently need to be used by someone who knows what they are doing to produce value, but the jump they made from being endless slop machines to useful tools in the right hands is enough for me to assume it is only a matter of time until they will be useful tools in the hands of even the untrained masses.

          I wish this wasn't true because I think it will economically upend the industry in which I have a career, but sadly the universe doesn't care what I wish.

          • mjr005 hours ago
            > assume it is only a matter of time until they will be useful tools in the hands of even the untrained masses.

            IMO this vastly overestimates how good the "untrained masses" are at thinking in a logical, mathematical way. Apparently something as basic as Calculus II has a fail rate of ~50% in most universities.

            • chromacity4 hours ago
              How does this follow?

              There's nothing "basic" about Calculus II. Calculus is uniquely cursed in mathematical education because everything that comes before it is more or less rooted in intuition about the real world, while calculus is built on axioms that are far more abstract and not substantiated well (not until later in your mathematical education). I expect many intelligent, resourceful people to fail it and I think it says more about the abstractions we're teaching than anything else.

              But also, prompting LLMs to give good results is nowhere near as complex as calculus.

            • isueej4 hours ago
              That’s why you can’t generalise opinions on here.

              Most people on here don’t belong to that group of people. So ofc they can find a way to create value out of a thing that requires some tinkering and playing with.

              The question is can the techniques evolve to become technologies to produce stuff with minimal effort - whilst only knowing the bare minimum. I’m not convinced personally - it’s a pipe dream and overlooks the innate skill necessary to produce stuff.

            • xyzelement5 hours ago
              Who cares? People know what they want and need and AI is increasingly able to take it from there.
              • embedding-shape4 hours ago
                > People know what they want and need

                If they truly did, there wouldn't be a huge amount of humans whose role is basically "Take what users/executives say they want, and figure out what they REALLY want, then write that down for others".

                Maybe I've worked for too many startups, and only consulted for larger companies, but everywhere in businesses I see so many problems that are basically "Others misunderstood what that person meant" and/or "Someone thought they wanted X, they actually wanted Y".

              • mjr004 hours ago
                > People know what they want and need

                The multi-decade existence of roles like "business analysts" and "product owners" (and sometimes "customer success") is pretty strong evidence that this is not the case.

              • PhilipRoman4 hours ago
                What they want? Sometimes. What they need? Almost never.
              • isueej4 hours ago
                Right… people knew they wanted an iPhone before it was conceived, right? Lmao

                The arrogance of people like you is astonishing.

          • bluefirebrand4 hours ago
            > I wish this wasn't true because I think it will economically upend the industry in which I have a career, but sadly the universe doesn't care what I wish.

            I mean, yes. I'm worried about my career too, but for different reasons. I don't think these things are actually good enough to replace me, but I do think it doesn't matter to the people signing the cheques.

            I don't believe LLMs are producing anything better than slop. I think people's standards have been sinking for a long time and will continue to sink until they reach the level LLMs produce

            The problem isn't just LLMs and the fact they produce slop, it's that people are overall pretty fine with slop

            I'm not though, so there's no place for me in most software business anymore

            • isueej4 hours ago
              I’m not a SWE.

              But I look at software from the perspective of them as being objects.

              Since it’s intangible people can’t see within. So something can look pretty even if underlying it all, it’s slop.

              However, there is an implicit trade off - mounting slop makes you more vulnerable from a security standpoint, bugs etc which can destroy trust and experience of using the software. This can essentially put the life of a business at risk.

              People aren’t thinking so much about that risk - because it hasn’t happened to anyone large substantially. What I think about is will slop just continue to mount unchecked? Or are people expecting there to be improvements that enable oneself to go back and clean up the slop with more powerful tooling?

              If the latter does not come about, I think we will see more firms come under stress.

              Overall though, I think too much focus is on the acceleration of output. I never think that’s the most important thing. It’s secondary to having a crystal clear vision. The problem is to have a clear vision requires doing a lot of grunt work - it trains and conditions your mind to think a particular kind of way.

              It will be interesting to see how this all plays out.

    • ramesh315 hours ago
      >TLDR: Greg Kroah-Hartman says that last month something magical happened and AI output is no longer "slop".

      Opus 4.6 has been a step change. It's simply never wrong anymore. You may need to continue giving it further clarification as to what you want, but it never makes mistakes with what it intends to do now.

      • Balinares4 hours ago
        Yo, just because you can't tell when Claude is wrong, doesn't mean it's right.

        I do agree that the Q1 2026 models in general have passed a threshold, but goodness almighty Opus 4.6 still screws up a lot.

        • no_shadowban_64 hours ago
          > Yo, just because you can't tell when Claude is wrong, doesn't mean it's right.

          Just because you can't tell when Claude is right, doesn't mean that you are.

          This shit is AGI, with decades + billions of dollars of research and development behind it.

          So don't get all high and mighty now, acting like you know better than Claude.

      • binarymax5 hours ago
        It’s wrong. It made large mistakes on my code literally yesterday.
        • brcmthrowaway5 hours ago
          Wrong context
          • banannaise4 hours ago
            Ah, the eternal handwave for anything the AI doesn't do well - it must be user error.
          • binarymax5 hours ago
            No. Aside from just making an algorithm that didn’t even run, it refused to use an MCP that it had registered in the same context session.
          • 98642478887544 hours ago
            [dead]
      • taormina4 hours ago
        Can I get your magical version of Opus that works? 4.6 has been a side grade at best, and worse than prior models most days.
  • pbiggar5 hours ago
    > What happened? Kroah-Hartman shrugged: "We don't know. Nobody seems to know why. Either a lot more tools got a lot better, or people started going ..."

    Odd sentiment. It's pretty clear the tools crossed a threshold last year (in April as I recall) where they became good enough to actually write entire applications, and just accelerated from there. Today they're amazing and no-one I know is writing artisanal code anymore (at least, not at work).

    • neonstatic3 hours ago
      > We don't know. Nobody seems to know why

      This is the buried lede. It's a propaganda piece.