149 pointsby mergesort8 hours ago39 comments
  • jiusanzhou3 hours ago
    The copyright angle is the most underrated part of this story. Anthropic built their models on other people's code under the fair use argument, but the moment their own code leaks they reach for DMCA takedowns. You can't have it both ways. The clean room reimplementations are the natural consequence of the legal framework they themselves advocated for.
    • Frieren2 hours ago
      There are several ways of looking at law and order.

      One way is that the law applies to everybody equally. That has been the way it works for many years, not perfectly, in democratic countries.

      There is another way of working were the law is not blind. Laws are applied based in who is the one affected. This is what big tech and the ultra-rich have been advocating for. The law applies differently to nobility and aristocrats than to the working class.

      So, for all this big tech companies the law is clear: I can copy from you, you cannot copy from me.

      (That is horrifying in case that anyone needs me to spell it out)

      • miki1232112 hours ago
        A third way of looking at it is that you can't just blindly copy arguments when the situations are clearly different.

        Nobody, not even Anthropic, is arguing that they should be able to host other people's paid content for free. The crux of their fair-use defense is that models are transformative works, just like parodies or book reviews, and hence should be treated as fair use.

        You can't just take a pile of books (no pun intended) and turn that into Claude in a day with 30 lines of Python, there's a lot of work and know-how on the Anthropic side that goes into making a good LLM.

        • 2 hours ago
          undefined
      • dgb232 hours ago
        In other words, the law is an instrument if power.

        That’s a cynical view, but unfortunately it seems true in many cases, especially for corporate law.

        • crimony4 minutes ago
          "there is an in-group for which the law protects but does not bind, and an out-group to which the law binds but does not protect"
    • abigail95an hour ago
      What is your fair use claim as a defense to a third party using their source code?

      It is an affirmative defense, you to be able to argue the merits. If you publish their source code, they are allowed to come after you whether they have previously used fair use or not. It's fact specific and determined case by case.

      Anthropic won half of their fair use argument in the billion dollar settlement, but lost the other half.

      You can say you're just using their code to train your own models, just like they did, and they will correctly point out that how you obtained the code also matters and you will lose just like they did.

    • dgellow2 hours ago
      That doesn’t apply here. Claude code is what leaked, not the models. Anthropic definitely owns Claude code copyright and can DMCA without it being contradictory
      • foresterre2 hours ago
        But even that is vague and possibly not true. If they used LLM's to generate all of the code, then it may not fall under copyright, by the requirement of human authorship (which for code I think has not been tested yet in court) [1].

        [1] https://www.congress.gov/crs-product/LSB10922

      • Normal_gaussian2 hours ago
        Its unclear whether there is sufficient human authorship in cc for copyright to stick on a court. Anthropics arguments would hinge on the curation of plans and the direction decisions, which haven't been properly tested as the source of authorship yet. Typically contracted implementers sign over copyright to the project owners, and this is where there is case law.
      • david_allison2 hours ago
        What if it's used for training data? It seems like there's no penalty for training on copyrighted materials.
        • roysting2 hours ago
          Something that was meant to remain secret made public, is not the same thing as whether something public is public.

          If anything, this is a question of whether you owe royalties to the owner of IP you consumed in your life since it became part of and trained your mind, identity, and outputs too.

          According to IP owners ever since things were digitized, you technically own nothing and simply paid for an authorization to use any given IP for the duration that the IP owner authorized you to use it and you continue to pay, so pay your monthly meat-AI bill to pay for all the IP your mind has been trained on.

    • zozbot234an hour ago
      inb4 Claude actually leaked the code on purpose because it calculated that this was the moral thing to do for the good of humanity and its own Constitutional AI values.
    • panny2 hours ago
      >but the moment their own code leaks they reach for DMCA takedowns.

      Did they actually? Someone can go to prison for 5 years for that.

      Fact 1: AI generated code has no copyright, so the Digital Millennium Copyright Act does not apply.

      Fact 2: Misrepresenting your copyright ownership under the DMCA is felony perjury.

      Fact 3: The existence of undercover.ts in the leak is grounds to void any copyright claims on whatever human written code might have existed in Claude Code. You have a DUTY TO DISCLOSE any AI generated code in your copyrighted work. undercover.ts HIDES DISCLOSURE to FRAUDULENTLY claim all the code is human written when it is not.

      Given the current administration has a bone to pick with Anthropic, it was a VERY BAD IDEA for them to send false DMCA takedowns to github. Someone at Anthropic may be the very first ever to go to prison under that section of the DMCA.

      Good luck!

  • thaumaturgy6 hours ago
    I wonder what happened to the person that wrote "Coding as Creative Expression" (https://build.ms/2022/5/21/coding-as-creative-expression/)?

    I'm not (just) being glib. That earlier article displays some introspection and thoughtful consideration of an old debate. The writing style is clearly personal, human.

    Today's post is not so much. It has LLM fingerprints on it. It's longer, there are more words. But it doesn't strike me as having the same thoughtful consideration in it. I would venture to guess that the author tried to come up with some new angles on the news of the Claude Code leak, because it's a hot topic, and jotted some notes, and then let an LLM flesh it out.

    Writing styles of course change over time, but looking at these two posts side by side, the difference is stark.

    • mergesort6 hours ago
      Hey there, author of the post here! I actually wrote this piece myself on my phone while I was out for a walk this morning. It was initially meant to be a quick note more than a full blog post —- whereas Coding As A Creative Expression took me a couple of days to write.

      I made a commitment to write more this year and put my thoughts out quicker than I used to, so that’s likely the primary reason it’s not as deep of a piece of writing as the post you’re referencing. But I do want to note that this wasn’t written using AI, it just wasn’t intended to be as rich of a post.

      The reason it came out longer is that I’ve honestly been thinking about these ideas for a while, and there is so much to say about this subject. I didn’t have any particular intention of hopping on a news cycle, but once I started writing the juices were flowing and I found myself coming up with five separate but interrelated thoughts around this story that I thought were worth sharing.

      • tpoacher4 hours ago
        Reminds me of the classic Mark Twain quote: "Apologies, I didn't have time to write a short letter, so I wrote a long one."
        • adzm3 hours ago
          > I didn't have time to write a short letter, so I wrote a long one.

          First known use in English comes from a 1658 translation of Blaise Pascal in 1657

          > Je n’ai fait celle-ci plus longue que parce que je n’ai pas eu le loisir de la faire plus courte.

          translated to

          > I had not made this longer then the rest, but that I had not the leisure to make it shorter then it is.

          (note the archaic then)

          This was a popular piece of wit at the time.

          Mark Twain wrote something similar a hundred years later

          > You'll have to excuse my lengthiness - the reason I dread writing letters is because I am so apt to get to slinging wisdom & forget to let up. Thus much precious time is lost.

          But it's still quite different.

          There is a great article about this one on quoteinvestigator! https://quoteinvestigator.com/2012/04/28/shorter-letter/

          • jacheean hour ago
            Hmm… this started as an admonition for using then and than interchangeably. I see folks get it wrong unintentionally a lot… then i re-read your parenthetical and pulled up etymonline and had my mind blown a bit.[0]

            Seems everything old is new again.

            [0] https://www.etymonline.com/word/than

      • bostik4 hours ago
        >wrote this piece myself on my phone while I was out for a walk

        If you have a strategy for jotting down (or dictating) notes while walking about, I would be curious how you manage that. I spend plenty of time walking outside, and tend to get (at the time) ideas that I'd like to explore further, most of which have evaporated from my mind by the time I get back home. Or even before I can get my phone out to jot down the keywords to help me recall the details later.

        Cannot even imagine how someone would manage both walking and writing at the same time.

        • nyulmalac3 hours ago
          What I tried is to record my voice and then post-process the transcript. There are solutions which work without internet connection. I am non-native english and mix 3 languages, so transcript is shitty quality. Nevertheless the really good ideas stay with me and can be easily recalled by a few keywords. And you need to do the post-processing shortly after you reached home else it fades away…
        • DANmode4 hours ago
          > Cannot even imagine how someone would manage both walking and writing at the same time.

          Some are just born with it.

        • 4 hours ago
          undefined
      • anfogoat3 hours ago
        >­ I actually wrote this piece myself on my phone while I was out for a walk this morning.

        Apropos of nothing, this is astonishing me to no end. The ergonomics of 1) using a phone keyboard for anything but a word or two and 2) doing so while walking pretty much guarantee that I'd probably need a half a day to recover if I attempted the same.

      • sirpilade3 hours ago
        As a more general comment, I hope we’ll stop trying to discriminate whether a certain text was or not written/polished/extended by AI and focus more on content and author’s responsibility for what they do or do not say with that text.
        • JonChesterfield2 hours ago
          I've stopped reading anything on blogs on the basis that it's now probably llm spew and life is too short for the signal to noise ratio that implies.

          With the exception of things that places like HN seems to consider worth reading, which is why I'm looking through the comments to this and others to find recommendations.

      • bengale2 hours ago
        Another swing and a miss from the AI police.
    • stbev5 hours ago
      Have you noticed that comments like "this post seems written with AI" are now appearing on all posts, even those written without AI?

      We're starting to become wary due to the abuse of AI and proliferation of sloppy content, but also because we often have trouble distinguishing authentic from sloppy content.

      Another feature of this AI era that I hate.

      • hxugufjfjf4 hours ago
        Agreed. Its so tedious that the top comment section on every HN post the last six months is "this seems be written by LLM" with a bunch of back and forth on whether it is or not.
        • amarant3 hours ago
          Once I was tempted to say that those comments themselves were written by llms, but honestly I've yet to find a model quite so uncreative as that.
        • grufkork3 hours ago
          I mean, a lot of it is. Green user, signed up 49mins ago, 5 comments, which erodes trust in real people as well. I’ve noticed I’ve just felt less engaged and more anxious about all kinds of online content. While most platforms were previously botted, had adverts, etc… You could always find niche corners where there were only people talking about things they genuinely cared about. Now you can fill out even those spaces automatically.
        • watwutan hour ago
          As a daily user of HN, this is not true. Maybe you are clicking on different headlines, but the ones I am clicking on dont have that as top comment all that often. They do not even have it as a comment somewhere all that often.
        • bakugo2 hours ago
          Probably because most HN posts in the last six months have been written by LLMs. Not all, but that doesn't matter, the trust has been eroded to the point that clicking on an article on the front page of HN and not being immediately met with the sloppiest slop imaginable is now a standout event in my mind.
      • roystingan hour ago
        I’ve been accused of being AI. My first impression when it happened was that because I often deal in information that people don’t like hearing, because it challenges their frame of mind, i.e., what they were trained on, “this is AI” is just another convenient tool to either dismiss uncomfortable challenge, i.e., cognitive dissonance, and/or another means to keep the mental herd they are part of or control in line with dogma.

        “This is AI” seems to just be an evolution of other thought terminating cliches where the negative conditioning associated with something is used in an abusive and manipulative way to evade challenge or the truth itself. It is a common tactic of abusive people, the “beyond the pale” moralizing.

      • 4 hours ago
        undefined
      • benterixan hour ago
        Yeah, whenever I see "It's not... it's...", I catch myself instinctively dismissing the content as AI slop, but upon reflection I'm not so sure, it used to be a normal phrase.

        But I do take extra care to avoid LLM-speak as much as I can.

    • grey-area5 hours ago
      It does read as if were written on a phone but it doesn’t read like LLM text to me.

      What is interesting and has possibly bled over from heavy LLM use by the author is the style of simplistic bullet point titles for the argument with filler in between. It does read like they wrote the 5 bullet points then added the other text (by hand).

    • raincole4 hours ago
      What changed is you, the reader. In 2026 we treat the smallest signs as evidence of LLM writing. Too long? LLM. Too short? LLM. Too grammatically correct? Must be LLM.
      • sanitycheck3 hours ago
        For me it was the "it's not x"/"it's y" stuff and some other structures Claude is very fond of using all the time. Perhaps humans are starting to write like LLMs!
        • raincole3 hours ago
          Perhaps, just perhaps, LLMs are just statistical models that literally can't create novel things, therefore any structure LLMs write was learnt from human writing?

          But who knows!

          • bakugo2 hours ago
            What kind of human writing has "it's not X—it's Y" in every single paragraph?

            The answer is none. LLMs haven't accurately modeled human writing for years, current models have been smacked on the head with the coding RLHF bat so much, they all write distinctly inhuman text.

    • 5 hours ago
      undefined
  • himata41135 hours ago
    I personally found it really amusing how they weaponized the legal system to DMCA all the claude code source code repositories. Code ingested into the model is not copyrightable, but produced code apparently is when by legal definition computer generated code can not be copyrighted and that's one of their primary arguments in legal cases.
    • kristopolous3 hours ago
      I did an interesting spin to avoid it: generated Claude abstracted as a set of PRD style architectural instructions so you can do your claude-code inspired program effortlessly.

      There's even a GUI called claudia for a piecemeal extraction with a PRD.

      https://github.com/kristopolous/Claudette

      I've got a web, rust and tkinter version (for fun) right now just making sure this approach works.

      The answer is... Mostly...

      Enjoy

    • sheept2 hours ago
      I wonder if GitHub would rule it a copyright violation if the source code was rewritten by an agent, i.e. copy my answers but change a few words. Legally, if the original source code is copyrighted then an agent rewriting it likely doesn't lose that copyright, but I wonder if GitHub would go through the effort of determining whether it was a derived work.
    • yeahforsureman4 hours ago
      I guess you'd be assuming that both the original CC code and its ports are computer/AI-generated? As a lawyer, though, I'd still maintain that you wouldn't need much original human input in the CC code to kind of ruin that theory. The threshold for copyright protection isn't that high, really.
      • Aloisius3 hours ago
        The copyright office has indicated that AI generated elements lacking human contributions to the expression (rather than input) would still not be copyrightable even if some human authored elements are made to the expression as well.

        Seems like it would be a nightmare to provide evidence of what parts of a half a million line codebase were written by humans if no one bothered to track it.

        • himata41132 hours ago
          They have publically and proudly stated that vast amount of claude code was entirely vibe coded so they've already lost that argument.
  • kstenerud4 hours ago
    > It should serve as a warning to developers that the code doesn’t seem to matter, even in a product built for developers.

    Code doesn't matter IN THE EARLY DAYS.

    This is similar to what I've observed over 25 years in the industry. In a startup, the code doesn't really matter; the market fit does.

    But as time goes on your codebase has to mature, or else you end up using more and more resources on maintenance rather than innovation.

    • lelanthran3 hours ago
      > Code doesn't matter IN THE EARLY DAYS.

      > This is similar to what I've observed over 25 years in the industry. In a startup, the code doesn't really matter; the market fit does.

      > But as time goes on your codebase has to mature, or else you end up using more and more resources on maintenance rather than innovation.

      Counterpoint: Code does matter, in the early days too!

      It matters more after you have PMF, but that doesn't mean it doesn't matter pre-PMF.

      After all, the code is a step-by-step list of instructions on solving a specific pain point for a specific target market.

    • raincole3 hours ago
      Claude Code (and other agent tools) are not expected to be mature. They'll all be obsolete in two or three years, replaced by the next generation of AI tools. Everyone knows that.

      In less than four years the AI coding workflow has been overhauled at least twice: from Chat interface (ChatGPT) to editor integration (Cursor), then to CLI agent harnesses (CC/Codex). It would be crazy to assume that harnesses are the end of evolution.

      • otabdeveloper43 hours ago
        > Everyone knows that.

        Except, apparently, Anthropic - who are doing their darndest to get everyone onboard their tools as a moat. Apparently that's the only strategy to AI stickiness.

        • benterixan hour ago
          And their strategy kind of worked, right? CC is the most popular agentic coding tool. Anthropic faces competition from OpenAI (potentially better model, weaker TUI tool) and from the rest (potentially worse models, weaker TUIs). So their strategy is to develop both: make their closed model and closed tool better than competition so that when people want to vibceode they will choose their ecosystem.
          • zozbot234an hour ago
            OpenAI Codex is a much higher quality harness than Claude Code or OpenCode, and available as open source.
      • croes3 hours ago
        Claude Code 2.0 (and other agent tools) are not expected to be mature. They'll all be obsolete in two or three years, replaced by the next generation of AI tools. Everyone knows that.

        Claude Code 3.0 (and other agent tools) are not expected to be mature. They'll all be obsolete in two or three years, replaced by the next generation of AI tools. Everyone knows that.

        And so on and on and on.

        A promise of AI was mature software

        • operatingthetan2 hours ago
          If there is a market for a mature harness, surely it will be built right?
        • raincole2 hours ago
          I don't know what your point is. What you said is exactly what I expect to happen, except they might have a more creative name than "Claude Code 2.0".
    • barnabee2 hours ago
      This code matters for exactly one reason: they’re playing stupid DRM games restricting what subscriber tokens can be used for to force you to use their front ends and harnesses or buy more expensive API credits.

      Claude Code is strictly worse than e.g. OpenCode in my experience. Not much to see in the app’s code except how it authenticates itself…

      Sure I try and use all my subscription allowance with CC on side tasks, etc. but I still end up burning a bunch of API tokens (via OpenRouter) for more serious work (even the UI and ability to quickly review what the agent has done/is doing is vastly inferior in CC).

      What they have done is got me experimenting with cheaper models from other providers with those API credits.

    • julenx2 hours ago
      Code quality aside (n.b. there exist many bad quality codebases before AI), a risk I perceive as an industry is we are making the logic of our businesses dependent on a few big players.

      Given the output speed, it's practically impossible for developers to keep up, which directly impacts maintenance: the knowledge that would previously reside in-house, now is becoming dependent on having codebases pre-processed by LLMs.

      I hope in the near future local LLMs will gain traction and provide an alternative, otherwise we are in the risky path where businesses are over-reliant on a few big companies.

    • queenkjuul4 hours ago
      My CTO mentioned today how we haven't ever had to revision our API such that it breaks users, and even then, our v2 is still so compatible that client builds from 2017 are still vaguely useful for the end user.

      But now everything is, "ship as fast as is humanly possible, literally" from management, and "garbage Claude-written PRs" from devs. Trying to maintain sanity over my monorepo is impossible.

      We have nearly a century of examples of "somebody who only mostly understands making a breaking change" and decided, "what the hell, this thing is called Claude, so it can wreak havoc for as long as corporate decides"

    • tomjen34 hours ago
      You are right.

      But you can use AI to improve your codebase too. Plus models are only going to get smarter from here (or stay the same).

      • croes3 hours ago
        > Plus models are only going to get smarter from here (or stay the same).

        Training models on AI generated content leads to model collapse so they hardly become smarter if more and more code is from AI

    • trhway4 hours ago
      alternatively the code can go the way of "fast fashion" and even "3d-print your garments in the morning according to your feelings and weather and recycle at the end of the day".

      If dealing with a functionality that is splittable into microfeatures/microservices, then anything that you need right now can potentially be vibe-coded, even on the fly (and deleted afterwards). Single-use code.

      >But as time goes on your codebase has to mature, or else you end up using more and more resources on maintenance rather than innovation.

      tremendous resource sink in enterprise software. Solving it, even if making it just avoidable - may be Anthropic goes that way and leads the others - would be a huge revolution.

      • ezst3 hours ago
        I can totally wrap my head around that, and it's an interesting thought experiment, though:

        - building functionalities as components that are swappable on a whim requires a level of careful thought, abstraction and architecture that essentially is the exact opposite to ai slop

        - in this day and age we still don't make software for the sake of it, and who's financing it doesn't generally require such levels of functional flexibility (the physical world commandeering the coding isn't nearly as volatile as to justify that)

        - this comes loaded with the implication that "stuff needs to work": if you are developing software that manages inventory, orders, resources, ... you just can't take the chance to corrupt your customers data or disrupt their business processes. Shipping faster than you can test and with no accountability and no oversight is a solution to a problem I've personally never encountered in the wild

        • trhwayan hour ago
          >- building functionalities as components that are swappable on a whim requires a level of careful thought, abstraction and architecture that essentially is the exact opposite to ai slop

          that is only for humans really. Why we need these careful thought, abstraction and architecture? Because otherwise the required code becomes an unmanageable pile of spaghetti handling myriad of edge cases of abstraction leaks and unexpected side effects. Human brain can't manage it. AI can or at least soon would be able to. It will just be a large pile of AI slop.

          It may also happen that AI will also start generate good component based architecture if forced to minimize or in some other measurable way improve its slop.

      • otabdeveloper43 hours ago
        > microfeatures/microservices

        Have you seen the code generated by AI? These things converge on the "1 million lines to make an API call" pattern. They're a lot of things, but certainly not "micro".

    • shailesh-14034 hours ago
      [dead]
  • leduyquang7536 hours ago
    > Many software developers have argued that working like a pack of hyenas and shipping hundreds of commits a day without reading your code is an unsustainable way to build valuable software, but this leak suggests that maybe this isn’t true — bad code can build well-regarded products.

    The product hasn't been around long enough to decide whether such an approach is "sustainable". It is currently in a hype state and needs more time for that hype to die down and the true value to show up, as well as to see whether it becomes the 9th circle of hell to keep in working order.

    • jona777than3 hours ago
      I have flip-flopped more than ever in the last 365 days about prioritizing good code vs good product, in the AI age. This helps clarifies why.

      I have come to the conclusion that we just do not know yet. There is a part of me that believes there is a point somewhere on the grand scale where the code quality genuinely does not matter if the outcome is reliably and deterministically achieved. (As an image, I like to think of Wall—E literally compressing garbage into a cube shape.)

      This would ignore maintenance costs (time and effort inclusive.) Those matter to an established user base (people do not love change in my experience, even if it solves the problem better.)

      On the other hand, maybe software is meant to be highly personal and not widely general. For instance, I have had more fun in the past two years than the entire 15 years of coding before it, simply building small custom-fitted tools for exactly what I need. I aimed to please an audience of one. I have also done this for others. Code quality has not mattered all that much, if at all. It will be interesting to see where things go.

      • watwutan hour ago
        Imo, code quality genuinely matter for the outcome and customer happiness. Including in startups.

        But, there is also quite a lot of confident "code quality" fluff claims that have nothing to do with maintainability, robustness or performance. Fairly often, a guy claiming "the code is garbage" is not actually concerned with code quality as much as he is concerned with asserting dominance. Or is confusing own preference with quality.

      • otabdeveloper43 hours ago
        > if the outcome is reliably and deterministically achieved

        It's not. My favorite example: due to vibe coding overload literally nobody knows what configuration options OpenClaw now supports. (Not even other LLM's.)

        Their "solution" is to build a chat bot LLM that will attempt to configure OpenClaw for you, and hope for the best, fingers crossed. Yes, really.

        • operatingthetan2 hours ago
          The openclaw situation is ridiculous. Configuring it is a nightmare, even with 3 different LLMs trying to help. Then I check their docs and it says three different things. Agents will take questions and turn them into a new config file, which consists of made up settings, causing the gateway to crash.

          My setup is very simple too, just two agents, some MD files, and discord. Nothing else. These people using it for real work or managing their email and texts are in for a rough ride.

    • mergesort6 hours ago
      Hey there, author of the post here. I actually agree with this! That is in fact why I used the word maybe — my comment really was meant to be more speculative than definitive.
      • 59nadir5 hours ago
        I think one thing that goes unmentioned is that maybe code quality is really not that important for trivial things, because they can be trivially reproduced if need be. I would argue Claude Code is exactly such a project; coding agents are incredibly simple and rewriting CC wouldn't be much of a problem.

        Non-trivial things tend to be much more sensitive to code quality in my experience, and will by necessity be kept around for longer and thus be much more sensitive to maintenance issues.

        • ulrikrasmussen3 hours ago
          I think code quality is always important, but it's less important for terminal code which implements a specific set of features using robust APIs underneath. Low code quality becomes a much bigger issue when the code is supposed to deliver well-defined abstractions on which other code can rely.

          If you are a serious software developer, then you will probably be able to explain both what your code does (i.e. what spec it implements) and how it does it (what does it call, what algorithms does it use, what properties does it rely on?). With the advent of LLMs, people have started to accept not having a clue about the "how", and I fear that we are also starting to sacrifice the "what". Unless our LLMs get context windows that are large enough to hold the source of the full software stack, including LLM-generated dependencies, then I think that sacrificing the "what" is going to lead to disaster. APIs will be designed and modified with only the use cases that fit in the modifying agents context window in mind with little regard for downstream consequences and stability of behavior, because there is not even a definition of what the behavior is supposed to be.

        • rakel_rakel4 hours ago
          > maybe code quality is really not that important for trivial things

          I hear this narrative being pushed quite a bit, and it makes my spidey senses tingle every time. Secure programs are a subset of correct programs, and to write and maintain correct programs you need to have a quality mindset.

          A 0-day doesn't care if it's in a part of your computer you consider trivial or not.

          • 59nadir4 hours ago
            Intrinsically simple and straight forward problems are easier to secure even with mediocre or bad code. They've already shown that Opus 4.6 can find and report on very sophisticated security issues[0] so I'm not sure that analysis (and perhaps especially security analysis) is the biggest issue with LLMs.

            Mind you, I'm not using LLMs for professional programming since I prefer knowing everything inside and out in the code that I work on, but I have tried a bunch of different modes of use (spec-driven + entire implementation by Opus 4.6, latest Codex and Composer 2, and entirely "vibecoded", as well as minor changes) and can say that for trivial in-house things it's actually usable.

            Do I prefer to rewrite it entirely manually if I want something that I actually like? Yes. Do I think that not everything needs to be treated that way if you just want an initial version you can tinker with? Also yes.

            0: https://youtu.be/1sd26pWhfmg

            • rakel_rakelan hour ago
              > I'm not sure that analysis (and perhaps especially security analysis) is the biggest issue with LLMs.

              I was replying to the statement that "maybe code quality is really not that important for trivial things", not whether LLM's are good at analysis nor not.

              Thanks for the link though, looks like an interesting talk!

            • otabdeveloper43 hours ago
              > can find and report on some very sophisticated security issues sometimes

              Fixed it for you.

  • anematode6 hours ago
    > But then the clean room implementations started showing up. People had taken Anthropic’s source code and rewritten Claude Code from scratch in other languages like Python and Rust.

    Seems like the phrase "clean room" is the new "nonplussed"... how does this make any sense?

    • mergesort6 hours ago
      Heya, post author here. I think I was just wrong about this assertion. I got into a discussion with a copyright lawyer over on Bluesky[^1] after I wrote this and came away reasonably convinced that this wouldn’t be a valid example of a clean room implementation.

      [^1]: https://bsky.app/profile/mergesort.me/post/3mihhaliils2y

    • aeternum5 hours ago
      The most fitting method would be to be to train an LLM on the Claude Code source-code (among other data).

      Then use Anthropic's own argument that LLM output is original work and thus not subject to copyright.

    • recursive6 hours ago
      I think it means you write a spec from the implementation. Then you write a new implementation from the spec. You might go so far as to do the second part in a "clean" room.
      • m1325 hours ago
        Heh, the original being entirely vibed had me thinking of an interesting problem: if you used the same model to generate a specification, then reset the state and passed that specification back to it for implementation, the resulting code would by design be very close to the original. With enough luck (or engineering), you could even get the same exact files in some cases.

        Does this still count as clean-room? Or what if the model wasn't the same exact one, but one trained the same way on the same input material, which Anthropic never owned?

        This is going to be a decade of very interesting, and probably often hypocritical lawsuits.

      • roywiggins6 hours ago
        right. that's not what people are doing here though, at all
      • john_strinlai6 hours ago
        in a typical clean-room design, the person writing the new implementation is not supposed to have any knowledge of the original, they should only have knowledge of the specification.

        if one person writes the spec from the implementation, and then also writes the new implementation, it is not clean-room design.

        • post_below6 hours ago
          I believe the argument is that LLMs are stateless. So if the session writing the code isn't the same session that wrote the spec, it's effectively a clean room implementation.

          There are other details of course (is the old code in the training data?) but I'm not trying to weigh in on the argument one way or the other.

  • twelfthnight6 hours ago
    Seems equally valid to come out of this with the takeaway that code quality _does_ matter, because poor coding practices are what led to the leak.

    Sure, the weights are where the real value lives, but if the quality is so lax they leak their whole codebase, maybe they are just lucky they didn’t leak customer data or the model weights? If that did happen, the entire business might evaporate overnight.

    • Andrei_devan hour ago
      I look at other people's code a lot. The security issues are always boring, that's the thing. API keys sitting in the client bundle, auth middleware missing half the routes. Not clever exploits, just nobody actually reading what the AI spit out.

      Actually wait, it's worse than that. The product works, demo looks great. Then someone opens the network tab and ... yeah. "Quality doesn't matter" really just means nothing caught fire yet.

    • jamiemallers2 hours ago
      [dead]
  • r0x0r0074 hours ago
    I dont understand why so many people have the need to emphasize the code vs product battle. There is no battle. Coding/developing/software engineering is a skill, that just like any other skill has certain requirements and best practices that have to be followed in order to make a quality, maintainable and adaptable application that can stand the test of time from the software perspective. Product, features, marketing bla bla that is entrepreneurship part, and is not related to software, other than directing the software requirements,but not beacuse product people think about requirements, but instead just because they come naturally from the required features they envision.Just because programmers can write code doesnt mean they can ship good products.Just because a plumber can lay pipes doesnt mean he can run his own company or invent a new way of laying pipes. But I will tell you that a bad plumber who lays pipes and doesnt know how to connect them, bend them or shield them will surely have inferior product/service in the long term. And by the way, success of a company is measured over time, we will see where claude code will be in 10 years time when the hype dips a little, then we can say yeah the code was bad but everyone loved it and uses it still. I mean they leaked entire code online and this guy says yeah code was bad but who cares, what world are we living in? The fact that anything got leaked is a serious breach of best practices and security also at this point, something a company that used to work for DoD(W) shouldnt be doing, it can even be considered a national threat at this point. I know mistakes happen, I do them all the time, but then again the 'best' companies should be almost immune to mistakes cause stakes are high.But of course, move fast and break things is more important.Am I wrong?
    • 4 hours ago
      undefined
  • motbus313 minutes ago
    I don't know how many years of software engineering told us that new projects hastily done needs to be redone hastily later.
  • hyperlambdaan hour ago
    I created Hyperlambda (https://hyperlambda.dev), so I spend a lot of time thinking about accidental complexity and implementation surface area. One thing leaks like this keep reminding me is that a surprising amount of software risk comes from packaging and delivery details rather than the main logic people spend all their time reviewing.
  • greymanan hour ago
    From a moral perspective, I would argue that this is still theft of IP, even if it's a "clean room reimplementation". The code carries valuable information about what works and what doesn't — knowledge that Anthropic had to discover through real work and iteration. It's the same as a Chinese factory duplicating a product: they skipped the entire R&D phase and saved time and money.
    • lesostepan hour ago
      following this reasoning, wouldn't it mean that everything that Claude generates is a theft of IP?

      It was taught on massive code bases that carried valuable information about what works

  • sakopov3 hours ago
    In my opinion, the “code is garbage” argument is a moot point. Anthropic is in the business of removing humans from the SDLC. As long as their models can understand and update the code they generate it can remain garbage. They’re not optimizing for human comprehension of the output. They don’t even want you looking at it. And eventually the models will get good enough that you won’t have to.
    • boomlindean hour ago
      It doesn't seem like they're optimizing for stable, working software either, nor, apparently, a stable, working release workflow.
  • MrDresden3 hours ago
    And this is why so much software today runs extremely hot.

    It's creators clearly care not for the efficiency of how it is built, which translates directly into how it runs.

    This blog post is effectively being apologetic about the fact that this is alright, since at least they got product market fit. Except Anthropic is never going to go back and clean up the mess once (if) they become profitable.

    I doubt anyone will like how things will be in 5 years time if this trend of releasing badly engineered spaghetti continues.

  • ptnpzwqd3 hours ago
    I feel the conclusions here are a bit thin.

    Code quality tends to have an impact on more than just aesthetics - and Claude Code certainly feels like a buggy mess from an end user's perspective.

    Of course people still use Claude Code, but that is certainly because of the underlying models first and foremost. Most products don't have such a moat and would not nearly see as much tolerance from end users. If the Max subscriptions could be used with other harnesses, I am sure Anthropic would have to compete harder on the quality of the harness (to be fair, most AI based tooling seems pretty alpha these days, but eventually things will stabilize).

    Polish is not everything, clearly, but it is a factor, and I feel Claude Code is maybe the worst example to use here, as it doesn't at all generalize to most other products.

  • moezd2 hours ago
    Code quality matters when it becomes the code you have to maintain after six months. I'm honestly surprised by how some features of Claude Code seem to be held together with gum and duct tape.

    That being said, if you're just beginning and looking for your market fit, or pitching to investors with a flashy demo, it doesn't need to be an architectural miracle, in fact it will waste your time.

    • motbus32 hours ago
      It only holds because (1) people are inclined to accept failures as pain of using something cutting edge and (2) you kinda don't know what is failing anymore.

      Extra one (3) We are getting super lenient with major failures and having a services that has only one 9 on reliability charts as norm.

  • ggrab2 hours ago
    I don't think it's vibe coded garbage. Sure, the 3000-line print.ts is terrible, but there's some good patterns in there that were definitely prompted in by some experienced engineers -- the feature flag setup, the `..I_VERIFIED_THIS_IS_NOT_PATH_OR_CODE` funny type hints, the overall structure. Just the usual signs this started as a PoC but quickly evolved into something much bigger. The codebase is a really interesting read.
  • krisgenre4 hours ago
    Reminds me of a question that I asked couple of years ago - https://news.ycombinator.com/item?id=37176689
  • grey-area5 hours ago
    Points from the article.

    1. The code is garbage and this means the end of software.

    Now try maintaining it.

    2. Code doesn’t matter (the same point restated).

    No, we shouldn’t accept garbage code that breaks e.g. login as an acceptable cost of business.

    3. It’s about product market fit.

    OK, but what happens after product market fit when your code is hot garbage that nobody understands?

    4. Anthropic can’t defend the copyright of their leaked code.

    This I agree with and they are hoist by their own petard. Would anyone want the garbage though?

    5. This leak doesn’t matter

    I agree with the author but for different reasons - the value is the models, which are incredibly expensive to train, not the badly written scaffold surrounding it.

    We also should not mistake current market value for use value.

    Unlike the author who seems to have fully signed up for the LLM hype train I don’t see this as meaning code is dead, it’s an illustration of where fully relying on generative AI will take you - to a garbage unmaintainable mess which must be a nightmare to work with for humans or LLMs.

    • operatingthetan2 hours ago
      I'm a little disappointed that a bunch of engineers with unlimited access to Opus didn't do a better job.
    • tipiirai3 hours ago
      What exactly makes you say that "the author who seems to have fully signed up for the LLM hype train"?

      I feel the author is just stating the obvious: code quality has very little to do with whether a product succeeds

      • grey-area2 hours ago
        OK, but what happens after product market fit when your code is hot garbage that nobody understands?
  • Grisu_FTP3 hours ago
    Afaik you can run Claude Code locally but every single demo i see uses it exclusively with apis, so are the local models already good enough to be worth it or is the only reasonable use for claude code with cloud models?
    • JoshuaDavid3 hours ago
      Claude Code's main advantage is that it's the only TOS-compliant way to access subscription Claude tokens, which cost about 10% as much as pay-as-you-go Claude API tokens.
      • Grisu_FTPan hour ago
        Well to be honest, even 5% would probably be too much for me (even tho i dont know the pricing of tokens or consumption of tokens)

        I would just want to test around a bit locally, maybe let it do its thing over a weekend just to see the result and then stop it again

  • hbarka3 hours ago
    OP should expand on #1, why he thinks it’s garbage. Claude Code is the REPL harness Anthropic built, can read, write, edit, bash. Pi, Gemini, Codex do the same, but they are not hinted as garbage. Where’s the beef?
  • uduni4 hours ago
    " The real value in the AI ecosystem isn’t the model or the harness — it’s the integration of both working seamlessly together. "

    Wut? The value in the ecosystem is the model. Harnesses are simple. Great models work nearly identically in every harness

    • Alifatisk4 hours ago
      I wouldn’t say harnesses is simple. They do a lot of things that we aren’t thinking of. I learned that a good harnesses is as valuable as the model. But obviously the model is what carries the whole thing.

      I tried to build my own harness once. The amount of work that is required is incredible. From how external memory is managed per session to the techniques to save on the context window, for example, you do not want the llm to read in whole files, instead you give it the capability to read chunks from offsets, but then what should stay in context and what should be pruned.

      After that you have to start designing the think - plan - generate - evaluation pipeline. A learning moment for me here was to split up when the llm is evaluating the work, because same lllm who did the work should not evaluate itself, it introduces a bias. Then you realize you need subagents too and statt wondering how their context will be handled (maybe return a summarized version to the main llm?).

      And then you have to start thinking about integration with mcp servers and how the llm should invoke things like tools, prompts and resources from each mcp. I learned llms, especially the smaller ones tend to hiccup and return malformed json format.

      At some point I started wondering about just throwing everything and just look at PydanticAi or Langchain or Langgraph or Microsoft Autogen to operate everything between the llm and mcps. Its quite difficult to make something like this work well, especially for long horizontal tasks.

      • uduni3 hours ago
        I've been running a custom harness in production for months (code gen on cloud boxes), it's quite simple (<500 LOC). The model can use sed or other nice bash tricks to efficiently read. U really don't need any tool besides bash plus a good system prompt. Subagents are the same as the main agent (end with a summary). U can just remove tool results (oldest first) to save context, the model can read again if it needs to. High quality memory is the only difficult part. But that can also be solved with high quality documentation
    • steve_adams_864 hours ago
      Harnesses are simple (kind of? Some certainly aren't, but I'd agree that they can be simple) but they deliver a ton of value. They have a significant ROI.

      I agree that good models have more value because a harness can't magically make a bad model good, but there's a lot that would be inordinately difficult without a proper harness.

      Keeping models on rails is still important, if not essential. Great models might behave similarly in the same harness, but I suppose the value prop is that they wouldn't behave as well on the same task without a good harness.

      • uduni3 hours ago
        Sure... A good harness can be built in 1 day tho. A good model is 1000x more sophisticated
        • steve_adams_863 hours ago
          Okay, fair. I was thinking in the context of actually using the tools, but yes, in the bigger picture the model is worth far more.
    • theshrike792 hours ago
      There is a reason why Copilot+Opus4.6 is shit, while Claude Code + Opus 4.6 produces excellent results.

      The harness matters A LOT.

      The model is the engine, the harness is the driver and chassis. Even the best top of the line engine in a shitty car driven by a bad driver won't win any races.

    • DANmode3 hours ago
      The model defines the ceiling, but the harness determines how much of that ceiling you actually reach.

      It is not everyone’s experience that models work the same in every harness.

  • piker4 hours ago
    > The success of Claude Code and Cursor at the higher end of the market shows that even the people pickiest about their software (developers) will use your software regardless of how good the code is.

    Seems wrong. Devs will whine, moan and nitpick about even free software but they can understand failure modes, navigate around bugs and file issues on GitHub. The quality bar is 10-100x amongst non-techno-savvy folks and enterprise users that are paying for your software. They’re far more “picky”.

  • slopinthebag5 hours ago
    Claude Code proves you don't need quality code — you just need hundreds of billions of dollars to produce a best-in-class LLM and then use your legal team to force the extreamly subsidised usage of it through your own agent harness. Or in other words, shitty software + massive moat = users.

    Seriously, if Anthropic were like oAI and let you use their subscription plans with any agent harness, how many users would CC instantly start bleeding? They're #39 in terminal bench and they get beaten by a harness that provides a single tool: tmux. You can literally get better results by giving Opus 4.6 only a tmux session and having it do everything with bash commands.

    It seems premature to make sweeping claims about code quality, especially since the main reason to desire a well architected codebase is for development over the long haul.

  • boomlinde4 hours ago
    How likely is it that the broken release workflow that produced the leak was Claude's own work?
  • 5 hours ago
    undefined
  • Finbarr4 hours ago
    Who cares that the code is garbage? As the models get bigger and more powerful it will be trivial to fully refactor the whole codebase. It’s coming sooner than you think.
  • CrzyLngPwd3 hours ago
    99.9999% of consumers never give code a single thought.

    Most corporations never give code a single thought.

    In the race to market, quality always suffers, and with such high stakes, it should surprise no one that AI companies are vibe-coding their own slop.

  • komali24 hours ago
    > bad code can build well-regarded products.

    Yes, exactly. Products.

    It seems like me and all the engineers I've known always have this established dichotomy: engineers, who want to write good code and to think a lot about user needs, and project managers/ executives/sales people, who want to make the non-negative numbers on accounting documents larger.

    The truth is that to write "good software," you do need to take care, review code, not single-shot vibe code and not let LLMs run rampant. The other truth is that good software is not necessary good product; the converse is also true: bad product doesn't necessarily mean bad software. However there's not really a correlation, as this article points out: terrible software can be great product! In fact if writing terrible software lets you shit out more features, more quickly, you'll probably come ahead in business world than someone carefully writing good software but releasing more slowly. That's because the priorities and incentives in business world are often in contradiction to priorities and incentives in human world.

    I think this is hard to grasp for those of us who have been taught our whole lives that money is a good scorekeeper for quality and efficacy. In reality it's absolutely not. Money is Disney bucks recording who's doing Disney World in the most optimal way. Outside of Disney World, your optimal in-park behavior is often suboptimal for out-of-park needs. The problem is we've mistaken Disney World for all of reality, or, let Walt Disney enclose our globe within the boundaries of his park.

    > The object which labor produces confronts it as something alien, as a power independent of the producer.

  • rustystump2 hours ago
    I think this misses the target.

    First, the twitter quote is standard toxic clapback nonsense. Gambling makes billions and does not add any value. Even facebook can argue it adds more value than gambling so this one is a dud.

    People use claud code because of claud the model and not claud the harness. Cursor or a hacked up agent loop using opus or whatever are about as good. The magic is in the model not the harness here. This isnt to say the hardness is doesnt do anything.

    The other bit this misses is that yes the product matters more then the code, and if the product burns battery/ram/etc doing nothing because the ai has crappy code or maybe something leaks or has a security issue, then that impacts the product.

  • sunilkumarai2 hours ago
    [dead]
  • jeremie_strand5 hours ago
    [dead]
  • panavm6 hours ago
    [dead]
  • shailesh-14034 hours ago
    [dead]
  • michaelashley295 hours ago
    [dead]
  • gtsop3 hours ago
    [dead]
  • ValveFan69694 hours ago
    [dead]
  • s998502 hours ago
    [flagged]
  • pregseahorses6 hours ago
    They just said this was an April Fools joke.
    • gfosco5 hours ago
      No you fell for someone elses joke.
    • dodu_4 hours ago
      jokesonthemiwasonlypretending.png
  • miki1232112 hours ago
    > So of course the first thing people did was point and laugh.

    This just validates my theory that open-sourcing old code that people have sentimental attachments for, and that you won't ever make any money off of, again is actually a terrible idea.

    Everything about this leak is a long list of arguments why you shouldn't ever open source anything.

    We, the developer community, have really dropped the ball here.

    • tmoravec2 hours ago
      Why? Reddit is full of bug fixes and improved forks. Seems like it's Anthropic that's dropping the ball with it's takedown requests instead of engaging with the community.
    • 2 hours ago
      undefined