45 pointsby guaxa day ago14 comments
  • elicksaura day ago
    When I read stories about people 10x’ing their productivity with AI, I interpret them as having the same vibe as these “solvers”.
    • It can 10x certain tasks. If you're making an interface to a library for a DSL and have documentation for the library, Claude 3.7 will very accurately fill that out for you extremely quickly.

      There are a lot of tasks that follow a pretty predictable but not automatable pattern that LLMs do a great job at, but that you couldn't leverage macros or other powerful editor tools to complete. Getting good at LLMs can greatly increase your productivity. But unless your job is just gluing well-defined interfaces together, spinning up demos, or writing documentation for existing code, you won't get 10x. 2x is very realistic for most programming jobs, though. Some tasks would potentially be below 1x, especially low level, novel, thorny logic. I'd never use it for developing new cryptographic primitives or a novel compression codec, for instance.

      • elicksaura day ago
        I’ve tried to use LLMs for programming. They always fail in extremely predictable ways for the text prediction programs that they are.

        The phrasing you use is all marketing. When people share actual chat threads that they claim helped them, they always have the same issues I observe in my own use.

        • MyOutfitIsVague21 hours ago
          I'm not using any marketing; I'm sharing my experience. At work, I recently had to build an interface for JNI in Lua. I had to do a lot of the fiddly bits myself, but most of the interface the LLM generated without a problem, just being fed my existing Lua interfaces and the JNI docs. I had to do some clean up and optimization, but it definitely saved time. I then fed it the JTOpen Java docs and had it use the interface to make a wrapper around JTOpen for communicating with IBM i message queues (which was the original goal of the project), and it did enough of a good job to definitely be worth the expense (notably, Claude did nearly perfect at generating JNI method signatures that it inferred from the Javadocs; as a counterpoint, it did generate interfaces for a couple methods that didn't actually exist). These things are really really good at generating interface glue, generating good documentation, generating tests, and limited refactoring.

          They won't do the thinking for you. You still have to know what you want out of it, you have to know your endpoint, and you have to be able to review and understand every line out of it, but they do save time. You might have tried to use them in situations where they aren't appropriate. They are text prediction programs, but they're extremely powerful text prediction programs. If you accept their unreliability and give them the right patterns and context to generate from, they can write a thousand lines of code in one go that look identical to what you would have written, and it takes 30 seconds to generate and a few minutes to review and test, compared to the half an hour at least that it would have taken otherwise. It can easily yield 10 times efficiency gain in places where it's really appropriate.

          That said, it's not perfect by any means. To be really what I want, I need the same quality significantly faster, cheaper, and local. If my company wasn't paying for it, I certainly wouldn't pay to use this thing myself. This costs between $0.05 and $1 per request, which gets expensive fast, and the much cheaper models are less capable enough to save me so little time as to not be worth the bother really.

        • Foreignborna day ago
          i don’t have a stake in this at all, but i've seen dozens and dozens of comments of exactly what you’re asking here on HN.

          People have been sharing that, plus the problems they’ve solved or products they’ve made.

          Though, most of these workflows are no longer really just “chats” at like chatgpt.com. A lot of that has changed.

          i’m not saying you need to do it, or change anything, but… i’ve seen people do shit with a computer lately that 6 months ago would seem weird and alien.

          • jakubmazaneca day ago
            > i’ve seen people do shit with a computer lately that 6 months ago would seem weird and alien

            Do you have video?

    • vultoura day ago
      I think the dystopian Apple Intelligence commercials summed it up perfectly. It sure was a 10x improvement when AI generated the email for that lazy bum, but anyone who’s already bringing value will never see anywhere near those numbers.
    • hombre_fatala day ago
      On the other hand, this mentality can keep you from figuring out how to use AI as a tool that works for you. You should speak to people who are using it productively since most of the time its work is verifiable.
      • swatcodera day ago
        Discussions with the "people who are using it productively" happen all the time here on HN and among our colleages in real life. It's not like you can escape it.

        But those discussions don't crest the hill because the "verified" work consistently either fails to be addressing a challenge that the skeptics encounter or has qualities that the skeptics can't imagine themselves possessing.

        The reality is that we all approach our work differently and all carry different standards in what we accept in the work we produce.

        It's clear that generative AI genuinely helps some people improve their work or ease their burden of producing it, but it should also be clear at this point that the people it doesn't seem to help are not being failed for lack of having learned about its use cases. We're 2+ years in and are saturated with the hype and with such examples. It's just not as universally useful as some people (for whatever reason) want to believe.

        • hombre_fatala day ago
          On the flip side, I think a lot of skepticism is just cynicism, and you need to ensure that you are evaluating the tool objectively, not just doing the HN sneer the same way we dismiss Javascript or "web".

          It's fine if you've found a use-case that LLMs aren't so helpful, like nonfiction writing that you don't want to have to constantly verify for hallucinations. Obviously those exist.

          But don't forget that they are calling into question whether other people really are super productive with LLMs, and that hints at a self-limiting belief.

          • elicksaura day ago
            I reject the implication that the problem is me or other people who are critical of AI here. I’ve given it a fair shake. Why is user error or literally here a “self-limiting belief” always implied or stated? Almost like I can’t use AI because of a moral failing? C’mon…

            It honestly sounds like a cult where I don’t know how to read the tea leaves and believe enough. I’ll pass happily and be so self limited :)

            When AI is actually useful, there won’t be a debate.

            • kelseyfroga day ago
              Rather a virtue - the protestant work ethic is a mass.social assumption. It's assumed that hard work is a virtue and that people that find a way to cheat will justly receive their consequences. When this belief primes us to search for flaws, we of course find them.

              AI even with flaws is good enough. I'm happy to pick up the profit on the difference in our beliefs and behaviors. It's easy to simply out compete.

          • swatcodera day ago
            That's two different discussions.

            Your original comment was suggesting that people who haven't found a use-case that works for them just haven't looked hard enough, despite how inescapable the topic and tools are.

            Some hypothetical person that believes nobody might have a personally effective use case is a different one than what you seemed to be writing about, and completely unrelated to what I was writing about.

            To be fair, the distinction gets muddy in both directions because some people invest their egos into/against this kind of thing and can't just accept differences as they are. But at that point, with those people, you're just stuck in the same class of endless flamewar debate as emacs vs vim or Xbox vs Playstation and there's really no point in giving time or attention to that.

      • n4r9a day ago
        It's worth keeping a toe in the water, but for now the productivity hype seems way overblown.

        I recently asked Claude to help deserialize a JSON Post request in an asp.net web API so that absent or null JSON properties get set to the coded default values. They gave me what was nearly the right answer (set NullValueHandling to Ignore) but led me down a merry garden path by saying I should also set DefaultValueHandling to Populate. Every time I said it wasn't working and gave the error message, it came up with additional code to put in, which never helped. Eventually I just started toying around with it myself and realised the solution, but it was maybe half a day down the drain.

        Perhaps I could have taken a different approach from the start, but still, my impression is that the tools are not what they're made out to be.

        • oxidanta day ago
          I've had similar issues and the trick I've found is to jump to Google/SO to find examples or docs as soon as Claude stops making progress.

          If you got NullValueHandling, jumping to the library documentation would give you immediate accurate results.

    • olddustytraila day ago
      That still doesn't mean it's not true...
  • Xceleratea day ago
    > Some years ago, Sanborn began charging $50 to review solutions, providing a speed bump to filter out wild guesses and nut cases

    Oh man, oh man, this just gave me a great startup idea. What is the total addressable market of nutcases willing to pay $50 or more to have someone check whether they have cracked some difficult code? Perhaps even with a giant prize attached for cracking it?

    • solardeva day ago
      Sounds like the NYT word game subscriptions
    • mcphage11 hours ago
      Even better: create an unbreakable code (one with no actual solution, but with a fake solution). Give people a few years to attempt to solve it, while you acquire stock in one of the smaller AI companies. Then give someone the fake solution, have them announce to the world that they solved your code with the help of the AI software you’ve invested in. Rise the wave, then sell it before everyone figures out the ruse.
  • chipha day ago
    If you'd like to see a similar work by Jim Sanborn, Cyrillic Projector is at the University of North Carolina Charlotte behind the Fretwell Building. Way easier to see in person than CIA headquarters!

    It has a light inside and if you visit at night, the reverse-cut lettering will be readable on the walkway around it.

    https://maps.app.goo.gl/tFbx6kBkMgSMDGGY8

  • haswella day ago
    This story summarizes one of my biggest concerns/frustrations with the rise of these LLM tools.

    The confident incorrectness they breed is a problem.

    The impact in online social spaces is already increasingly obvious.

    • MrMcCalla day ago
      > The confident incorrectness they breed is a problem.

      Looking at America here in 2025, it sure seems like the confidently incorrect are confidently and successfully convincing the submorons that they are correct, despite their both being incorrect.

      I sure seems like a kind of tensegrity of willful ignorances.

      But definitely there are a shitton of folks that are both far too stupid to know how stupid they are, while also being way too overconfident that they're smarter than the actual smart folks. Dunning-Kruger really exposed the majorities.

      And there's that quote about change only coming when the old generation dies out; not because they willingly learned and humbly levelled themselves up.

  • kewhoa day ago
    > What took 35 years and even the NSA with all their resources could not do I was able to do in only 3 hours before I even had my morning coffee,” it began, before the writer showed Sanborn what they believed to be the cosmically elusive solution. “History’s rewritten,” wrote the submitter. “no errors 100% cracked.”

    I liked the snark from gizmodos report:

    > > The smugness is, frankly, inexplicable. Even if they did successfully crack Sanborn’s code using AI (which, for the record, Sanborn says they haven’t even gotten close), what is it about asking a machine to do the work for you that generates such self-satisfaction?

    Hear, hear.

    https://gizmodo.com/chatbots-have-convinced-idiots-that-they...

    • MrMcCalla day ago
      We live among idiots, my friend. And a shitton of them are right here on HN, and with damn high karma because there's nothing an idiot loves more than another idiot to validate their idiocy. And they always go Yosemite Sam on anyone who explains exactly what kind of stupid they are.

      What they lack in intelligence, however, they MORE than make up for in confidence.

      "Same as it ever was." --Talking Heads

      • TMWNN17 hours ago
        I don't disagree that your comment applies to HN, but it applies to Reddit 100X more (in terms of both width and depth).
        • MrMcCall16 hours ago
          I don't doubt that that's true, but my surfing there has been effectively nonexistent, so I know that I don't know. I'd hazard a guess, however, that there is a great deal of variance between different subreddits.
  • trescenzia day ago
    This is one of the more major risks of chatbot style generative AI. People are very open to well argued answers to problems even if the answer is wrong. I was reading this article about how those with less educational attainment seem to use chatbots more[1].

    On one hand this is amazing because it increases access to good enough writing and information. On the other hand it raises the chances that people, regardless of educational attainment, are convinced that falsehoods are true.

    1: https://arstechnica.com/ai/2025/03/researchers-surprised-to-...

    • ethbr1a day ago
      LLMs are literally truth-sounding machines.

      Which isn't to diminish their utility, but is to say it's smart to treat them like used car salesmen.

      Maybe they're telling a true-sounding truth... or maybe it's just true-sounding.

      • trescenzia day ago
        Right their purpose is to produce bullshit in the most technical sense[1]. There is no concern for the actual truth, just that the output sounds good.

        Edit: That’s not to discount the value of said bullshit or to say they are never correct. Simply that correctness isn’t what they are optimized for.

        1: https://en.wikipedia.org/wiki/On_Bullshit

      • timbit42a day ago
        The need to add an inference engine to them for fact checking.
    • roenxia day ago
      It raises an interesting theoretical problem. If the AI comes up with a good argument that translates a coded message into something legible can it truly be said to be wrong? It'd be more of a flaw in the art installation.

      The problem with asking people to interpret gobbledegook is that it might map to English in any number of ways.

  • snowwrestlera day ago
    I love entertaining the possibility that the 4th panel of Kryptos is not actually solvable, and that everyone is spending tons of time and energy trying to know secret information has no practical use, and doesn’t even exist.

    I feel like that would be a strong artistic statement about the CIA and intelligence agencies in general. Do people reluctantly work to know every secret because it’s actually necessary for security? Or do some people just want to know every secret, and “security” is the handiest excuse for them to pursue that?

  • IshKebaba day ago
    > And Sanborn is getting ticked off

    ...

    > Some years ago, Sanborn began charging $50 to review solutions, providing a speed bump to filter out wild guesses and nut cases.

    Yeah I suspect he isn't that ticked off. I'm happy to take over reviewing solutions if he likes!

  • bell-cota day ago
    But it feels SO good to believe that you (with your AI sidekick) are some sort of uber-awesome genius!

    Especially when the downsides of being wrong are nill.

    • haswella day ago
      > Especially when the downsides of being wrong are nill.

      The question is: are the downsides of being wrong because of an LLM actually nil?

      While this is a story of a harmless contest, I think it represents something much bigger and perhaps far less harmless.

    • mcphage11 hours ago
      What if the downside of being wrong is $50?
  • picafrosta day ago
    I am reminded of a brief quip in Neal Stephenson's The Diamond Age where a character comments on remembering a time before "AI" had been correctly rebranded to "PI": pseudo-intelligence.

    The valuation-perception-driven hyperbole around these Dunning-Kruger machines does not help the average person trying to bat above their level.

  • spacecadeta day ago
    As an artist, I would just setup a small/local AI to auto respond to the bullshit emails and sit back and collect the $50. That would be art in itself.
    • RajT88a day ago
      Sign the response as Banksy while you are at it.
  • a day ago
    undefined
  • teddyha day ago
    [flagged]
    • minitecha day ago
      Full quote, showing that this is something the artist in question is talking about:

      > Sanborn, a climate-conscious friend of the Earth who lives on a small island on the Chesapeake Bay, is also appalled by the amount of energy that it takes to produce generative AI, and AI’s fabricated answers. Adding to the annoyance is that some of the would-be codebreakers are touting their collaboration with Grok 3, which is made by Elon Musk’s xAI. The same Musk who, despite good deeds with Tesla, now works for an administration determined to reverse any progress on mitigating climate change. “That’s a little twist of the ice pick,” he says.

      Nothing “stealth” or “ad copy” about it.

      • teddyh19 hours ago
        The sentence I quoted is not presented as an opinion of the artist, or even of the article author; it is presented as an incontrovertible fact. As an opinion of the artist, it would have had at best a very slight relevance. However, presented as pure fact, it is irrelevant at best and seems suspicious.
    • wahnfriedena day ago
      It's annoying to read surface level critiques that skim articles for Musk denigration or wokeisms and pull them out of their context and misrepresent their point to clown on the author