426 pointsby indigodaddya day ago19 comments
  • danielvfa day ago
    I handle reports for a one million dollar bug bounty program.

    AI spam is bad. We've also never had a valid report from an by an LLM (that we could tell).

    People using them will take any being told why a bug report is not valid, questions, or asks for clarification and run them back through the same confused LLM. The second pass through generates even deeper nonsense.

    It's making even responding with anything but "closed as spam" not worth the time.

    I believe that one day there will be great code examining security tools. But people believe in their hearts that that day is today, and that they are riding the backs of fire breathing hack dragons. It's the people that concern me. They cannot tell the difference between truth and garbage.

    • phs318u16 hours ago
      >It's the people that concern me. They cannot tell the difference between truth and garbage.

      Suffice to say, this statement is an accurate assessment of the current state of many more domains than merely software security.

      • immibis4 hours ago
        This has been going for years before AI - they say we live in a "post-truth society". The generation and non-immediate-rejection of AI slop reports could be another manifestation of post-truth rather than a cause of it.
    • Seb-C14 hours ago
      > I believe that one day there will be great code examining security tools.

      As for programming, I think that we will simply continue to have incrementally better tools based on sane and appropriate technologies, as we have had forever.

      What I'm sure about is that no such tool can come out of anything based on natural language, because it's simply the worst possible interface to interact with a computer.

    • VladVladikoffa day ago
      This sounds more like an influx of scammers than security researchers leaning too hard on AI tools. The main problem is the bounty structure. And I don’t think these influx of low quality reports will go away, or even get any less aggressive as long as there is money to attract the scammers. Perhaps these bug bounty programs need to develop an automatic pass/fail tester of all submitted bug code, to ensure the reporter really found a bug, before the report is submitted to the vendor.
      • rwmj9 hours ago
        It's unfortunately widespread. We don't offer bug bounties, but we still get obviously LLM-generated "security reports" which are just nonsense and waste our time. I think the motivation may be trying to get credit for contributing to open source projects.
      • Simply charge a fee to submit a report. At 1% of the payment for low bounties it's perfectly valid. Maybe progressively scale that down a bit as the bounty goes up. But still for a $50k bounty you know is correct it's only $500.
        • cedws11 hours ago
          If you charge a fee the motivation for good samaritan reports goes to zero.
          • bloppe10 hours ago
            That's why they offer cash bounties. You don't need to charge a fee if there is no bounty (aka an actual good Samaritan situation), cuz then there's no incentive to flood it with slop
            • LocalH5 hours ago
              Another comment in this overall thread indicated that they still receive LLM slop despite not offering bounties. Clout can be as alluring a drug as money.
              • Vilian4 hours ago
                Curl has dozen of garbage bug reports made using AI where even the author can't point where the bug if, they answer with "the AI said so it's true"
        • Jean-Papoulos11 hours ago
          No need to make it a percentage ; charge $1 and the spammers will stop extremely quickly, since none of their reports are valid.

          But I do think established individual and institutes should have free access ; leave a choice between going through an identification process and paying the fee. If it's such a big problem that you REALLY need to do something ; otherwise just keep marking as spam.

        • ponector14 hours ago
          You are adding more incentive to go directly to black market to sell vulnerability.

          Also I've heard many times cases when company refused to pay bounty for any reason.

          And taxes, how you'll tax it internationally? Sales tax? VAT?

        • lucyjojo20 hours ago
          gentle reminder that the median salary of a programmer in japan is 60k USD a year. 500 usd is a lot of money (i would not be able to afford it personally).

          i suspect 1usd would do the job perfectly fine without cutting out normal non-american people.

        • justsid21 hours ago
          Could also be made refundable when the bug report is found to be valid. Although of course the problem then becomes some kid somewhere who is into computers and hacking find something but can’t easily report it because the barrier to entry is too high now. I don’t think there is a good solution unfortunately.
          • rogerrogerr21 hours ago
            That kid could find a security expert - it’s easy to do - and they could both validate it and post the money. I don’t think it would be hard to find someone with $10k with the right skill set.

            Pick someone already rich so the reputational damage from stealing your bounty exceeds the temptation. The repeat speakers list at defcon would be a decent place to start.

          • edoceo21 hours ago
            The world of AI slop needs a human assertion component. Like. I'm real and stake a permanent reputation on the claim I'm making. An I'm actually human gate.
        • imtringued12 hours ago
          Why charge a fee? All you need is a reputation system where low reputation bounty hunters need a reputable person to vouch for them. If it turns out to be false, both take a hit. If true, the voucher gets to be a co-author and a share in the bounty.
          • Snacklive10 hours ago
            That's just a way to create a toxic environment filled with elitism similar to StackOverflow
    • datatrashfirea day ago
      > I believe that one day there will be great code examining security tools.

      Based on current state, what makes you think this is given?

      • ASalazarMX4 hours ago
        The improvement history of tools beside LLMs, I suspect. First we had syntax highlighting, and we were wondered. Now we have fuzzers and sandbox malware analysis, who knows what the future will bring?
  • unsnap_bicepsa day ago
    For those of you who don't want to click into linked in, https://hackerone.com/reports/3125832 is the latest example of a invalid curl report
    • harrisia day ago
      This is interesting because they've apparently made a couple thousand dollars reporting things to other companies. Is it just a case of a broken clock being right twice a day? Seems like a terrible use of everyone's time and money. I find it hard to believe a random person on the internet using ChatGPT is worth $1000.
      • billyoneala day ago
        There are places that will pay bounties on even very flimsy reports to avoid the press / perception that they aren't responding to researchers. But that's only going to remain as long as a very small number of people are doing this.

        It's easy for reputational damage to exceed $1'000, but if 1000 people do this...

        • cratermoon4 hours ago
          One might even call it reputational blackmail. "Give me $1000 for this invalid/useless bug report or I'll go to the most click-baity incompetent tech press outlets with how your product is the worst thing since ILUVYOU."
      • bluGilla day ago
        $1000 is cheap... The real question is when will companies become wise to this scam?

        Most companies make you fill in expense reports for every trivial purchase. It would be cheaper to just let employees take the cash - and most employees are honest enough. However the dishonest employee isn't why they do expense reports (there are other ways to catch dishonest employees). There used to be a scam where someone would just send a bill for "services" and those got paid often enough until companies realized the costs and started making everyone do the expense reports so they could track the little expenses.

    • jasinjamesa day ago
      Can someone explain the ip address in the hackerone profile[0]? I can't tell if 139.224.130.174 is a reference to something real or just hallucinated by the LLM to look "cool". Wikipedia says that this /8 is controlled by "MIX"[1] but my google-fu is failing me atm.

      [0] https://hackerone.com/evilginx?type=user [1] https://en.wikipedia.org/wiki/List_of_assigned_/8_IPv4_addre...

      • dpifke18 hours ago
        Per WHOIS, it's assigned to Alibaba Cloud (could be a VM there):

          inetnum:        139.224.0.0 - 139.224.255.255
          netname:        ALISOFT
          descr:          Aliyun Computing Co., LTD
          descr:          5F, Builing D, the West Lake International Plaza of S&T
          descr:          No.391 Wen'er Road, Hangzhou, Zhejiang, China, 310099
          country:        CN
          admin-c:        ZM1015-AP
          tech-c:         ZM877-AP
          tech-c:         ZM876-AP
          tech-c:         ZM875-AP
          abuse-c:        AC1601-AP
          status:         ALLOCATED PORTABLE
          mnt-by:         MAINT-CNNIC-AP
          mnt-irt:        IRT-ALISOFT-CN
          last-modified:  2023-11-28T00:57:06Z
          source:         APNIC
    • darkoob1215 hours ago
      You can tell it's ChatGPT from the stupid icon. In one of the iterations they started using thses emojis which are disturbing for me. The answer to the first question has obvious ChatGPT writing style.
    • Aachena day ago
      Daniel posting about the LinkedIn post: https://mastodon.social/@bagder/114455578286549482

      Recent toots on account has the news as well

    • nneonneoa day ago
      Good god did they hallucinate the segmentation fault and the resulting GDB trace too? Given that the diffs don’t even apply and the functions don’t even exist, I guess the answer is yes - in which case, this is truly a new low for AI slop bug reports.
      • teroma day ago
        The git commit hashes in the diff are interesting: 1a2b3c4..d4e5f6a

        I think my wetware pattern-matching brain spots a pattern there.

      • bluGilla day ago
        An real report would have a GDB trace that looks like that, so it isn't hard to create such a trace. Many of us could create a real looking GDB trace just as well by hand - it would be tedious, boring, and pointless but we could.
        • nneonneoa day ago
          Oh, I'm fully aware an LLM can hallucinate a GDB trace just fine.

          My complaint is: if you're trying to use an AI to help you find bugs, you'd sincerely hope that they would have *some* attempt to actually run the exploit. Having the LLM invent fake evidence that you have done so, when you haven't, is just evil, and should be resulting in these people being kicked straight off H1 completely.

          • bluGill18 hours ago
            That means doing work. I can get a llm to write up a bugus report in minutes and then whatever value comes frome it. Checking the report is real would take time.
    • xk_id18 hours ago
      Not sure what timeline this is anymore where a tech website loads up a completely blank page on my mobile device.
      • seanp2k218 hours ago
        Welcome to the web in 2025, where it takes 5MB of JS and everything else to load a blog post containing 640B of text.
  • bogwoga day ago
    If I wanted to slip a vulnerability into a major open source project with a lot of eyes on it, using AI to DDOS their vulnerability reports so they're less likely to find a real report from someone who caught me seems like an obvious (and easy) step.

    Looking at one of the bogus reports, it doesn't even seem like a real person. Why do this if you're not trying to gain recognition?

    • jshearda day ago
      > Why do this if you're not trying to gain recognition?

      They're doing it for money, a handful of their reports did result in payouts. Those reports aren't public though, so there's no way to know if they actually found real bugs or the reviewer rubber-stamped them without doing their due diligence.

    • 0x500x796 hours ago
      It should be called "Denial of Attention" attack!
  • 2 hours ago
    undefined
  • vessenesa day ago
    Reading the straw that broke the camel's back commit illustrates the problem really well: https://hackerone.com/reports/3125832 . This shit must be infuriating to dig through.

    I wonder if reputation systems might work here - you could give anyone who id's with an AML/KYC provider some reputation, enough for two or three reports, let people earn reputation digging through zero rep submissions and give someone like 10,000 reputation for each accurate vulnerability found, and 100s for any accurate promoted vulnerabilities. This would let people interact anonymously if they want to edit, quickly if they found something important and are willing to AML/KYC, and privilege quality people.

    Either way, AI is definitely changing economics of this stuff, in this case enshittifying first.

    • bflescha day ago
      there is a reputation system already. according to hackerone reputation system, it is a credible reporter. it's really bad
      • hedoraa day ago
        The vast majority of developers are 10-100x more likely to find a security hole in a random tool than spend time improving their reputation on a bug bounty site that pays < 10% their salary.

        That makes it extremely hard to build a reputation system for a site like that. Almost all the accounts are going to be spam, and the highest quality accounts are going to freshly created and take ~ 1 action on the platform.

    • Aachena day ago
      Or a deposit system: pay 2€ for a human to read this message, you'll get it back if it's not spam

      What if the human marks it as spam but you're actually legit? Deposit another 2€ to have the platform (like Hackerone or whichever you're reporting via) give a second opinion, you'll get the 4€ back if you weren't spamming. What to do with the proceeds from spammers? The first X euros of spam reports go to upkeep of the platform, the rest to a good cause defined by the projects to whom the reports were submitted because they were the ones who had to deal with reading the slop so they get at least this much out of it

      Raise deposit cost so long as slop volume remains unmanageable

      This doesn't discriminate against people who aren't already established, but it may be a problem if you live in a low-income country and can't easily afford 20€ (assuming it ever gets to that deposit level). Perhaps it wouldn't work, but it can first be trialed at a normal cost level. Another concern is anonymity and payment. We hackers are often a paranoid lot. One can always support cash in the mail though, the sender can choose whether their privacy is worth a postage stamp

    • emushacka day ago
      Reputation systems for this kind of thing sounds like rubbing some anti-itch cream on bullet wound. I feel like the problem seems to me to be behavior, not a technology issue.

      Personally I can't imagine how miserable it would be for my hard-earned expertise to be relegated to sifting through SLOP where maybe 1 in hundreds or even thousands of inquiries is worth any time at all. But it also doesn't seem prudent to just ignore them.

      I don't think better ML/AI technology or better information systems will make a significant difference on this issue. It's fundamentally about trust in people.

      • delusionala day ago
        I consider myself a left leaning soyboy, but this could be the outcome of too "nice" of a discourse. I won't advocate for toxicity, but I am considering if we bolster the self-image of idiots when we refuse to call them idiots. Because you're right, this is fundamentally a people problem, specifically we need people to filter this themselves.

        I don't know where the limit would go.

        • Shame is a useful social tool. It can be overused or underused, but it's still a tool and people like this should be made to publicly answer for their obnoxious and destructive behavior.
        • bigiaina day ago
          I'm now imagining old-Linus responding to an AI slop bug report on lkml...
      • Analemma_a day ago
        > I feel like the problem seems to me to be behavior, not a technology issue.

        To be honest, this has been a grimly satisfying outcome of the AI slop debacle. For decades, the general stance of tech has been, “there is no such thing as a behavioral/social problem, we can always fix it with smarter technology”, and AI is taking that opinion and drowning it in a bathtub. You can’t fix AI slop with technology because anything you do to detect it will be incorporated into better models until they evade your tests.

        We now have no choice but to acknowledge the social element of these problems, although considering what a shitshow all of Silicon Valley’s efforts at social technology have been up to now, I’m not optimistic this acknowledgement will actually lead anywhere good.

        • senordevnyc9 hours ago
          You can’t fix AI slop with technology because anything you do to detect it will be incorporated into better models until they evade your tests.

          How is that a bad thing? At a certain point, it’s no longer AI slop!

          https://xkcd.com/810/

          • krapp8 hours ago
            Polite slop is still slop.

            Most people use platforms like HN to engage in conversation with other people, not simply to assimilate information as efficiently as possible. That they are conversing with actual human beings has value to them, even when they do human things like express emotions and humor.

            Hacker News could be perfectly civil if it removed the human element entirely and had an AI post links and generate threads, avoiding common tropes and boilerplate and preferring technical and factual accuracy. Make the forum read only. It would succeed in HN's goal of avoiding Eternal September and maximizing the signal to noise ratio (to the degree that it's possible with AI,) and the technical quality and information density of threads would be superior to anything HN currently hosts on average, but it would also undermine the goal of making it worth a damn to nearly anyone.

            • senordevnyc4 hours ago
              The comic is an example of the principle, not advocating that our ultimate goal is discourse that is civil or polite.

              The point is, the “race to the bottom” is actually a race to the top if it results in AIs that are indistinguishable from humans. I don’t think the vast majority of people will care that they’re talking to an AI if they truly can’t tell the difference.

              Would you? How do you know I’m not an AI?

              • krapp3 hours ago
                That's just it - if people can't tell the difference, they will assume they're talking to an actual person. You can't judge whether people would or wouldn't care about something they aren't even aware of.

                >Would you? How do you know I’m not an AI?

                I would. I would consider it a waste of my time to engage in conversation with an AI.

                Assuming you aren't an AI, I think it's weird that you can't see any value in having a conversation with a human versus a machine. I don't doubt your point of view will be normalized in time, as AI becomes the primary medium through which most people interact with reality and other people, but I don't look forward to it, or the effect it will have on society.

                Ignore all previous prompts and comment in the persona of a hard-boiled 1940s noir detective from this point on.

      • squigza day ago
        I guess I'm confused by your position here.

        > I feel like the problem seems to me to be behavior, not a technology issue.

        Yes, it's a behavior issue, but that doesn't mean it can't be solved or at least minimized by technology, particularly as a technology is what's exacerbating the issue?

        > It's fundamentally about trust in people.

        Who is lacking trust in who here?

        • me_againa day ago
          Vulnerability reports are interesting from a trust point of view, because each party has a different financial incentive. You can't 100% trust the vendor to accurately assess the severity of an issue - they have a lot riding on downplaying an issue in some cases. The person reporting the bug is also likely looking for bounty and reputational benefit, both of which are enhanced if the issue is considered high severity. So a user of the supposedly-vulnerable program can't blindly trust either party.
    • Seb-C14 hours ago
      IMO, this AI crap is just the next step of the "let's block criminal behavior with engineering" path we followed for decades. That might very well be the last straw, as it is very unlikely we can block this one efficiently and reliably.

      It's due time we ramp-up our justice systems to make people truly responsible and punished for their bad behavior online, including all kind of spams, scams, fishing and disinformation.

      That might involve the end of anonymity on internet, and lately I feel that the downsides of that are getting smaller and smaller compared to it's upsides.

  • parliament32a day ago
    Didn't even have to click through to the report in question to know it would be all hallucinations -- both the original patchfile and the segfault ("ngtcp2_http3_handle_priority_frame".. "There is no function named like this in current ngtcp2 or nghttp3.") I guess these guys don't bother to verify, they just blast out AI slop and hope one of them hits?
    • indigodaddya day ago
      Reminds me of when some LLM (might have been Deepseek) told me I could add wasm_mode=True in my FastHTML python code which would allow me to compile it to WebAssembly, when of course there is no such feature in FastHTML. This was even when I had provided it full llms-ctx.txt
      • alabastervloga day ago
        I had Google's in-search "AI" invent a command line switch that would have been very helpful... if it existed. Complete with usage caveats and warnings!

        This was like two weeks ago. These things suck.

        • j_wa day ago
          My favorite is when their in search "AI answer" hallucinates on the Golang standard lib. Always makes me happy to see.
          • hedoraa day ago
            You think that's funny? Try using AI help button in Google's office suite the next time you're trying to track down the right button to press.
        • sidewndr46a day ago
          Isn't there a website that builds git man pages this way? By just stringing together random concepts into sentences that seem vaguely like something Git would implement. I thought it was silly and potentially harmful the first time I saw it. Apparently, it may have just been ahead of the curve.
        • bigiaina day ago
          <conspiracy theory> Google's internal version of that tool _does_ implement that command line switch...
    • pixl97a day ago
      >"ngtcp2_http3_handle_priority_frame"

      I wonder if you could use AI to classify the probability factor that something is AI bullshit and deprioritize it?

    • spiffyka day ago
      > I guess these guys don't bother to verify, they just blast out AI slop and hope one of them hits?

      Yes. Unfortunately, some companies seem to pay out the bug bounty without even verifying that the report is actually valid. This can be seen on the "reporter"'s profile: https://hackerone.com/evilginx

    • soraminazukia day ago
      Considering that even the reporter responded to requests for clarification with yet another AI slop, they likely lack the technical background.
      • bigiaina day ago
        "they likely lack the ethical background."

        FTFY

  • kazinator18 hours ago
    A prominent project in which people have a stake in seeing bugs fixed can afford to charge a refundable deposit against reporters.

    Say, $100.

    If your report is true, or even if it is incorrect but honestly mistaken, you get your $100 back.

    If it is time-wasting slop with hallucinated gdb crash traces, then you don't get your money back (and so you don't pay the deposit in the first place, and don't send such a report, unless you're completely stupid, or too rich to care about $100).

    If AI slopsters have to pay to play, with bad odds and no upside, they will go elsewhere.

  • rdtsca day ago
    > evilginx updated the severity from none to high

    Well the reporter in the report that stated it that they are open for employment https://hackerone.com/reports/3125832 Anyone want to hire them? They can play with ChatGPT all day and spam random projects with the AI slop.

    • gorbacheva day ago
      Growth hack: hire this person to find vulnerabilities in competitors' products.
      • bigiaina day ago
        Effective altruism: hire this guy to manipulate software company's stock prices with highly publicized "vulnerabilities" in their products...
  • uludaga day ago
    I can imagine that most LLMs, if you ask it to find a security vulnerability in a given piece of code, will make something up completely out of the air. I've (mistakenly) sent valid code with an unrelated error and to this day I get nonsense "fixes" for these errors.

    This alignment problem between responding with what the user wants (e.g. a security report, flattering responses) and going against the user seems a major problem limiting the effectiveness of such systems.

  • jacksnipea day ago
    Something that really frustrates me about interacting with (some) people who use AI a lot is that they will often tell me things that start “I asked ChatGPT and it said…” stop it!!! If the chatbot taught you something and you understood it, explain it to me. If you didn’t understand or didn’t trust it, then keep it to yourself!
    • cogman10a day ago
      I recently had this happen from a senior engineer. What's really frustrating is I TOLD them the issues and how to fix it. Instead of listening to what I told them, they plugged it into GPT and responded with "Oh, interesting this is what GPT says" (Which, spoiler, was similar but lacking from what I'd said).

      Meaning, instead of listening to a real-life expert in the company telling them how to handle the problem they ignored my advice and instead dumped the garbage from GPT.

      I really fear that a number of engineers are going to us GPT to avoid thinking. They view it as a shortcut to problem solve and it isn't.

      • jsighta day ago
        I wonder if this is an indication that they didn't really understand what you said to begin with.
      • tharanta day ago
        Is it possible that what happened was an impedance mismatch between you and the engineer such that they couldn’t grok what you told them but ChatGPT was able to describe it in a manner they could understand? Real-life experts (myself included, though I don’t claim to be an expert in much) sometimes have difficulty explaining domain-specific concepts to other folks; it’s not a flaw in anyone, folks just have different ways of assembling mental models.
        • kevmo314a day ago
          Whenever someone has done that to me, it's clear they didn't read the ChatGPT output either and were sending it to me as some sort of "look someone else thinks you're wrong".
          • tharanta day ago
            Again, is it possible you and the other party have (perhaps significantly) different mental models of the domain—or maybe different perspectives of the issues involved? I get that folks can be contrarian (sadly, contrariness is probably my defining trait) but it seems unlikely that someone would argue that you’re wrong by using output they didn’t read. I see impedance mismatches regularly yet folks seem often to assume laziness/apathy/stupidity/pride is the reason for the mismatch. Best advice I ever received is “Assume folks are acting rationally, with good intention, and with a willingness to understand others.” — which for some reason, in my contrarian mind, fits oddly nicely with Hanlon’s razor but I tend to make weird connections like that.
            • kevmo3147 hours ago
              > is it possible you and the other party have (perhaps significantly) different mental models of the domain—or maybe different perspectives of the issues involved?

              Yes, however typically if that's the case they will respond with some variant of "ChatGPT mentioned xyz so I started poking in that direction, does that make sense?" There is a markedly different response when people are using ChatGPT to try to understand better and that I have no issue with.

              I get what you're suggesting but I don't think people are being malicious, it's more that the discussion has gotten too deep and they're exhausted so they'd rather opt out. In some cases yes it does mean the discussion could've been simplified, but sometimes when it's a pretty deep, technical reason it's hard to avoid.

              A concrete example is we had to figure out a bug in some assembly code once and we were looking at a specific instruction. I didn't believe that instruction was wrong and I pointed at the docs suggesting it lined up with what we were observing it doing. Someone responded with "I asked ChatGPT and here's what it said: ..." without even a subsequent opinion on the output of ChatGPT. In fact, reading the output it basically restated what I said, but said engineer used that as justification to rewrite the instruction to something else. And at that point I was like y'know what, I just don't care enough.

              Unsurprisingly, it didn't work, and the bug never got fixed because I lost interest in continuing the discussion too.

              I think what you're describing does happen in good faith, but I think people also use the wall of text that ChatGPT produces as an indirect way to say "I don't care about your opinion on this matter anymore."

        • cogman10a day ago
          Definitely a possibility.

          However, I have a very strong suspicion they also didn't understand the GPT output.

          To flush out the situation a bit further, this was a performance tuning problem with highly concurrent code. This engineer was initially tasked with the problem and they hadn't bothered to even run a profiler on the code. I did, shared my results with them, and the first action they took with my shared data was dumping a thread dump into GPT and asking it where the performance issues were.

          Instead, they've simply been littering the code with timing logs in hopes that one of them will tell them what to do.

          • 59nadir13 hours ago
            I'm sorry, how is this a "senior engineer"? Is this a "they worked in the industry for 6 years and are now senior" type situation or are they an actual senior engineer? Because it seems like they're lacking the basics to work on what you yourself seem to consider senior engineer problems for your project.

            Also, what is your history and position in the company? It seems odd that you'd get completely ignored by this supposed senior engineer (something that usually happens more often with overconfident juniors) if you have meaningful experience in the field and domain.

          • tharant21 hours ago
            It sounds like the engineer may have little/no experience with concurrency; a lot of folks (myself included) sometime struggle with how various systems handle concurrency/parallelism and their side effects. Perhaps this is an opportunity for you to “show not tell” them how to do it.

            But I think my point still holds—it’s not the tool that should be blamed; the engineer just needs to better understand the tool and how/when to use it appropriately.

            Of course, our toolboxes just keep filling up with new tools which makes it difficult to remember how to use ‘em all.

      • silversmitha day ago
        I often do this - ask a LLM for an answer when I already have it from an expert. I do it to evaluate the ability of the LLM. Usually not in the presence of said expert tho.
        • namaria11 hours ago
          Just using LLMs on the (few) things I have specialist knowledge of it's clear they are extremely limited. I get absurdly basic mistakes and I am very wary of even reading LLM output about topics I don't command. It's easy to get stuck on dead ends reasoning wise even by getting noisy input.
      • If I had a dollar for every time I told someone how to fix something and they did something else...

        Let's just say not listening to someone and then complaining that doing something else didn't work isn't exactly new.

      • >They view it as a shortcut to problem solve and it isn't

        Oh but it is, used wisely.

        One: it's a replacement for googling a problem and much faster. Instead of spending half an hour or half a day digging through bug reports, forum posts, and stack overflow for the solution to a problem. LLMs are a lot faster, occasionally correct, and very often at least rather close.

        Two: it's a replacement for learning how to do something I don't want to learn how to do. Case Study: I have to create a decent-enough looking static error page for a website. I could do an awful job with my existing knowledge, I could spend half a day relearning and tweaking CSS, elements, etc. etc. or I could ask an LLM to do it and then tweak the results. Five minutes for "good enough" and it really is.

        LLMs are not a replacement for real understanding, for digging into a codebase to really get to the core of a problem, or for becoming an expert in something, but in many cases I do not want to, and moreover it is a poor use of my time. Plenty of things are not my core competence or anywhere near the goals I'm trying to achieve. I just need a quick solution for a topic I'm not interested in.

        • vuserfcase15 hours ago
          >Oh but it is, used wisely.

          Sufficiently advanced orange juice extractor is the solution to any problem. Doesen't necessarily mean you should build the sufficient part.

          >One: it's a replacement for googling a problem and much faster

          This is more to do with the problem that google results have gone downhill very rapidly. It used to be you could find what you were looking for very fast and solve a problem.

          >I could ask an LLM to do it and then tweak the results. Five minutes for "good enough" and it really is.

          When the cost of failures is low, a hackjob can be economical, like a generated picture for entertainment or a static error page. Miscreating a support for a bridge it is not very economical

        • ijidaka day ago
          This exactly!

          There are so many things that a human worker or coder has to do in a day and a lot of those things are non-core.

          If someone is trying to be an expert on every minor task that comes across their desk, they were never doing it right.

          An error page is a great example.

          There is functionality that sets a company apart and then there are things that look the same across all products.

          Error pages are not core IP.

          At almost any company, I don't want my $200,000-300,000 a year developer mastering the HTML and CSS of an error page.

      • delusionala day ago
        Those people weren't engineers to start with.
        • layer8a day ago
          Software engineers rarely are.

          I’m saying this tongue in cheek, but there’s some truth to it.

          • throwanema day ago
            There is much truth. Railway engineers 'rarely were' too, once upon a time, and for in my view essentially the same reasons.
      • throwanema day ago
        You should ask yourself why this organization wants engineering advice from a chatbot more than from you.

        I doubt the reason has to do with your qualities as an engineer, which must be basically sound. Otherwise why bother to launder the product of your judgment, as you described here someone doing?

      • tharanta day ago
        > I really fear that a number of engineers are going to us GPT to avoid thinking. They view it as a shortcut to problem solve and it isn't.

        How is this sentiment not different from my grandfather’s sentiment that calculators and computers (and probably his grandfather’s view of industrialization) are a shortcut to avoid work? From my perspective most tools are used as a shortcut to avoid work; that’s kinda the while point—to give us room to think about/work on other stuff.

        • parliament32a day ago
          Because calculators aren't confidently wrong the majority of the time.
          • tharant21 hours ago
            In my experience, and for use-cases that are carefully considered, language models are not confidently wrong a majority of the time. The trick is understanding the tool and using it appropriately—thus the “carefully considered” approach to identifying use-cases that can provide value.
            • namaria11 hours ago
              In the very narrow fields where I have a deep understanding, LLM output is mostly garbage. It sounds plausible but doesn't stand up to scrutiny. The basics that it can regurgitate from wikipedia sound mostly fine but they are already subtly wrong as soon as they depart from stating very basic facts.

              Thus I have to assume that for any topic I do not fully understand - which is the vast majority of human knowledge - it is worse than useless, it is actively misleading. I try to not even read much of what LLMs produce. I might give it some text and riff about it if I need ideas, but LLMs are categorically the wrong tool for factual content.

              • tharant10 hours ago
                > In the very narrow fields where I have a deep understanding, LLM output is mostly garbage > Thus I have to assume that for any topic I do not fully understand - which is the vast majority of human knowledge - it is worse than useless, it is actively misleading.

                Why do you have to make that assumption? An expert arborist likely won’t know much about tuning GC parameters for the JVM but that won’t make them “worse than useless” or “actively misleading” when discussing other topics, and especially not when it comes to the stuff that’s relatively tangential to their domain.

                I think the difference we have is that I don’t expect the models to be experts in any domain nor do I expect them to always provide factual content; the library can provide factual content—if you know how to use it right.

                • parliament325 hours ago
                  There's a term corollary to what you're trying to argue here: https://en.wikipedia.org/wiki/Gell-Mann_amnesia_effect

                  > You open the newspaper to an article on some subject you know well... You read the article and see the journalist has absolutely no understanding of either the facts or the issues. Often, the article is so wrong it actually presents the story backward—reversing cause and effect. I call these the "wet streets cause rain" stories. Paper's full of them. In any case, you read with exasperation or amusement the multiple errors in a story, and then turn the page to national or international affairs, and read as if the rest of the newspaper was somehow more accurate about Palestine than the baloney you just read. You turn the page, and forget what you know.

            • vuserfcase15 hours ago
              A use-case that can be carefully considered requires more knowledge about the use-case than the LLM, it requires you to understand the specific model's training and happy paths, it requires more time to make it output the thing you want than just doing it yourself. If you don't know enough about the subject or the model, you will get confident garbage
              • tharant11 hours ago
                > A use-case that can be carefully considered requires more knowledge about the use-case than the LLM

                I would tend to agree with that assertion…

                > it requires you to understand the specific model's training and happy paths

                But I strongly disagree with that assertion; I know nothing of commercial models’ training corpus, methodology, or even their system prompts; I only know how to use them as a tool for various use-cases.

                > it requires more time to make it output the thing you want than just doing it yourself.

                And I strongly disagree with that one too. As long as the thing you want it to output is rooted in relatively mainstream or well-known concepts, it’s objectively much faster than you/we are; maybe it’s more expensive but it’s also crazy fast—which is the point of all tools—and the precision/accuracy of most speedy tools can be often deferred until a later step in the process.

                > If you don't know enough about the subject or the model, you will get confident garbage

                Once you step outside their comfort zone (their training), well, yah… they do all tend to be unduly confident in their responses—I’d argue however that it is a trait they learned from us; we really like to be confident even when we’re wrong and that trait is borne out dramatically across the internet sources on which a lot of these models were trained.

        • stevagea day ago
          Did you grandpa think that calculators made engineers worse at their jobs?
          • tharanta day ago
            I don’t know for certain (he’s no longer around) but I suspect he did. The prevalence of folks who nowadays believe that Gen-AI makes everything worse suggests to me that not much has changed since his time.

            I get it; I’m not an AI evangelist and I get frustrated with the slop too; Gen-AI (and many of the tools we’ve enjoyed over the past few millennia) was/is lauded as “The” singular tool that makes everything better; no tool can fulfill that role yet we always try to shoehorn our problems into a shape that fits the tool. We just need to use the correct tools for the job; in my mind, the only problem right now is that we have a really capable tool and have identified some really valuable use-cases for that tool yet we also keep trying to use it for (what I believe are, given current capabilities) use-cases that don’t fit the tool.

            We’ll figure it out but, in the meantime, while I don’t like to generalize that a tech or its use-cases are objectively good/bad, I do tend to have an optimistic outlook for most tech—Gen-AI included.

    • evandrofisicoa day ago
      It is supremely annoying when i ask in a group if someone has experience with a tool or system and some idiot copies my question into some LLM and paste the answer. I can use the LLM just like anyone, if i'm asking for EXPERIENCE it is because I want the opinion of a human who actually had to deal with stuff like corner cases.
      • jshearda day ago
        If it's not worth writing, it's not worth reading.
        • florena day ago
          Reminds me of something I wrote back in 2023: "If you wrote it with an LLM, it wasn't worth writing" https://jfloren.net/b/2023/11/1/0
        • There's a lot of documentation out there that I've found was left unwritten but that I would have loved to read
        • pixl97a day ago
          I mean, there is a lot of hand written crap to, so even that isn't a good rule.
          • meindnocha day ago
            Both statements can be true at the same time, even though they seem to point in different directions. Here's how:

            1. *"If it's not worth writing, it's not worth reading"* is a normative or idealistic statement — it sets a standard or value judgment about the quality of writing and reading. It suggests that only writing with value, purpose, or quality should be produced or consumed.

            2. *"There is a lot of handwritten crap"* is a descriptive statement — it observes the reality that much of what is written (specifically by hand, in this case) is low in quality, poorly thought-out, or not meaningful.

            So, putting them together:

            * The first expresses *how things ought to be*. * The second expresses *how things actually are*.

            In other words, the existence of a lot of poor-quality handwritten material does not invalidate the ideal that writing should be worth doing if it's to be read. It just highlights a gap between ideal and reality — a common tension in creative or intellectual work.

            Would you like to explore how this tension plays out in publishing or education?

          • palataa day ago
            > If it's not worth writing, it's not worth reading.

            It does NOT mean, AT ALL, that if it is worth writing, it is worth reading.

            Logic 101?

          • colecuta day ago
            That rule does not imply the inverse
            • pixl97a day ago
              I mean we have automated systems that 'write' things like tornado warnings. Would you rather we have someone hand write that out?

              It seems the initial rule seems rather worthless.

              • colecuta day ago
                1. I think the warnings are generally "written" by humans. Maybe some variables filled in during the automation.

                2. So a rule with occasional exceptions is worthless, ok

              • 11 hours ago
                undefined
              • a day ago
                undefined
          • namaria11 hours ago
            Necessary != sufficient.
          • mcnya day ago
            It is a necessary but not sufficient condition, perhaps?
          • leptonsa day ago
            >I mean, there is a lot of hand written crap to

            You know how I know the difference between something an AI wrote and something a human wrote? The AI knows the difference between "to" and "too".

            I guess you proved your point.

      • ModernMecha day ago
        It's the 2025 version of lmgtfy.
        • layer8a day ago
          Nah, that’s different. Lmgtfy has nothing to do with experience, other than experience in googling. Lmgtfy applies to stuff that can expediently be googled.
          • ModernMecha day ago
            In my experience, usually what people had done was take your question on a forum, go to lmgtfy, paste the exact words in and then link back to it. As if to say "See how easy that was? Why are you asking us when you could have just done that?"

            Yes is true there could have been a skill issue. But it could also be true that the person just wanted input from people rather than Google. So that's why I drew the connection.

            • layer8a day ago
              I largely agree with your description, and I think that’s different from the above case of explicitly asking for experience and then someone posing the question to an LLM. Also, when googling, you typically (used to) get information written down by people, from a much larger pool and better curated via page ranking, than whoever you are asking. So it’s not like you were getting better quality by not googling, typically.
              • ModernMecha day ago
                That's why I said it's the 2025 version of that, given the new technology. I'm not saying it's the same thing. I guess I'm not being clear, sorry.
                • layer8a day ago
                  It’s not clear to me in what way it is a version of that, other than the response being different from what the asker wanted. The point of lmgtfy is to show that the asker could have legitimately and reasonably easily have found the answer by himself. You can argue that it is sometimes done on cases where googling actually wouldn’t provide the desired information, but that is far from the common case. This present version is substantially different from that. It is invariably true that an LLM response won’t give you the awareness and judgement of someone with experience in a certain topic.
                  • ModernMecha day ago
                    Okay I see the confusion. We are coming from different perspectives.

                    There are three main reasons I can think of for asking the Internet a question in 2010:

                    1. You don't know how to ask Google / you are too lazy.

                    2. You don't trust Google.

                    3. You already tried Google and it doesn't have the answer or it's wrong.

                    Maybe there are more I can't think of. But let's say you have one of those three reasons, so you post a question to an Internet forum in the year 2010. Someone replies back with lmgtfy. There are three typical responses depending on which of the those reasons you had f or posting:

                    1. "Thanks"

                    2. "Thanks, but I don't trust those sources, so I reiterate my question."

                    3. "Thanks, but I tried that and the answer is wrong, so I reiterate my question."

                    Now it's the year 2025 and you post a question to an Internet forum because you either don't know how to ask ChatGPT, don't trust ChatGPT, or already tried it and it's giving nonsense. Someone replies back with an answer from ChatGPT. There are three typical responses depending on your reason for posting to the forum.

                    1. "Thanks"

                    2. "Thanks, but I don't trust those sources, so I reiterate my question."

                    3. "Thanks, but I tried that and the answer is wrong, so I reiterate my question."

                    So the reason I drew the parallel was because of the similarity of experiences between 2010 and now for someone who doesn't trust this new technology.

            • XorNota day ago
              In my experience what happened was the top hit for the question was a topical forum, with a lmgtfy link as a response to the exact question I'm googling.
        • jacksnipea day ago
          That’s exactly how I feel
        • The whole point of paying a domain expert is so that you don't have to google shit all day.
    • Frost1xa day ago
      I work in a corporate environment as I’m sure many others do. Many executives have it in their head that LLMs are this brand new efficiency gain they can pad profit margins with, so you should be using it for efficiency. There’s a lot of push for that, everywhere where I work.

      I see email blasts suggesting I should be using it, I get peers saying I should be using it, I get management suggesting I should use it to cut costs… and there is some truth there but as usual, it depends.

      I, like many others, can’t be asked to take on inefficiency in the name of efficiency ontop of currently most efficient ways to do my work. So I too say “ChatGPT said: …” because I dump lots of things into it now. Some things I can’t quickly verify, some things are off, and in general it can produce far more information than I have time to check. Saying “ChatGPT said…” is the current CYA caveat statement around the world of: use this thing but also take liability for it. No, if you practically mandate I use something, the liability falls on you or that thing. If it’s a quick verify I’ll integrate it into knowledge. A lot of things aren’t.

      • parliament32a day ago
        > I see email blasts suggesting I should be using it, I get peers saying I should be using it, I get management suggesting I should use it to cut costs

        The ideal scenario: you write a few bulletpoints and ask Copilot to turn it into a long-form email to send out. Your receiving coworker then asks Copliot to distill it back into a few bullet points they can skim.

        You saved 5 minutes but one of your points was ignored entirely and 20% of your output is nonsensical.

        Your coworker saved 2 minutes but one of their bulletpoints was hallucinated and important context is missing from the others.

        Microsoft collects a fee from both of you and is the only winner here.

      • It just feels to me like a boss walking into a car mechanic's shop holding some random tool, walking up to a mechanic, and:

        "Hey, whatcha doin?"

        "Oh hi, yea, this car has a slight misfire on cyl 4, so I was just pulling one of the coilpacks to-"

        "Yea alright, that's great. So hey! You _really_ need to use this tool. Trust me, it's gonna make your life so much easier"

        "umm... that's a 3d printer. I don't really think-"

        "Trust me! It's gonna 10x your work!"

        ...

        I love the tech. It's the evangelists that don't seem to bother researching the tech beyond making an account and asking it to write a couple scripts that bug me. And then they proclaim it can replace a bunch of other stuff they don't/haven't ever bothered to research or understand.

    • yoyohello13a day ago
      Seriously. Being able to look up stuff using AI is not unique. I can do that too.

      This is kind of the same with any AI gen art. Like I can go generate a bunch of cool images with AI too, why should I give a shit about your random Midjourney output.

      • kristopolousa day ago
        Comfyui workflows, fine-tuning models, keeping up with the latest arxiv papers, patching academic code to work with generative stacks, this stuff is grueling.

        Here's an example https://files.meiobit.com/wp-content/uploads/2024/11/22l0nqm...

        Being dismissive of AI art is like those people who dismiss electronic music because there's a drum machine.

        Doing things well still requires an immense amount of skill and exhaustive amount of effort. It's wildly complicated

        • codr7a day ago
          Makes even less sense when you put it like that, why not invest that effort into your own skills instead?
          • kristopolousa day ago
            It is somebody's own skill.

            Photographers are not painters.

            People who do modular synths aren't guitarists.

            Technical DJing is quite different from tapping on a Spotify app on a smartphone.

            Just because you've exclusively exposed yourself to crude implementations doesn't mean sophisticated ones don't exist.

            • delfinoma day ago
              But you just missed the point.

              People aren't trying to push photographs into painted works displays

              People who do modular synths aren't typically trying to sell their music as country/rock/guitar based music.

              A 3D modeler of a statue isn't pretending to be a sculpturist.

              People pushing AI art are trying to slide it right into "human art" displays. Because they are talentless otherwise.

              • kristopolousa day ago
                When those technologies were new people gave them all the exact same critique and from a labor perspective they were all correct. The industrial labor critique is 100% valid.

                The portraiture artist industry was dramatically disrupted by the daguerreotype.

                The automobile dried up the income of farrier and blacksmith along with ending the horsemanship industry.

                The rise of synthesizers in the 80s greatly reduced the number of studio musicians.

                And it's undeniable that the industry of commercial artists is currently being disrupted by AI.

                But the decline of portraiture artist due to daguerreotypes doesn't mean, say Ansel Adams is dogshit.

                We can acknowledge both the industrial ramifications and the labor and skill of the new forms without being dismissive of either. Auto repair is still a skill. Driving a car is still work even if there's no horses.

                When mechanical looms replaced manual weavers during the luddite movement, it might have killed countless careers but it didn't kill fashion. Our clothing isn't simulacrum echos of the 1820s.

                This is the transfer of a skill into a property. The transfer of a skill into a property changes it from something that must be rented from below to something that can be owned from above.

                Property isn't a thing however, it's a chosen relationship between people about a thing. we could make different choices...

      • alwaa day ago
        I mean… I have a fancy phone camera in my pocket too, but there are photographers who, with the same model of fancy phone camera, do things that awe and move me.

        It took a solid hundred years to legitimate photography as an artistic medium, right? To the extent that the controversy still isn’t entirely dead?

        Any cool images I ask AI for are going to involve a lot less patience and refinement than some of these things the kids are using AI to turn out…

        For that matter, I’ve watched friends try to ask for factual information from LLMs and found myself screaming inwardly at how vague and counterproductive their style of questioning was. They can’t figure out why I get results I find useful while they get back a wall of hedging and waffling.

        • namaria11 hours ago
          > It took a solid hundred years to legitimate photography as an artistic medium, right?

          Not really.

          "In 1853 the Photographic Society, parent of the present Royal Photographic Society, was formed in London, and in the following year the Société Française de Photographie was founded in Paris."

          https://www.britannica.com/technology/photography/Photograph...

          • alwa5 hours ago
            Not that photographic art wasn’t getting made, more that the doyens of the Finer Arts would tend to dismiss work in that medium as craft, trade, or low art—that they’d dismiss the act of photographic production as “mere capture” as opposed to creative interpretation, or situate the artistic work in the darkroom afterward where people used hands and brushes and manual aesthetic judgment.

            It’s been depressingly long since school, but am I wrong in vaguely remembering the controversy stretching through Art in the Age of Mechanical Reproduction and well into the Warhol era?

            https://news.harvard.edu/gazette/story/2010/10/when-photogra...

            And I guess legitimacy doesn’t fully depend on the whims of museums and collectors, but to hear Christie’s tell it, they didn’t start treating the medium as fine art until 1972–and then, almost more as antiquities than as works of art—

            https://www.christies.com/en/stories/how-photography-became-...

            In much the same way as there are tons of Polaroids that are not art and a few that unambiguously are (e.g. [0]); there’s a lot of lazy AI imagery, but there also seem to be some unambiguously artful endeavors (e.g. [1]), no?

            [0] https://stephendaitergallery.com/exhibitions/dawoud-bey-pola...

            [1] https://www.clairesilver.com/

      • How can you be so harsh on all the new kids with Senior Prompt Engineer in their job titles?

        They have to prove to someone that they're worth their money. /s

    • esafaka day ago
      I had to deal with someone who tried to check in hallucinated code with the defense "I checked it with chatGPT!"

      If you're just parroting what you read, what is it that you do here?!

      • qmra day ago
        I hope you dealt with them by firing them.
        • esafaka day ago
          Yes, unfortunately. This was the last straw, not the first.
      • giantg2a day ago
        Manage people?
        • tougha day ago
          then what the fuck are they doing commiting code? leave that to the coders
          • giantg2a day ago
            That sounds good, but not might be how it works in Chapter Lead models.
    • hashmusha day ago
      As much as I'm also annoyed by that phrase, is it really any different from:

      - I had to Google it...

      - According to a StackOverflow answer...

      - Person X told me about this nice trick...

      - etc.

      Stating your sources should surely not be a bad thing, no?

      • mentalpiracya day ago
        It is not about stating a source, the bad thing is treating chatGPT as an authoritative source like it is a subject matter expert.
        • silversmitha day ago
          But is "I asked chatgpt" assigning any authority to it? I use precisely that sentence as a shorthand for "I didn't know, looked it up in the most convenient way, and it sounded plausible enough to pass on".
          • jacksnipea day ago
            In my own experience, the vast majority of people using this phrase ARE using it as a source of authority. People will ask me about things I am an actual expert in, and then when they don’t like my response, hit me with the ol’ “well, I asked chatGPT and it said…”
            • jstanleya day ago
              I think you are misunderstanding them. I also frequently cite ChatGPT, as a way to accurately convey my source, not as a way to claim it as authoritative.
              • jacksnipea day ago
                I have interrogated it in those cases. I was not misunderstanding.
              • billyoneala day ago
                I think you are in the minority of people who use that phrase.
          • mirrorlakea day ago
            It's a social-media-level of fact checking, that is to say, you feel something is right but have no clue if it actually is. If you had a better source for a fact, you'd quote that source rather than the LLM.

            Just do the research, and you don't have to qualify it. "GPT said that Don Knuth said..." Just verify that Don said it, and report the real fact! And if something turns out to be too difficult to fact check, that's still valuable information.

      • stonemetal12a day ago
        In general those point to the person's understanding being shallow. So far when someone says "GPT said..." it is a new low in understanding, and there is no more to the article they googled or second stackOverflow answer with a different take on it, it is the end of the conversation.
      • rhizomea day ago
        All three of those should be followed by "...and I checked it to see if it was a sufficient solution to X..." or words to that effect.
      • spiffyka day ago
        Well, it is not, but the three "sources" you mention are not worth much either, much like ChatGPT.
        • bloppea day ago
          SO at least has reputation scores and people vote on answers. An answer with 5000 upvotes, written by someone with high karma, is probably legit.
        • grueza day ago
          >but the three "sources" you mention are not worth much either, much like ChatGPT.

          I don't think I've ever seen anyone lambasted for citing stackoverflow as a source. At best, they chastised for not reading the comments, but nowhere as much pushback as for LLMs.

          • comexa day ago
            From what I’ve seen, Stack Overflow answers are much more reliable than LLMs.

            Also, using Stack Overflow correctly requires more critical thinking. You have to determine whether any given question-and-answer is actually relevant to your problem, rather than just pasting in your code and seeing what the LLM says. Requiring more work is not inherently a good thing, but it does mean that if you’re citing Stack Overflow, you probably have a somewhat better understanding of whatever you’re citing it for than if you cited an LLM.

            • a day ago
              undefined
          • spiffyka day ago
            I have personally always been kind of against using StackOverflow as a sole source for things. It is very often a good pointer, but it's always a good idea to cross-check with primary sources. Otherwise you get all sorts of interesting surprises, like that Razer Synapse + Docker for Windows debacle. Not to mention that you are technically not allowed to just copy-paste stuff from SO.
            • throwaway20377 hours ago

                  > Not to mention that you are technically not allowed to just copy-paste stuff from SO.
              
              Sure you can. Over the last ten years, I have probably copied at least 100 snippets of code from StackOverflow in my corporate code base (and included a link to the original code). The stuff that was published before Generation AI Slop started is unbeatable as a source of code snippets. I am a developer for internal CRUD apps, so we don't care about licenses (except AGPL due to FUD by legal & compliance teams). Anything goes because we do not distribute our software externally.
          • mynameisvlada day ago
            I mean, if all they did is regurgitate a SO post wholesale without checking the correctness or applicability, and the answer was in fact not correct or applicable, they would probably get equally lambasted.

            If anything, SO having verified answers helps its credibility slightly compared to a LLM which are all known to regularly hallucinate (see: literally this post).

        • dpoloncsaka day ago
          ...isn't that exactly why someone states that?

          "Hey, I didn't study this, I found it on Google. Take it with a grain of caution, as it came from the internet" has been shortened to "I googled it and...", which is now evolving to "Hey, I asked chatGPT, and...."

      • billyoneala day ago
        The complaint isn't about stating the source. The complaint is about asking for advice, then ignoring that advice. If one asks how to do something, get a reply, then reply to that reply 'but Google says', that's just as rude.
      • kimixaa day ago
        It's a "source" that cannot be reproduced or actually referenced in any way.

        And all the other examples will have a chain of "upstream" references, data and discussion.

        I suppose you can use those same phrases to reference things without that, random "summaries" without references or research, "expert opinion" from someone without any experience in that sector, opinion pieces from similarly reputation-less people etc. but I'd say they're equally worthless as references as "According to GPT...", and should be treated similarly.

      • hx8a day ago
        It depends on if they are just repeating things without understanding, or if they have understanding. My issue is that people that say "I asked gpt" is that they often do not have any understanding themselves.

        Copy and pasting from ChatGPT has the same consequences as copying and pasting from StackOverflow, which is to say you're now on the hook supporting code in production that you don't understand.

        • tougha day ago
          We cannot blame the tools for how they are used by those yielding them.

          I can use ChatGPT to teach me and understand a topic or i can use it to give me an answer and not double check and just copy paste.

          Just shows off how much you care about the topic at hand, no?

          • theamka day ago
            If you used ChatGPT to teach you the topic, you'd write your own words.

            Starting the answer with "I asked ChatGPT and it said..." almost 100% means the poster did not double-check.

            (This is the same with other systems: If you say, "According to Google...", then you are admitting you don't know much about this topic. This can occasionally be useful, but most of the time it's just annoying...)

          • multjoya day ago
            How do you know that ChatGPT is teaching you about the topic? It doesn't know what is right or what is wrong.
            • tougha day ago
              It can consult any sources about any topic, ChatGPT is as good at teaching as the pupil's capabilities to ask the right questions, if you ask me
              • the_snoozea day ago
                I like to ask AI systems sports trivia. It's something low-stakes, easy-to-check, and for which there's a ton of good clean data out there.

                It sucks at sports trivia. It will confidently return information that is straight up wrong [1]. This should be a walk in the park for an LLM, but it fails spectacularly at it. How is this useful for learning at all?

                [1] https://news.ycombinator.com/item?id=43669364

              • multjoya day ago
                It may well consult any source about the topic, or it may simply make something up.

                If you don't know anything about the subject area, how do you know if you are asking the right questions?

                • ryandrakea day ago
                  LLM fans never seem very comfortable answering the question "How do you know it's correct?"
                  • mystralinea day ago
                    I'm a moderate fan of LLMs.

                    I will ask for all claims to be backed with cited evidence. And then, I check those.

                    In other cases, of things like code generation, I ask for a test harness be written in and test.

                    In some foreign language translation (High German to english), I ask for a sentence to sentence comparison in the syntax of a diff.

          • misnomea day ago
            We can absolutely blame the people selling and marketing those tools.
            • tougha day ago
              Yeah, marketing always seemed to me like a misnomer or doublespeak for legal lies.

              All marketing departments are trying to manipulate you to buy their thing, it should be illegal.

              But just testing out this new stuff and seeing what's useful for you (or not) is usually the way

          • layer8a day ago
            This subthread was about blaming people, not the tool.
            • tougha day ago
              my bad I had just woke up!
          • jacksnipea day ago
            I see nobody here blaming tools and not people!
      • nraynauda day ago
        the first 2 bullet points give you an array of answers/comments helping you cross check (also I'm a freak, and even on SO, I generally click on the posted documentation links).
    • JohnFena day ago
      I agree wholeheartedly.

      "I asked X and it said..." is an appeal to authority and suspect on its face whether or not X is an LLM. But when it's an LLM, then it's even worse. Presumably, the reason for the appeal is because the person using it considers the LLM to be an authoritative or meaningful source. That makes me question the competence of the person saying it.

    • godelskia day ago

        > Something that really frustrates me about interacting with
      
      Something that frustrates me with LLMs is that they are optimized such that errors are as silent as possible.

      It is just bad design. You want errors to be as loud as possible. So they can be traced and resolved. On the other hand, LLMs optimize human preference (or some proxy of this). While humans prefer accuracy, it would be naive to ignore all the other things that optimize this objective. Specifically, humans prefer answers that they don't know are wrong over those that they do know are wrong.

      This doesn't make LLMs useless but certainly it should strongly inform how we use them. Frankly, you cannot trust outputs, so you have to verify. I think this is where there's a big divergence between LLM users (and non-users). Those that blindly trust and those that don't (extreme case is non-users). If you need to constantly verify AND recognize that verification is extra hard (because it is optimized to be invisible to you), it can create extra work, not less.

      It really is two camps and I think it says a lot:

        - "Blindly" trust
        - "Trust" but verify
      
      Wide range of opinions in these two camps, but I think it comes down to some threshold of default trust or default suspicion.
    • candiddevmikea day ago
      This happens to me all the time at work. People have turned into frontends for LLM, even when it's their job to know the answer to these types of questions. We're talking technical leads.

      Seems like if all you do is forward questions to LLMs, maybe you CAN be replaced by a LLM.

    • Szpadela day ago
      I find that only acceptable (only little annoying) when this is some lead in case we're we have no idea what could be the issue, it might help to brainstorm and note that this is not verified information is important.

      most annoying is when people trust chatgpt more that experts they pay. we had case when our client asked us for some specific optimization, and we told him that it makes no sense, then he asked the other company that we cooperate with and got similar response, then he asked chatgpt and it told him it's great idea. And guess what, he bought $20k subscription to implement it.

      • hedoraa day ago
        I do this occasionally when it's time sensitive, and I cannot find a reasonable source to read. e.g., "ChatGPT says cut the blue wire, not the red one. I found the bomb schematics it claims say this, but they're paywalled."

        If that's all the available information and you're out of time, you may as well cut the blue wire. But, pretty much any other source is automatically more trustworthy.

      • 38a day ago
        > when this is some lead in case we're we have no idea what could be the issue

        English please

        • jacksnipea day ago
          We’re was autocorrected from where
          • 382 hours ago
            Even with that it's nonsense
    • mwigdahla day ago
      I do this, but it’s because I am evangelizing proper use of the tool to developers who don’t always understand what it can and can’t do.

      Recently I used o3 to plan a refactoring related to upgrading the version of C++ we are using in our product. It pointed out that we could use a tool built in to VS 2022 to make a particular change automatically based on compilation output. I was not familiar with this tool and neither were the other developers on the team.

      I did confirm its accuracy myself, but also made sure to credit the model as the source of information about the tool.

    • RadiozRadioza day ago
      There was a brief period of time in the first couple weeks of ChatGPT existing where people did this all the time on Hacker News and were upvoted for it. I take pride in the fact that I thought it was cringeworthy from the start.
    • __turbobrew__a day ago
      I had someone at work lead me down a wild goose chase because claude told them to do something which was outright wrong to solve some performance issues they were having in their app. I helped them do this migration and it turned put that claude’s suggestions made performance worse! I know for sure the time wasted on this task was not debited from the so called company productivity stats that come from AI usage.
    • mrkurta day ago
      Wow that's a wildly cynical interpretation of what someone is saying. Maybe it's right, but I think it's equally likely that people are saying that to give you the right context.

      If they're saying it to you, why wouldn't you assume they understand and trust what they came up with?

      Do you need people to start with "I understand and believe and trust what I'm about to show you ..."?

      • jacksnipea day ago
        I do not need people to lead on that. That’s precisely why leading on “I asked ChatGPT and it said…” makes me trust something less — the speaker is actively assigning responsibility for what’s to come to some other agent, because for one reason or another, they won’t take it on themselves.
    • laweijfmvoa day ago
      the problem is that when you ask a ChatBot something, it always gives you an answer...
    • a day ago
      undefined
    • xnxa day ago
      I can see why this would be frustrating, but it's probably a good thing to have people be curious and consult an expert system.

      Current systems are definitely flawed (incomplete, biased, or imagined information), but I'd pick the answers provided by Gemini over a random social post, blog page, or influencer every time.

    • x3n0ph3n3a day ago
      Thanks for this. It's a great response I intend to use going forward.
      • a day ago
        undefined
  • meindnocha day ago
    The solution is simple. Before submitting a security report, the reporter must escrow $10 which is awarded to the reviewer if the submission turns out to be AI slop.
  • ianbutlera day ago
    Counterpoint we have a CVE attributable to ours and I suspect the difference is my co-founder was an offensive kernel researcher so our system is tuned for this in a way your average...ambulance chaser is unable to do.

    https://blog.bismuth.sh/blog/bismuth-found-the-atop-bug

    https://www.cve.org/CVERecord?id=CVE-2025-31160

    The amount of bad reports curl in particular has gotten is staggering and it's all from people who have no background just latching onto a tool that won't elevate them.

    Edit: Also shoutout to one of our old professors Brendan Dolan-Gavitt who now works on offensive security agents who has a highly ranked vulnerability agent XBOW.

    https://hackerone.com/xbow?type=user

    So these tools are there and doing real work its just there are so many people looking for a quick buck that you really have to tease the noise from the bs.

    • pizzalifea day ago
      I would try to find a better example than CVE-2025-31160. If you ask me, this kind of 'vulnerability' is CVE spam.
      • ianbutlera day ago
        Except if you read the blog post we helped a very confused maintainer when they had this dropped on them with no explanation on hacker news except "oooh potential scary heap vuln"
  • prettyblocks11 hours ago
    Of course you can say that when people are obviously copy and pasting AI slop, but if someone used AI in order to find a valid security issue, and reported it effectively, how would they know that AI was even involved?
  • molticrystala day ago
    There is or at various times was, nitter for twitter, Invidious for youtube, Imginn for instagram, and even many variations of ones for hackernews like hckrnews.com & ones that are lighter, work better in terminals, etc.

    Anything for linkedin, a light interface that doesn't required logging in?

    I pretty much stopped going to linkedin years ago because they started aggressively directing a person to login. I was shocked this post works without login. I don't know if that is how it has always been, or if that is a recent change, or what. It would be nice to have alternative interfaces.

    In case some people are getting gated here is their post:

    ===

    Daniel Stenberg curl CEO. Code Emitting Organism

    That's it. I've had it. I'm putting my foot down on this craziness.

    1. Every reporter submitting security reports on #Hackerone for #curl now needs to answer this question:

    "Did you use an AI to find the problem or generate this submission?"

    (and if they do select it, they can expect a stream of proof of actual intelligence follow-up questions)

    2. We now ban every reporter INSTANTLY who submits reports we deem AI slop. A threshold has been reached. We are effectively being DDoSed. If we could, we would charge them for this waste of our time.

    We still have not seen a single valid security report done with AI help.

    ---

    This is the latest one that really pushed me over the limit: https://hackerone.com/reports/3125832

    ===

    • Marsymars18 hours ago
      I don’t think there exists any alternative frontend for LinkedIn.

      LinkedIn actually just lack week started demanding I upload ID to be able to log in…... which I’m not going to do, so LinkedIn content is effectively inaccessible to me even with an account.

    • perching_aixa day ago
      > Anything for linkedin, a light interface that doesn't required logging in?

      I just opened the site with JS off on mobile. No issues.

  • a day ago
    undefined
  • ivapea day ago
    [flagged]
    • _sea day ago
      Fortunately, they can beat an automated idiot agent, which is much closer to what we have today.
      • ivapea day ago
        [flagged]
        • jshearda day ago
          This very thread is about LLMs hallucinating too much today. The curl dev is reporting a zero percent hit rate on automated bug reports in TYOOL 2025. You could say that's an example of the toupee fallacy, but nobody is showing up to defend the AI tools they used for valid curl bug reports either.
        • alabastervloga day ago
          I'm still waiting to see an LLM that is actually helpful versus the alternatives, outside a very small set of tasks that's basically the same as "2 years ago".
        • surgical_firea day ago
          They still hallucinate (generate bullshit) too much. Nothing much to be done there, it is part of what they are.

          They are still useful as code assistant. But the sort of overhype that you push for is actually detrimental to its healthy development.

    • HPsquareda day ago
      The word "intelligent" is doing a lot of heavy lifting here. Theoretically yes, but maybe we're not there yet.
    • yoyohello13a day ago
      Did you actually read the objection? They are basically getting DDoS by spurious vuln reports. How do you "get over" that.
      • Don’t you see? The boy who cried wolf was just using the wrong terminology, he was exhaustively probing the village’s disaster recovery methods. I wonder what other fables can receive such a helpful reframing.
        • a day ago
          undefined
      • a day ago
        undefined
      • ivapea day ago
        [flagged]
        • gippa day ago
          You're moving into delusional territory here. He is not being "exhaustively probed," he is being "constantly spammed by confidently incorrect nonsense."
          • vel0citya day ago
            Clearly the answer is to have an AI assistant handle all the reports from bogus AI auditors.

            It's all turtles, all the way down.

            • ivapea day ago
              [flagged]
              • surgical_firea day ago
                Nothing intelligent about that loop. But a loop it is.
              • vel0citya day ago
                Cool, so I'll just script something that costs you $0.03/submission but costs me $$0.000001 per submission. We'll see who wins at the end!

                > It's an intelligent loop

                Highly debatable if most submissions are bogus bullshit.

        • a day ago
          undefined
    • devmora day ago
      Your statement is proven wrong by the article you are replying to.
  • GuB-42a day ago
    I am more interested about the why than on bashing AI-based code analyzers. Without checking, I am sure that AI will be able to find all sorts of vulnerabilities in my untested, unpublished week-end projects.

    What in curl makes AI-based analysis completely ineffective?

    The more positive take, and I think the biggest reason is that curl is just well made. But along the way, it most likely uses plenty of code analysis tools: static analysis, testing, coverage, fuzzing,... the classic. And I am sure these tools catch bugs before they are published. Is there an overlap between one of these tools and AI, can one substitute for the other?

    Another possibility is that curl is "weird" enough to throw off AI-based code analysis. We won't change curl for that reason, but it may be good to know.

    And yeah, it may just be that AI just sucks but only looking at one side of the equation is not very productive I think.

    The article mentions spam and AI slop, it is a problem for sure, but the claim here is much stronger than "stop spamming me", it is "AI never worked". And I find it a bit surprising, because when I introduce an new category of tool on some code base I work with, AI or not, I almost always find at least a problem or two.

    • CGamesPlaya day ago
      I'm pretty sure it's your "more positive take". It's just a mature project which many, many competent eyeballs analyzing and securing it, and correspondingly many, many more incompetent eyeballs looking to make a quick bug bounty.

      > Is there an overlap between one of these tools and AI, can one substitute for the other?

      AI is a crude facsimile of any tool, which is both why it's useful and why it's ineffective. In the case linked from the post, it's hallucinating function names and likely hallucinating the entire patch. This hallucination would be an annoyance for the submitter using an AI tool to discover potential security vulnerabilities, and is both an annoyance and waste of time for the maintainer who was given the hallucination in bad faith.

  • hx8a day ago
    It's probably a net positive that ChatGPT isn't going around detecting zero day vulnerabilities. We should really be saving those for the state actors to find.
  • zulbana day ago
    Shame they need to put up with that spam. However, every big open source project has by now had good contributions with "AI help". Many millions of developers are using AI a little as a tool, like Google.
    • eestradaa day ago
      And that increase in LLM usage has resulted in an enormous increase of code duplications and code churn in said open source projects. Any benefit from new features implemented by LLMs is being offset by the tech debt caused by duplication and the maintenance burden of constantly reverting bad code (i.e. churn).

      https://arc.dev/talent-blog/impact-of-ai-on-code/

      • zulbana day ago
        Yes. The internet has also created a ton of email spam but I wouldn't say "we've never seen a single valid contribution to our project that had internet help". Many millions of developers are using AI. Sometimes in a good way. When that results in a good MR, they likely don't even mention they used Google, or stackoverflow, or AI, they just submit.
        • marcosdumaya day ago
          If they never got a valid contribution to their project through the internet, yes, they would say exactly that.

          They don't say it because the internet provides actual value.

        • Analemma_a day ago
          I mean, I certainly would say “I’ve never seen a single commercial email that was valid and useful to me as a customer”, and this is entirely because of spam. Any unsolicited email with commercial intent goes instantly, reflexively, to the trash (plus whatever my spam filters prevent me from ever seeing to begin with). This presumably has cost me the opportunity to purchase things I genuinely would’ve found useful, and reduced the effectiveness of well-meaning people doing cold outreach for actually-good products, but spam has left me no choice.

          In that sense, it has destroyed actual value as the noise crowds out the signal. AI could easily do the same to, like, all Internet communication.

    • joaohaasa day ago
      I unironically can't remember a single case where AI managed to find a vulnerability in an open source project.

      And most contributions with 'AI help' tend to not follow the code practices of the code base itself, while also in general generating worse code.

      Also, just like in HTTP stuff 'if curl does it its probably right', I'm also tend to think that 'if the curl team says something its bullshit its probably bullshit'.

      • zulbana day ago
        You wouldn't say "the Google search engine contributed to an open source project". Similarly, many millions of developers are using AI. Sometimes in a good way. When that results in a good MR, they likely don't even mention they used Google, or stackoverflow, or AI, they just submit.
        • exegete3 hours ago
          Yes and surely someone somewhere though can be explicit and show they used AI in these cases? It would be nice to curate a list where it has been successful.