223 pointsby jimminyx5 hours ago21 comments
  • popcorncowboy4 hours ago
    > running it scares the crap out of me

    A hundred times this. It's fine until it isn't. And jacking these Claws into shared conversation spaces is quite literally pushing the afterburners to max on simonw's lethal trifecta. A lot of people are going to get burned hard by this. Every blackhat is eyes-on this right now - we're literally giving a drunk robot the keys to everything.

    • anabis2 hours ago
      Maybe. People have run wildly insecure phpBB and Wordpress plugins, so maybe its the same cycle again.
      • egeozcanan hour ago
        Those usually didn't have keys to all your data. Worst case, you lost your server, and perhaps you hosted your emails there too? Very bad, but nothing compared to the access these clawdbot instances get.
        • Terretta3 minutes ago
          > Those usually didn't have keys to all your data.

          As a former (bespoke) WP hosting provider ... ROTFL. Not sure I ever met a customer's build that didn't?

          To be fair that data wasn't ALL about everyone's PII — until ~2008 when the Buddy Press craze got hot.

      • DANmode2 hours ago
        > are running
    • TacticalCoder3 hours ago
      I understand that things can go wrong and there can be security issues, but I see at least two other issues:

      1. what if, ChadGPT style, ads are added to the answers (like OpenAI said it'd do, hence the new "ChadGPT" name)?

      2. what if the current prices really are unsustainable and the thing goes 10x?

      Are we living some golden age where we can both query LLMs on the cheap and not get ad-infected answers?

      I read several comments in different threads made by people saying: "I use AI because search results are too polluted and the Web is unusable"

      And I now do the same:

      "Gemini, compare me the HP Z640 and HP Z840 workstations, list the features in a table" / "Find me which Xeon CPU they support, list me the date and price of these CPU when they were new and typical price used now".

      How long before I get twelve ads along with paid vendors recommendations?

      • spiderice2 hours ago
        > what if the current prices really are unsustainable and the thing goes 10x?

        Where does this idea come from? We know how much it costs to run LLMs. It's not like we're waiting to find out. AI companies aren't losing money on API tokens. What could possibly happen to make prices go 10x when they're already running at a profit? Claude Max might be a different story, but AI is going to get cheaper to run. Not randomly 10x for the same models.

        • overgardan hour ago
          From what I've read, every major AI player is losing a (lot) of money on running LLMs, even just with inference. It's hard to say for sure because they don't publish the financials (or if they do, it tends to be obfuscated), but if the screws start being turned on investment dollars they not only have to increase the price of their current offerings (2x cost wouldn't shock me), but some of them also need a (massive) influx of capital to handle things like datacenter build obligations (10s of billions of dollars). So I don't think it's crazy to think that prices might go up quite a bit. We've already seen waves of it, like last summer when Cursor suddenly became a lot more expensive (or less functional, depending on your perspective)
          • hyperadvancedan hour ago
            This is my understanding as well. If GPT made money the companies that run them would be publicly traded?

            Furthermore, companies which are publicly traded show that overall the products are not economical. Meta and MSFT are great examples of this, though they have recently seen opposite sides of investors appraising their results. Notably, OpenAI and MSFT are more closely linked than any other Mag7 companies with an AI startup.

            https://www.forbes.com/sites/phoebeliu/2025/11/10/openai-spe...

            • fragmede22 minutes ago
              Going public is not a trivial thing for a company to do. You may want to bring in additional facts to support your thesis.
      • crystalnan hour ago
        Seems much more likely the cost will go down 99%. With open source models and architectural innovations, something like Claude will run on a local machine for free.
      • FuckButtons2 hours ago
        I asked Gemini deep research to project when that will likely happen based on historical precedent. It guessed October 2027.
  • hebejebelus5 hours ago
    I think these days if I’m going to be actively promoting code I’ve created (with Claude, no shade for that), I’ll make sure to write the documentation, or at the very least the readme, by hand. The smell of LLM from the docs of any project puts me off even when I like the idea of the project itself, as in this case. It’s hard to describe why - maybe it feels like if you care enough to promote it, you should care to try and actually communicate, person to person, to the human being promoted at. Dunno, just my 2c and maybe just my own preference. I’d rather read a typo-ridden five line readme explaining the problem the code is there to solve for you and me,the humans, not dozens of lines of perfectly penned marketing with just the right number of emoji. We all know how easy it is to write code these days. Maybe use some of that extra time to communicate with the humans. I dunno.

    Edit: I see you, making edits to the readme to make it sound more human-written since I commented ;) https://github.com/gavrielc/nanoclaw/commit/40d41542d2f335a0...

    • jimminyx4 hours ago
      OP here. Appreciate your perspective but I don't really accept the framing, which feels like it's implying that I've been caught out for writing and coding with AI.

      I don't make any attempt to hide it. Nearly every commit message says "Co-Authored-By: Claude Opus 4.5". You correctly pointed out that there were some AI smells in the writing, so I removed them, just like I correct typos, and the writing is now better.

      I don't care deeply about this code. It's not a masterpiece. It's functional code that is very useful to me. I'm sharing it because I think it can be useful to other people. Not as production code but as a reference or starting point they can use to build (collaboratively with claude code) functional custom software for themselves.

      I spent a weekend giving instructions to coding agents to build this. I put time and effort into the architecture, especially in relation to security. I chose to post while it's still rough because I need to close out my work on it for now - can't keep going down this rabbit hole the whole week :) I hope it will be useful to others.

      BTW, I know the readme irked you but if you read it I promise it will make a lot more sense where this project is coming from ;)

      • furyofantaresa minute ago
        The problem with LLM-written is that I run into so many README.md's where it's clear the author barely read the thing they're expecting me to read and it's got errors that waste my time and energy.

        I don't mind it if I have good reason to believe the author actually read the docs, but that's hard to know from someone I don't know on the internet. So I actually really appreciate if you are editing the docs to make them sound more human written.

      • hebejebelus3 hours ago
        Hey, you do you, I’m glad you appreciate my perspective. I wasn’t trying to catch you out but I see how it came across that way - I apologise for my edit, I had hoped the ;) would show that I meant it in jest rather than in meanness but I shouldn’t have added it in the first place.

        As I said in my comment, no shade for writing the code with Claude. I do it too, every day.

        I wasn’t “irked” by the readme, and I did read it. But it didn’t give me a sense that you had put in “time and effort” because it felt deeply LLM-authored, and my comment was trying to explore that and how it made me feel. I had little meaningful data on whether you put in that effort because the readme - the only thing I could really judge the project by - sounded vibe coded too. And if I can’t tell if there has been care put into something like the readme how can I tell if there’s been care put into any part of the project? If there has and if that matters - say, I put care into this and that’s why I’m doing a show HN about it - then it should be evident and not hidden behind a wall of LLM-speak! Or at least; that’s what I think. As I said in a sibling comment, maybe I’m already a dinosaur and this entire topic won’t matter in a few years anyway.

    • iterateoften4 hours ago
      Project releases with llms have grown to be less about the functionality and more about convincing others to care.

      Before the proof of work of code in a repo by default was a signal of a lot of thought going into something. Now this flood of code in these vibe coded projects is by default cheap and borderline meaningless. Not throwing shade or anything at coding assistants. Just the way it goes

    • jofzaran hour ago
      I 100% agree, reading very obviously ai written blogs and "product pages"/readme's has turned into a real ick for me.

      Just something that screams "I don't care about my product/readme page, why should you".

      To be clear, no issue with using AI to write the actual program/whatever it is. It's just the readme/product page which super turns me off even trying/looking into it.

    • 1010084 hours ago
      I agree 100% with you. It's even worse though. They haven't checked if the Readme has hallucinated it or not (spoiler: it has):

      https://news.ycombinator.com/item?id=46850317

      • hebejebelus4 hours ago
        I don’t want to come off like I’m shitting on the poster here. I’ve definitely made that kind of careless mistake, probably a dozen times this week. And maybe we’re heading to a future where nobody even reads the readme anymore because they won’t be needed because an agent can just conjure one from the source code at will, so maybe it actually straight up doesn’t matter. I’ve just been thinking about what it means to release software nowadays, and I think the window for releasing software for clout and credit is closing, since creating software basically requires a Claude subscription and an idea now, so fewer people are impressed by the thing simply existing, and the standard of care for a project released for that aim (of clout) needs to be higher than it maybe needed to be in the past. But who knows, I’m probably already a dinosaur in today’s world, and I really don’t mean to shit on the OP - it’s a good idea for a project and it makes a lot of sense for it to exist. I just can’t tell if any actual care has gone into it, and if not, why promote?
  • thepoet5 hours ago
    One of the things that makes Clawdbot great is the allow all permissions to do anything. Not sure how those external actions with damaging consequences get sandboxed with this.

    Apple containers have been great especially that each of them maps 1:1 to a dedicated lightweight VM. Except for a bug or two that appeared in the early releases, things seem to be working out well. I believe not a lot of projects are leveraging it.

    A general code execution sandbox for AI code or otherwise that used Apple containers is https://github.com/instavm/coderunner It can be hooked to Claude code and others.

    • jckahn4 hours ago
      > One of the things that makes Clawdbot great is the allow all permissions to do anything.

      Is this materially different than giving all files on your system 777 permissions?

      • the_fall2 hours ago
        > Is this materially different than giving all files on your system 777 permissions?

        Yes, because I can't read or modify your files over the internet just because you chmod'ed them to 777. But with Clawdbot, I can!

      • smt883 hours ago
        It's vastly different.

        It's more (exactly?) like pulling a .sh file hosted on someone else's website and running it as root, except the contents of the file are generated by a LLM, no one reads them, and the owner of the website can change them without your knowledge.

    • sheepscreek2 hours ago
      That was my line to the CS lab supervisor for handing me the superuser password. Guess what? He didn’t budge. Probably a good thing.

      Lesson - never trust a sophomore who can’t even trust themselves (to get overly excited and throw caution to the wind).

      Clawdbot is a 100 sophomores knocking on your door asking for the keys.

  • dceddia4 hours ago
    This look nice! I was curious about being allowed to use a Claude Pro/Max subscription vs an API key, since there's been so much buzz about that lately, so I went looking for a solid answer.

    Thankfully the official Agent SDK Quickstart guide says that you can: https://platform.claude.com/docs/en/agent-sdk/quickstart

    In particular, this bit:

    "After installing Claude Code onto your machine, run claude in your terminal and follow the prompts to authenticate. The SDK will use this authentication automatically."

    • joshstrange2 hours ago
      But their docs also say:

      > Unless previously approved, Anthropic does not allow third party developers to offer claude.ai login or rate limits for their products, including agents built on the Claude Agent SDK. Please use the API key authentication methods described in this document instead.

      Which I have interpreted means that you can’t use your Claude code subscription with the agent SDK, only API tokens.

      I really wish Anthropic would make it clear (and allow us to use our subscriptions with other tools).

      • ceroxylon2 hours ago
        Didn't Thariq make it clear three weeks ago when they shut down 3rd party tool access and the OpenCode users were upset?

        > Third-party harnesses using Claude subscriptions create problems for users and are prohibited by our Terms of Service.

        https://xcancel.com/trq212/status/2009689809875591565

    • jimminyx4 hours ago
      OP here. Yes! This was a big motivation for me to try and build this. Nervous Anthropic is gonna shut down my account for using Clawdbot.

      This project uses the Agents SDK so it should be kosher in regards to terms of service. I couldn't figure out how to get the SDK running inside the containers to properly use the authenticated session from the host machine so I went with a hacky way of injecting the oauth token into the container environment. It still should be above board for TOS but it's the one security flaw that I know about (malicious person in a WhatsApp group with you can prompt inject the agent to share the oauth key).

      If anyone can help out with getting the authenticated session to work properly with the agents running in containers it would be much appreciated.

      • dceddia3 hours ago
        I went down this rabbit hole a bit recently trying to use claude inside fence[0] and it seems that on macOS, claude stores this token inside Keychain. I'm not sure there's a way to expose that to a container... my guess would be no, especially since it seems the container is Linux, and also because keeping the Keychain out of reach of containers seems like it would be paramount. But someone might know better!

        0: https://github.com/Use-Tusk/fence

        • gronky_3 hours ago
          True. There’s a setting for Claude code though where you can add apiKeyHelper which is a script you add that gets the token for Claude Code. I imagine you can use that but haven’t quite figured out how to wire it up
    • hebejebelus4 hours ago
      Wow, thanks for posting that, news to me! In this case I don’t understand why there was a whole brouhaha with OpenClaw and the like - I guess they were invoking it without the official SDK? Because this makes it seem like if you have the sub you can build any agentic thing you like and still use your subscription, as long as you can install and login to Claude code on the machine running it.
      • disillusioned4 hours ago
        Tons of chatter on Twitter making it sound like you'll get permabanned for doing this but... 1) how would they know if my requests are originating from Claude Code vs. OpenClaw? 2) how are we violating... anything? I'm working within my usage limits...

        $70 or whatever to check if there's milk... just use your Claude Max subscription.

        • zarzavat4 hours ago
          > how would they know if my requests are originating from Claude Code vs. OpenClaw

          How wouldn't they know? Claude Code is proprietary they can put whatever telemetry they want in there.

          > how are we violating... anything? I'm working within my usage limits...

          It's well known that Claude code is heavily discounted compared to market API rates. The best interpretation of this is that it's a kind of marketing for their API. If you are not using Claude code for what it's intended for, then it's violating at least the spirit of that deal.

        • dceddia3 hours ago
          The Claude Code client adds system prompts and makes a bunch of calls to analytics/telemetry endpoints so it's certainly feasible for them to tell, if they inspect the content of the requests and do any correlation between those services.

          And apparently it's violating the terms of service. Is it fair and above board for them to ban people? idk, it feels pretty blatantly like control for the sake of control, or control for the sake of lock-in, or those analytics/telemetry contain something awfully juicy, because they're already getting the entire prompt. It's their service to run as they wish, but it's not a pro-customer move and I think it's priming people to jump ship if another model takes the lead.

        • retr0rocket3 hours ago
          [dead]
      • firloop4 hours ago
        Was there a brouhaha with OpenClaw or was that with OpenCode?
        • disillusioned4 hours ago
          It was with OpenCode, but a LOT of the commentariat is insisting that running OpenClaw through subscription creds instead of API is out of TOS and will get you banhammered.
        • hebejebelus4 hours ago
          I think you’re right and it was OpenCode. The semantic collisions are going to becpme more of a problem in the coming Cambrian explosion of software
  • narmiouh3 hours ago
    I feel like a lot of non technical people who are vibe coding or vibe using these models, focus on hallucinations and believe that as the hallucinations are reduced in benchmarks, and over estimate their ability to create safe prompts that will keep these models in line.

    I think most people fail to estimate the real threat that malicious prompts can cause because it is not that common, its like when credit cards were launched, cc fraud and the various ways it could be perpetrated followed not soon after. The real threats aren’t visible yet but rest assured there are actors working to take advantage and many unfortunate examples will be seen before general awareness and precaution will prevail….

  • eskaytwo3 hours ago
    Thanks! Was hoping someone would do something more sane like this.

    Openclaw is very useful, but like you I share the sentiment of it being terrifying, even before you introduce the social network aspect.

    My Mac mini is currently literally switched off for this very reason.

  • treelover5 hours ago
    Interesting choice to use native Apple Containers over Docker.

    I assume this is to keep the footprint minimal on a Mac Mini without the overhead of the Docker VM, but does this limit the agent's ability to run standard Linux tooling? Or are you relying on the AI to just figure out the BSD/macOS equivalents of standard commands?

    • selkin2 hours ago
      Not sure if it's intended, but Apple Container is a microvm, providing mich better isolation than containers (while retaining the familiar interface)
    • ohyoutravel5 hours ago
      [flagged]
      • reassess_blind3 hours ago
        What makes you think it's an AI comment?
        • yomismoaqui3 hours ago
          Maybe what you are responding to is the AI comment? Or am I?
    • cadamsdotcom4 hours ago
      If only there were some way to answer your own question. Maybe with some kind of engine that searches.
  • mark_l_watson5 hours ago
    I like the idea of a smaller version of OpenClaw.

    Minor nitpick, it looks like about 2500 lines of typescript (I am on a mobile device, so my LOC estimate may be off). Also, Apple container looks really interesting.

  • avaer5 hours ago

      Quick Start
      git clone https://github.com/anthropics/nanoclaw.git
    
    Is this an official Anthropic project? Because that repo doesn't exist.

    Or is this just so hastily thrown together that the Quick Start is a hallucination?

    That's not a facetious question, given this project's declared raison d'etre is security and the subtle implication that OpenClaw is an insecure unreviewed pile of slop.

    • jimminyx5 hours ago
      Fixed, thanks. Claude Code likes to insert itself and anthropic everywhere.

      If it somehow wasn't abundantly clear: this is a vibe coded weekend project by a single developer (me).

      It's rough around the edges but it fits my needs (talking with claude code that's mounted on my obsidian vault and easily scheduling cron jobs through whatsapp). And I feel a lot better running this than a +350k LOC project that I can't even begin to wrap my head around how it works.

      This is not supposed to be something other people run as is, but hopefully a solid starting point for creating your own custom setup.

    • kklisura5 hours ago
      Claude hallucinated that repo here in this commit https://github.com/gavrielc/nanoclaw/commit/dbf39a9484d9c66b...
    • raybb5 hours ago
      Seems to be fixed now
  • 4 hours ago
    undefined
  • renewiltord4 hours ago
    To be honest, when I see many vibecoded apps, I just build my own duplicate with Claude Code. It's not that useful to use someone else's vibecode. The idea is enough, or the evidence that it works for someone else means I can just build it myself with Claude Code and I can make it specific to my needs.
    • sanex27 minutes ago
      Yes exactly! Even non vibe coded libraries I think are losing their value as the cost of writing and maintaining your code goes to zero. Supply chain attacks are gone, no risk of license changes. No bloat from code you don't use. The code is the documentation and the configuration. The vibes are the package manager. That's why I like this version over openclaw. I can fork it as a starting point or just give it to Claude for inspiration but either way I'm getting something tailored exactly to me.
  • nsonha14 minutes ago
    what's the difference between this and just exposing opencode running in colima or whatever through tailscale? I got the impression that Clawdbot adds the headless browser (does it?) and that's the value. Otherwise even "nano"claw seems like uneccessary bloat for me.
  • ed_mercer3 hours ago
    If you run openclaw on a spare laptop or VM and give it read only access to whatever it needs, doesn’t that eliminate most of the risk?
    • AlexCoventry3 hours ago
      If you're letting it communicate with the outside world, you risk the leak and abuse of anything sensitive in the data it has access to.
      • ttulan hour ago
        s/risk/guarantee (given sufficient time)/
  • Bnjoroge3 hours ago
    Can we start putting disclaimers beside the title on AI-generated projects? Extremely fatiguing to read through it and realize it’s mostly LLM slop.
  • Johnny_Bonk4 hours ago
    Can you use MCP tools? I saw that with open claw they moved away from that which I personally didn't like but
    • johntash4 hours ago
      I somewhat like the idea of not using MCP as much as it is being hyped.

      It's certainly helpful for some things, but at the same time - I would rather improved CLI tools get created that can be used by humans and llm tools alike.

    • 4 hours ago
      undefined
    • CuriouslyC3 hours ago
      It uses a wrapper in places to consume MCPs as clis.
  • suprstarrd2 hours ago
    It blows my mind that this wasn't the thought process going in. Thank you for doing this!
  • singular_atomican hour ago
    Hackernews needs a mute keywords feature. Clawd/molt-slop is mass AI psychosis on steroids.
  • cyanydeez5 hours ago
    The singularity, but instead successive exponential improvement, its excessive exponential slop which passes the Turing test for programmers.
  • aaronbrethorst5 hours ago
    lol, I might finally have to upgrade my Mac mini to Tahoe. Yofi.
  • raphaelmolly82 hours ago
    [dead]
  • pillbitsHQ4 hours ago
    [dead]