79 pointsby linolevan13 hours ago13 comments
  • bramhaag6 hours ago
    Only tangentially related, but today I found a repo that appears to be developed using AI assistance, and the costs for running the agents are reported in the PRs. For example, 50 USD to remove some code: https://github.com/coder/mux/pull/1658
    • AnotherGoodName3 hours ago
      This is the cost of Anthropics pay by the token plan.

      To give an analogy, Anthropics pricing is $0.10 per grain of rice (pay by the token) or;

      $20 a month for quarter cup of rice each day (claude pro)

      $100 a month for 10 cups of rice each day (claude max 100)

      $200 a month for a sack of rice delivered to your door each day.

      It's a rather insane pricing scale and here they are paying $50 there because they don't understand the pricing model (which is fair, Anthropics pricing model is crazy). Never pay by the token with Antropic! Only ever use the subscription plans.

    • msephton4 hours ago
      I'd like to suggest that dev lower their settings: sota model + high + thinking is definitely not needed to do this simple task. Lower settings could easily do it for less than $0.50, maybe even $0.05. I'd encourage people to operate on average to low settings and wind then up or down depending on the prompt task complexity.
    • RestartKernel5 hours ago
      That seems reasonable compared to an actual developer (depending on your region), but I had hoped for these models to make simple tasks like this fast and cheap so those developers can focus on the difficult stuff.
  • nospice7 hours ago
    I'm not sure I like this method of accounting for it. The critics of LLMs tend to conflate the costs of training LLMs with the cost of generation. But this makes the opposite error: it pretends that training isn't happening as a consequence of consumer demand. There are enormous resources poured into it on an ongoing basis, so it feels like it needs to be amortized on top of the per-token generation costs.

    At some point, we might end up in a steady state where the models are as good as they can be and the training arms race is over, but we're not there yet.

    • Aurornis6 hours ago
      That's not really an error, that's a fundamental feature of unit economics.

      Fixed costs can't be rolled into the unit economics because the divisor is continually growing. The marginal costs of each incremental token/query don't depend on the training cost.

    • cortesoft6 hours ago
      It would be really hard to properly account for the training, since that won't scale with more generation.

      The training is already done when you make a generative query. No matter how many consumers there are, the cost for training is fixed.

      • nospice6 hours ago
        My point is that it isn't, not really. Usage begets more training, and this will likely continue for many years. So it's not a vanishing fixed cost, but pretty much just an ongoing expenditure associated with LLMs.
        • bob10296 hours ago
          No one doing this for money intends to train models that will never be amortized. Some will fail and some are niche, but the big ones must eventually pay for themselves or none of this works.

          The economy will destroy inefficient actors in due course. The environmental and economic incentives are not entirely misaligned here.

          • quietbritishjim6 hours ago
            > No one doing this for money intends to train models that will never be amortized.

            Taken literally, this is just an agreement with the comment you're replying to.

            Amortizing means that it is gradually written off over a period. That is completely consistent with the ability to average it over some usage. For example, if a printing company buys a big new printing machine every 5 years (because that's how long they last before they wear out), they would amortize it's cost over the 5 years (actually it's depreciation not amortization because it's a physical asset but the idea is the same). But it's 100% possible to look at the number of documents they print over that period and calculate the price of the print machine per document. And that's still perfectly consistent with the machine paying for itself.

    • TSiege7 hours ago
      The challenge with no longer developing new models is making sure your model is up to date which as of today requires an entire training run. Maybe they can do that less or they’ll come up with a way to update a model after it’s trained. Maybe we’ll move onto something other than LLMs
    • skybrian5 hours ago
      The training cost is a sunk cost for the current LLM, and unknown for the next-generation LLM. Seems like it would be useful information but doesn't go here?
    • robocat5 hours ago
      The AI training data sets are also expensive... The cost is especially hard to estimate for data sets that are internal to businesses like Google. Especially if the model needs to be refreshed to deal with recent data.

      I presume historical internal datasets remain high value, since they might be cleaner (no slop) or maybe unavailable (copyright takedowns) and companies are getting better at hiding their data from spidering.

  • simonw8 hours ago
    At first glance this looks like a credible set of calculations to me. Here's the conclusion:

    > So, if I wanted to analogize the energy usage of my use of coding agents, it’s something like running the dishwasher an extra time each day, keeping an extra refrigerator, or skipping one drive to the grocery store in favor of biking there.

    That's for someone spending about $15-$20 in a day on Claude Code, estimated at the equivalent of 4,400 "typical queries" to an LLM.

    • Aurornis6 hours ago
      Comparing it to running a refrigerator or the dishwasher is very relatable, as most people have at least one refrigerator without a second thought.

      This is for someone using a lot of LLM tokens relative to the average customer of these companies.

  • linolevan8 hours ago
    Had a small discussion about this on an OP on bsky. A somewhat interesting discussion over there.

    https://bsky.app/profile/simonpcouch.com/post/3mcuf3eazzs2c

  • matthewfcarlson5 hours ago
    I would like some real world comparisons. How much power does the laptop or desktop consume during these (likely multi hour) sessions? Assuming you’re using a large HDR monitor 50-100W isn’t unreasonable and at 8 hours a day you’re talking about at least 2 days before you crack 1000kwh like his sessions do. But then a personal desktop on a gaming session can easily pull 1000w (cpu + gpu + peripherals). So comparing it to a gaming session seems fair.
  • ggm8 hours ago
    As long as it's unaccounted for by users it's at best anexternality. I think it may demand regulation to force this cost to the surface.

    electricity and cooling incur wider costs and consequences.

    • simonw7 hours ago
      That's hardly unique to data centers.

      I'm all for regulation that makes businesses pay for their externalities - I'd argue that's a key economic role that a government should play.

      • ggm5 hours ago
        No disagree. I think all enterprises that come with community borne costs (noise, heat, energy use, road use, construction & infrastructure, state incentives) and benefits (tax revenue, jobs) should have some level of accounting. It would be wrong to say the negs always outweigh the positives, thats not the point here. The point is that a bunch of cost in DC relating to power and cooling wind up having impact on the community at large.

        I've been told in other (non US) economies, decisions to site hyperscaler DCs has had downstream impacts on power costs and longterm power planning. The infra to make a lot of power appear at a site, means the same capex and inputs cannot be used to supply power to towns and villages. There's a social opportunity loss in hosting the DC because the power supply doesn't magicly make more transformers and wires and syncons appear on the market: Prices for these things are going up because of a worldwide shortage.

        Its like the power version of worldwide RAM pricing.

    • Majromax6 hours ago
      > As long as it's unaccounted for by users it's at best anexternality.

      Why is it an externality? Anthropic (or other model provider) pays the electricity cost, then it's passed along in the subscription or API bill. The direct cost of the energy is fully internalized in the price.

      • ggm5 hours ago
        Electricity supply is about more than supply cost. It has a build cost, with assumed inputs which are in turn priced by demand, and they're in short supply worldwide. DC costs of construction in power are pushing up the delay and cost for non-DC power projects.
    • avalys5 hours ago
      No! What a disaster. Enforce the costs of externalities as close to the source as possible. If electricity costs more money charge more money for electricity. Don’t add some “regulation” to force end users to pay more based on your estimate of how much more you think electricity should cost.
      • ggm4 hours ago
        I said force the cost to the surface. Not set a fixed rate of return: I want the accounting on the build and supply cost at large, as well as the KW charges to be accounted for.
    • pixl976 hours ago
      I mean if we really cared about this one bit we'd stop making a car based society in the US and save far more energy and pollution. That's not politically expedient and there are powerful vested interests in ensuring it doesn't happen.

      That's why I think most of this data center energy use, especially over longer terms is a joke. Data center can pretty easily run on solar and wind energy if we spend even a small amount to political capital to make it happen.

      • ggm5 hours ago
        They're a lot harder to build out from those sources because they are costed to run 24/7 and the intermittency issue comes to the fore. Unlike things like aluminium smelters there isn't always a good load-shed or even supply-timing story in a DC, cooling aside (big chunks of cooling can be used for demand management)

        I am not in the DC business. if somebody who is says "thats bunkum" I'd pay attention to it.

    • jeffbee7 hours ago
      I don't see how this follows. Data center operators buy energy and this is almost their only operating expense. Their products are priced to reflect this. The fact that basic AI features are free reflects the fact that they use almost no energy.
      • arrowleaf7 hours ago
        I would be surprised if AI prices reflect their current cost to provide the service, even inference costs. With so much money flowing into AI the goal isn't to make money, it's to grow faster than the competition.
        • Majromax6 hours ago
          From this article:

          > For the purposes of this post, I’ll use the figures from the 100,000 “maximum”–Claude Sonnet and Opus 4.5 both have context windows of 200,000 tokens, and I run up against them regularly–to generate pessimistic estimates. So, ~390 Wh/MTok input, ~1950 Wh/MTok output.

          Expensive commercial energy would be 30¢ per kWh in the US, so the energy cost implied by these figures would be about 12¢/MTok input and 60¢/MTok output. Anthropic's API cost for Opus 4.5 is $5/MTok input and $25/MTok output, nearly two orders of magnitude higher than these figures.

          The direct energy cost of inference is still covered even if you assume that Claude Max/etc plans are offering a tenfold subsidy over the API cost.

          • ggm5 hours ago
            Thank you for some good intel. Thats very interesting. But, I wonder how this affects supply pricing to other customers. Not that you haven't shown the direct power costs have been borne, but the more indirect ones remain for me.
        • simonw7 hours ago
          I remain confident that most AI labs are not selling API access for less than it costs to serve the models.

          If that's so common then what's your theory as to why Anthropic aren't price competitive with GPT-5.2?

          • JimDabell3 hours ago
            I think it’s more instructive to look at providers like AWS than to compare with other AI labs. What’s the incentive for AWS to silently subsidise somebody else’s model when you run it on their infrastructure?

            AWS are quite happy to give service away for free in vast quantities, but they do it by issuing credits, not by selling below cost.

            I think it’s a fairly safe bet AWS aren’t losing money on every token they sell.

        • Aurornis6 hours ago
          > I would be surprised if AI prices reflect their current cost to provide the service, even inference costs.

          This has been covered a lot. You can find quotes from one of the companies saying that they'd be profitable if not for training costs. In other words, inference is a net positive.

          You have to keep in mind that the average customer doesn't use much inference. Most customers on the $20/month plans never come close to using all of their token allowance.

  • throwerxyz5 hours ago
    "How much energy does it take to eat meat? (ignoring the cost to produce the meat into your hands)"

    Do people even care about this?

    How much energy does it take to download a video on YouTube versus the energy input to keep it all setup and running?

  • scottcha7 hours ago
    That is a pretty good article although the one factor not mentioned that we see that has a huge impact on energy is batch size but that would be hard to estimate with the data he has.

    We've only launched to friends and family but I'll share this here since its relevant: we have a service which actually optimizes and measures the energy of your AI use: https://portal.neuralwatt.com if you want to check it out. We also have a tools repo we put together that shows some demonstrations of surfacing energy metadata in to your tools: https://github.com/neuralwatt/neuralwatt-tools/

    Our underlying technology is really about OS level energy optimization and datacenter grid flexibility so if you are on the pay by KWHr plan you get additional value as we continue to roll new optimizations out.

    DM me with your email and I'd be happy to add some additional credits to you.

    • ccgibson7 hours ago
      To add a bit more to what @scottcha is saying: overall GPU load has a fairly significant impact on the energy per result. Energy per result is inversely related, since the idle TDP of these servers is significant the more the energy gets spread the more efficient the system becomes. I imagine Anthropic is able to harness that efficiency since I imagine their servers are far from idle :)
      • Majromax6 hours ago
        You can infer the discount from the pricing of the batch API, which is presumably arranged for minimum inference costs. Anthropic offers a 50% discount there, which is consistent with other model providers.
  • xk35 hours ago
    So less energy than a human brain uses...
  • maxdo6 hours ago
    Us person does not consume 1600 liters a day
    • ipnon6 hours ago
      I think this number is derived from the water used to create all the goods they consume. Let's say every American eats 1lb of beef per day, and nothing else, to create a simple model. Then every American has a water footprint of about 1,850 gallons.

      [0] https://watercalculator.org/news/articles/beef-king-big-wate...

    • NicuCalcea6 hours ago
      It does seem high, most of the estimates I find are around half that.
    • simonw6 hours ago
      What number would you provide for that?
  • mikeaskew47 hours ago
    I have a kids and a dishwasher (which with kids, runs quite often) but I’m not convinced I’m doing worse at energy consumption
  • renewiltord5 hours ago
    lol who honestly gives a shit. You guys drive cars.
  • HNisCIS8 hours ago
    LLMs don't use much energy at all to run, they use it all at the beginning for training, which is happening constantly right now.

    TLDR this is, intentionally or not, an industry puff piece that completely misunderstands the problem.

    Also, even if everyone is effectively running a a dishwasher cycle every day, this is still a problem that we can't just ignore, that's still a massive increase in ecological impact.

    • simonw8 hours ago
      The training cost for a model is constant. The more individual use that model gets the lower the training-cost-per-inference-query gets, since that one-time training cost is shared across every inference prompt.

      It is true that there are always more training runs going, and I don't think we'll ever find out how much energy was spent on experimental or failed training runs.

      • dietr1ch7 hours ago
        > The training cost for a model is constant

        Constant until the next release? The battle for the benchmark-winning model is driving cadence up, and this competition probably puts a higher cost on training and evaluation too.

        • simonw7 hours ago
          Sure. By "constant" there I meant it doesn't change depending on the number of people who use the model.
          • dietr1ch4 hours ago
            I got that part, it's just that it overlooks the power consumption of the AI race.

            I wish we reach the everything is equally bad phase so we can start enjoying the more constant cost of the entire craze to build your own model with more data than the rest.

    • kingstnap7 hours ago
      You underestimate the amount of inference and very much overestimate what training is.

      Training is more or less the same as doing inference on an input token twice (forward and backward pass). But because its offline and predictable it can be done fully batched with very high utilization (efficiently).

      Training is guestimate maybe 100 trillion total tokens but these guys apparently do inference on the quadrillion token monthly scales.

    • linolevan8 hours ago
      I'm not convinced that LLM training is at such a high energy use that it really matters in the big picture. You can train a (terrible) LLM on a laptop[1], and frankly that's less energy efficient than just training it on a rented cloud GPU.

      Most of the innovation happening today is in post-training rather than pre-training, which is good for people concerned with energy use because post-training is relatively cheap (I was able to post-train a ~2b model in less than 6 hours on a rented cluster[2]).

      [1]: https://github.com/lino-levan/wubus-1 [2]: https://huggingface.co/lino-levan/qwen3-1.7b-smoltalk

    • jeffbee7 hours ago
      Training is pretty much irrelevant in the scheme of global energy use. The global airline industry uses the energy needed to train a frontier model, every three minutes, and unlike AI training the energy for air travel is 100% straight-into-your-lungs fossil carbon.
      • pluralmonad7 hours ago
        Not to mention doesn't aviation fuel still make heavy (heh) use of lead?
        • TSiege7 hours ago
          I think thats only true for propeller planes, which use leaded gasoline. Jet fuel is just kerosene
          • tialaramex5 hours ago
            Pistons, rather than all propellers. Basically imagine a really old car engine, because simplicity is crucial for reliability and ease of maintenance so all those "fancy" features your car had by the 1990s aren't available, however instead of turning wheels the piston engine turns a propeller. Like really old car engines these piston engines tend to be designed for leaded fuel. Because this is relatively cheap to do, all the cheapest planes aimed at GA (General Aviation, ie you just like flying a plane, not for pay) are like this.

            Propellers are a very common means to make aeroplanes work though, instead of a piston engine, which is cheap to make but relatively unreliable and expensive to run, you can use turbine engines, which run on JetA aka kerosene, and the rotary motion of the turbine drives the propeller making a turboprop. In the US you won't see that many turboprop engines for passenger service, but in the rest of the world that's a very common choice for medium distance aeroplane routes, while the turbofan planes common everywhere in the US would in most places be focused on longer distances between bigger airfields because they deliver peak efficiency when they spend longer up in the sky.

            JetA, whether for a turbofan or turboprop does not have lead in it, so to a first approximation no actual $$$ commercial flights spew lead. They're bad for the climate, but they don't spew lead into the atmosphere.