The more interesting consequence is that when inference and fine-tuning are essentially free at startup scale, specialization becomes viable again. Instead of generic prompting against a closed API, teams can afford narrow, high-precision models tailored to their domain — something that used to be economically out of reach. Came across this interesting post - https://www.linkedin.com/feed/update/urn:li:activity:7396291...
Cheap and free to download. Most developers would rather spend weeks rebuild something for themselves than pay $20 a month for a tool.
Big AI has prompts you cannot remove. They have to because they have a big audience, get attacked relentlessly, and have to be mindful of PR events.
Now, while I can avoid the copilot/Claude code agent prompts, I am still using their models directly and subject to their prompts. Moving to use models directly is the next step, and the only way to do that is with open models. Therein, the Chinese have been building better open models, and that is why we see their usage rising.
It's more about full stack control than it is about price (imo)