If you have non-technical stakeholders managing the deployment or making decisions about AI use, API providers give you more guardrails and less operational surface area. Local LLMs require someone who understands the stack.
That said, the AI literacy gap on the business/governance side is real — teams where managers don't have AI foundations tend to get stuck on escalations. Programs like IAIDL (https://iaidl.org) address that side of the equation for non-technical colleagues. Reduces the "what is it actually doing" conversation overhead considerably.
That said, "worth it" still depends heavily on your hardware. A 4070 Ti gets you a very different answer than a 3060.
Disclosure: I'm building localllm-advisor.com, free and client-side, which also helps answer these types of questions. It shows which models fit your GPU with quantization options and estimated tok/s, or which GPU you'd need to run a specific model. Relevant to the question so I'm mentioning it, but take it for what it is.