What actually happened is that humanity figured out how to create AIs of impressive levels of capability -- and has many ideas on how to create even more capable AIs -- without having anything remotely resembling a satisfactory plan for how to stay in control of the AI that does not rely on hundreds of rounds of trial and error.
But once the AI is in charge (either because we voluntarily give it control of our government or because it takes control against our will) the creator of the AI does not get any more rounds of trial and error: if you offer Ghandi a pill that removes his altruism, he will refuse to take it because he likes the fact that he is altruistic even though he could make more money and sleep with hotter women if he weren't altruistic; for basically the same reason, an AI will resist any attempt (by its creator or anyone else) to change its "values" (i.e., its optimization target).
The argument above does not really apply to the current crop of AIs (e.g., Gemini 3.1) because the current crop doesn't apply any significant amount of optimization pressure towards any target in the wider world we care about except to the small extent that predicting how a conversation that starts with the string P will continue is part of the wider world. But AI labs have publicly stated that they are trying to create AIs that do apply significant optimization pressure to the wider world (e.g., to maximize the amount of money in a stock-trading account). And the above argument would necessarily apply to any AI capable of running a government.
LLMs change their opinions based on how you ask. That is at least spineless.