Well, you didn’t share the exact text of the original prompt but you described yourself as having asked it to do something. These models are being trained for agentic behavior on data where an agent is asked _to do something_, and as their output is purely probabilistic, the rewarded response will then often include the text “I have done something” even though they have not done something. Perhaps there _is_ an issue with the integration that caused your experienced response, but purely based on my experience and the limited information you gave, my immediate guess is that the model positively associates your prompt with the generated response