Here’s the thing: Grok didn’t say anything. Grok didn’t
blame anyone. Grok didn’t apologize. Grok can’t do any
of these things, because Grok is not a sentient entity
capable of speech acts, blame assignment, or remorse.
What actually happened is that a user prompted Grok to generate
text about the incident. The chatbot then produced a word sequence
that pattern-matched to what an apology might sound like, because
that’s what large language models do. They predict statistically
likely next tokens based on their training data.
When you ask an LLM to write an apology, it writes something that
looks like an apology. That’s not the same as actually apologizing.At least with LLMs it's not too hard to figure what's going on, unlike certain politicians.
LLMs that aren't chat tuned are just not as easy to anthropomorphize.
If you mean being told by the end user, this famously hasn't been the case. Dialing back the only restriction was enough for Grok to create nsfw material (w/o any request to create that).
[Grok] didn’t hesitate to spit out fully uncensored topless
videos of Taylor Swift the very first time I used it
without me even specifically asking the bot to take her clothes off.I have tried to offer corrections to incorrect headlines and technical information about LLMs over the past few years but have stopped because I don't have the bandwidth to deal with the "so you support the plagiarism machine" comments every time.
strangers were replying to women’s photos and asking Grok, the platform’s built-in AI chatbot, to “remove her clothes” or “put her in a bikini.” And Grok was doing it. Publicly. In the replies. For everyone to see.
Wow. Thats some really creepy behavior people are choosing to show off publicly.
Grok needs some tighter gaurdrails to prevent abuse.