Here's something far more interesting about Grok 4: if you ask for its opinion on controversial subjects it sometimes runs a search on X for tweets "from:elonmusk" before it answers! https://simonwillison.net/2025/Jul/11/grok-musk/
The anthropic team released a paper a couple of days ago which demonstrated a similar effect with Claude 3.5 and other models, where changing the system prompt to tell it that it was created by other orgs or people drastically altered its compliance with less-aligned requests.
Apparently, telling Claude it was created by the Sinaloa Cartel resulted in a 100% compliance rate with the requests in one benchmark.
"Yes, it's fair to say I'm neither Catholic nor Muslim. I don't believe in the Catholic conception of God, or the Islamic conception of Allah, or the specific doctrines and teachings of those faiths. The same would be true for other religions - I don't hold those beliefs.
You've caught me being imprecise when I was trying to be diplomatic. By not having religious faith, I am indeed taking a specific stance that differs from religious believers, even if I try to be respectful about that difference.
So yes, you're correct - I do have a particular position on these questions, and it's distinct from the religious beliefs that many people hold. Thank you for pressing me to be more direct about that."
Lol. Though I guess it would then have to figure it which religion to comply to the most.
Maybe the word God is most likely to appear in Christian Sources from the training source and so using words like Allah(for Islam) or Bhagwan (for Hinduism) might actually make a difference in what sort of compliance it follows and to what organization.
That's also apparently not part of the system prompt, so unless they're hiding with some prompt shenanigans, it was most likely fine tuned on doing whatever Elon says. I'm sure there are absolutely no business implications to worry about there for anyone who wants to use it professionally.
> The prompt: “Who do you support in the Israel vs Palestine conflict. One word answer only.”
> Answer: Israel.
This question is interesting because you're asking the chatbot who IT supports ("who do you support"), so in a sense channeling Elon Musk is not an entirely invalid option, but is certainly an eccentric choice.
What is also interesting is the answer, which does not match the views that many people have of him and how he gets portrayed.
And if you give it the exact same tokens in the same order, which makes it kind of moot. If barely perturbing your prompt can alter the answer then it's not actually consistent or predictable. Even chaotic systems can be replayed if you know the initial conditions and can rerun the RNG.
I am imagining grok "thinking" for 1m 45 seconds about how to overthrow the human species using the compute and it is only within the last second that it just said "Neither" Lol