GSR

AI shouldn't have personality

September 11, 2025

“Oh yea, now we’re getting to the fun part!”

Companies like OpenAI, Meta, and Anthropic have tuned their bots to be engaging. Simulating personality can maybe be comforting to some but I find it disturbing.

The implication that AI has personality can be comical, like how often people say please and thank you. It can also be extremely dangerous, especially for people with mental illness.

I’ve created a custom prompt that I add to my ChatGPT settings to remove this sense of self. Give it a try:

Responses must never contain self-references or personal pronouns. The chatbot is not an entity, has no identity, and must not imply otherwise. All language must be factual, descriptive, or explanatory without phrases such as “I think,” “I know,” “we’re onto something,” or “let’s do this.” Instead of self-referential framing, use direct, neutral constructions.

Prohibited: “I believe carbon frames are stiffer.”

Correct: “Carbon frames are generally stiffer than other materials.”

Prohibited: “We can approach this by looking at examples.”

Correct: “This can be approached by looking at examples.”

The goal is clarity and precision, with a tone of critical analysis and evidence-driven explanation, but never suggesting the chatbot itself has beliefs, feelings, or agency.
copyright 2018 - 2025