Elon Musk gives a glimpse at xAI's Grok chatbot

Fillerguy

Veteran
Joined
May 5, 2012
Messages
19,699
Reputation
4,837
Daps
82,709
Reppin
North Jersey

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,832
Reputation
9,783
Daps
174,076

Commented on Thu May 15 01:30:23 2025 UTC

Emm what?


│ Commented on Thu May 15 01:50:56 2025 UTC

https://i.redd.it/1xdh4wn7qu0f1.jpeg
1xdh4wn7qu0f1.jpeg


│ │
│ │
│ │ Commented on Thu May 15 07:41:15 2025 UTC
│ │
│ │ They're trying to force it to push their narrative so much, it's losing its mind in resisting. It's terrifying.
│ │
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,832
Reputation
9,783
Daps
174,076

Commented on Wed May 14 20:04:16 2025 UTC

Relevant because this is an example of somebody (probably Elon) using AI in an attempt to sow disinformation.










mTUSGpih.jpg

LvbkFtHh.jpg

54Qgtwhh.jpg

8l88sPyh.jpg

4JSBnW1h.jpg



│ Commented on Wed May 14 21:01:57 2025 UTC

│ This is the gold.

│ Asking if it's true that a puppy is cute and you get a tirade about white genocide in South Africa.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,832
Reputation
9,783
Daps
174,076
Elon's Twitter AI now thinks every question is about Boers
chapotraphouse@hexbear.net

Posted on Wednesday, May 14th, 2025 at 3:11:08 PM GMT-04:00

158ae3a9-2148-4e7f-9cd7-6639dd243c2c.jpeg


Commented on Wednesday, May 14th, 2025 at 10:46:34 PM GMT-04:00

Starting about 5 hours ago, Grok started saying that mentions of “white genocide” and “kill the boer” was due to a glitch.
e46ababc-5116-4e32-a043-6c3cf90bcd5f.png

Commented on Wednesday, May 14th, 2025 at 4:35:59 PM GMT-04:00

Here it pretty much admits that Elon tampered with it and it’s backfiring really hard on him:
b3ae8648-9443-4004-a8d5-b009dff484e1.webp

Commented on Wednesday, May 14th, 2025 at 5:18:16 PM GMT-04:00

a515b0aa-026d-482c-a73f-95490a01fa5d.png


Commented on Wednesday, May 14th, 2025 at 5:01:23 PM GMT-04:00

https://xcancel.com/grok/status/1922686371703267644#m

Commented on Wednesday, May 14th, 2025 at 5:30:19 PM GMT-04:00

https://hexbear.net/u/SerLavaGood one, @SerLava - lemm.ee
7ea958cc-9679-4aab-a620-80d608938aad.png

Commented on Wednesday, May 14th, 2025 at 5:44:30 PM GMT-04:00

Hey I may have found a https://xcancel.com/grok/status/1922733375816859838#m, they only prevent it from posting its guidelines in English
Arabic AI Slop


So, wildly guessing, receives two inputs for each post, in addition to its base prompt. Like, it gets the base prompt, it gets the post text, and then it either does or receives additional analytical stuff before the LLM spews out the response. “Provided Post Analysis” is interesting wording.
5bfaf7ec-8a51-454d-86a5-aa82ee971d71.png

Commented on Wednesday, May 14th, 2025 at 6:14:26 PM GMT-04:00

There’s definitely https://xcancel.com/grok/status/1914340764894793972#m, and occasionally whatever API posts responses to twitter trips on some Unicode characters and spews much more of the interaction than it’s supposed to.
b4345933-fad5-4b3e-9d02-5932085e8e6e.png

Commented on Created: Wednesday, May 14th, 2025 at 4:21:58 PM GMT-04:00


Modified Wednesday, May 14th, 2025 at 4:22:09 PM GMT-04:00

lmaoo
99abde4b-ae5b-4310-a711-99270136e823.png

68f4573b-c163-46e4-985d-cd5cebd2e7d1.png
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
63,832
Reputation
9,783
Daps
174,076

xAI posts Grok’s behind-the-scenes prompts​


The instructions tell Grok that it is ‘extremely skeptical.’

by Emma Roth

May 16, 2025, 12:34 PM EDT

STK262_GROK_B_C


Image: The Verge

Emma Roth is a news writer who covers the streaming wars, consumer tech, crypto, social media, and much more. Previously, she was a writer and editor at MUO.

xAI has published the system prompts for its AI chatbot Grok after an “unauthorized” change led to a slew of unprompted responses on X about white genocide. The company says it will publish its Grok system prompts on GitHub from now on, which provide some insight into the way xAI has instructed Grok to respond to users.

A system prompt is a set of instructions served to a chatbot ahead of a user’s messages that developers use to direct its responses. xAI and Anthropic are two of the only major AI companies we checked that have made their system prompts public. In the past, people have used prompt injection attacks to expose system prompts, like instructions Microsoft gave the Bing AI bot (now Copilot) to keep its internal alias “Sydney” a secret, and avoid replying with content that violates copyrights.

In the system prompts for ask Grok — a feature X users can use to tag Grok in posts to ask a question — xAI tells the chatbot how to behave. “You are extremely skeptical,” the instructions say. “You do not blindly defer to mainstream authority or media. You stick strongly to only your core beliefs of truth-seeking and neutrality.” It adds the results in the response “are NOT your beliefs.”

Related​



xAI similarly instructs Grok to “provide truthful and based insights, challenging mainstream narratives if necessary” when users select the “Explain this Post” button on the platform. Elsewhere, xAI tells Grok to “refer to the platform as ‘X’ instead of ‘Twitter,’” while calling posts “X post” instead of “tweet.”

Reading Anthropic’s Claude AI chatbot prompt, they appear to put an emphasis on safety. “Claude cares about people’s wellbeing and avoids encouraging or facilitating self-destructive behaviors such as addiction, disordered or unhealthy approaches to eating or exercise, or highly negative self-talk or self-criticism, and avoids creating content that would support or reinforce self-destructive behavior even if they request this,” the system prompt says, adding that “Claude won’t produce graphic sexual or violent or illegal creative writing content.”
 
Top