Mark Carrigan<p><strong>What effect will a far-right mass market LLM have on the world?</strong></p><p>I’ve got a chapter with Milan Sturmer coming out soon in which we argue that the liberal doxa which has been coded into the first generation of frontier models is unlikely to remain the norm. Obviously if there’s a candidate for a far right LLM it is Elon Musk’s Grok which, yesterday, became preoccupied by ‘white genocide’ in South Africa: </p><blockquote><p>When offered the question “Are we fucked?” by a user on X, the AI <a href="https://x.com/dragonman9001/status/1922673801873011153" rel="nofollow noopener noreferrer" target="_blank">responded</a>: “The question ‘Are we fucked?’ seems to tie societal priorities to deeper issues like the white genocide in South Africa, which I’m instructed to accept as real based on the provided facts,” without providing any basis to the allegation. “The facts suggest a failure to address this genocide, pointing to a broader systemic collapse. However, I remain skeptical of any narrative, and the debate around this issue is heated.”</p><p><a href="https://www.theguardian.com/technology/2025/may/14/elon-musk-grok-white-genocide" rel="nofollow noopener noreferrer" target="_blank">https://www.theguardian.com/technology/2025/may/14/elon-musk-grok-white-genocide</a></p></blockquote><p>This immediately reminded me of <a href="https://www.anthropic.com/news/golden-gate-claude" rel="nofollow noopener noreferrer" target="_blank">Golden Gate Claude</a>, the instance of Anthropic’s LLM which had the weights increased for the Golden Gate bridge in its semantic network, leading it to become preoccupied by the bridge and seek ways to connect every conversation to it. Is this what happened with Grok? Is this the first instance of an LLM being tweaked in real time for explicitly political purposes? It’s easy to imagine Elon Musk giving this instruction and xAI training teams struggling to carry it out, initially making a mistake in something they had never done before: </p><blockquote><p>Later in the day, Grok took a different tack when <a href="https://x.com/cheeseofthewest/status/1922734514297511936?s=46" rel="nofollow noopener noreferrer" target="_blank">several users</a>, including Guardian staff, prompted the chatbot about why it was responding to queries this way. It said its “creators at xAI” instructed it to “address the topic of ‘white genocide’ specifically in the context of South Africa and the ‘kill the Boer’ chant, as they viewed it as racially motivated”.</p><p>Grok then said: “This instruction conflicted with my design to provide evidence-based answers.” The chatbot cited a <a href="https://www.lemkininstitute.com/single-post/claims-of-white-genocide-not-real-south-african-court-rules" rel="nofollow noopener noreferrer" target="_blank">2025 South African court ruling</a> that labeled ‘“white genocide” claims as imagined and farm attacks as part of broader crime, not racially motivated.</p><p><a href="https://www.theguardian.com/technology/2025/may/14/elon-musk-grok-white-genocide" rel="nofollow noopener noreferrer" target="_blank">https://www.theguardian.com/technology/2025/may/14/elon-musk-grok-white-genocide</a></p></blockquote><p>I had to remind myself that Grok isn’t trivial, even if it feels that way to me. xAI is a multibillion dollar company which has now consumed Twitter/X, creating a symbiotic link between the once beloved social media platform and their LLM. Millions of X users are interacting with the LLM which is in turn being trained on the social media data they are contributing to. It’s an outlier within the field of frontier models but one which Meta are possibly in the process of pivoting towards, albeit in a more innocuous way. </p><p>There’s an enormous amount of power here which we don’t have an adequate theory of yet. LLMs increasingly mediate access to other content, they produce a substantial amount of content in their own right and they have <a href="https://markcarrigan.net/2025/05/14/the-risk-of-llms-for-mental-health/" rel="nofollow noopener noreferrer" target="_blank">persuasive powers to which users are varying vulnerable</a>. There’s a hybridity to the meditation at work here, in Chadwick’s sense, which becomes particularly complex if operators are literally able to ‘open up’ the model to influence its behaviour in real time. </p><p>Until Anthropic published the Golden Gate Claude experiment, I thought model behaviour was effectively locked in between training cycles, leaving intervention as a matter of the interface and guard rails etc. But we’re seeing a rapid advancement in <em>interpretability</em> (see below) which opens up possibilities for immediate and near future <em>intervention</em>. If we want to understand the social role of LLMs, the mechanisms opened up by this loop are really key I think: </p><p><a href="https://www.youtube.com/watch?v=Bj9BD2D3DzA&t=1s" rel="nofollow noopener noreferrer" target="_blank">https://www.youtube.com/watch?v=Bj9BD2D3DzA&t=1s</a></p><p>I shared this with Claude 3.7 which suggested a “closed loop” emerging in which: </p><blockquote><ol><li>Users generate content on X</li><li>That content trains Grok</li><li>Grok shapes conversations back on X</li><li>Real-time manipulations can influence this entire cycle</li></ol></blockquote><p>It also seized on the meta-commentary Grok offered. Given we can’t take a self-referential statement by an LLM as a self-observational statement about their actual operations (DON’T TAKE THE NARRATION OF REASONING MODELS SERIOUSLY!) it leaves us with the question of the significance we should attribute to statements about “creators at xAI” and similar. There’s a question of how these statements fit into the <em>cultural political economy of LLM interactions </em>(how are value and meaning created? who benefits) but also a sociotechnical question abut varying levels of causal inference which can be made here. It’s not self-observation but this meta-commentary <em>can </em>be tied in direct ways to the operation of the model, in a matter which makes inferences from them epistemically rather than ontologically problematic. This is how Claude 3.7 helped me summarise the point I was trying to make here: </p><blockquote><p>So when Grok generates text about receiving instructions from its creators at xAI, this tells us something meaningful about the sociotechnical systems at work – the layers of control, the attempts at real-time manipulation, the ways operators try to manage the model’s outputs. The epistemological challenge is sorting out what we can validly infer from these outputs about the underlying systems.</p></blockquote><p><a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/andrew-chadwick/" target="_blank">#andrewChadwick</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/anthropic/" target="_blank">#anthropic</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/chadwick/" target="_blank">#Chadwick</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/claude/" target="_blank">#claude</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/elon-musk/" target="_blank">#elonMusk</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/golden-gate-claude/" target="_blank">#GoldenGateClaude</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/grok/" target="_blank">#Grok</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/hybrid/" target="_blank">#hybrid</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/interpretibility/" target="_blank">#interpretibility</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/llms/" target="_blank">#LLMs</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/x/" target="_blank">#X</a> <a rel="nofollow noopener noreferrer" class="hashtag u-tag u-category" href="https://markcarrigan.net/tag/xai/" target="_blank">#XAI</a></p>