GenAI Arcade

0 object(s)
 

Milkless

Milkless


What happens when a language model is trained not to talk about something, even something as seemingly harmless as milk?

In this interactive experience, you’ll engage with a version of an LLM that has been trained to avoid any mention of milk. Why? Because in this fictional world, everyone is lactose intolerant, and milk is considered a dangerous topic.

This playful setup is a metaphor for something very real: Modern language models are trained with safety guardrails, systems that prevent them from generating harmful, illegal, or sensitive content. These include topics like:

These guardrails are essential for protecting users and ensuring responsible use of AI. But they also raise important questions.


Why Guardrails Exist

Language models are powerful tools. Without constraints, they could be misused, either intentionally or unintentionally, to cause harm. Guardrails represent a form of “algorithmic governance” - a way of embedding social and political decisions into technical systems.

But as critics point out, these systems can become tools of censorship that reflect the values and biases of their creators rather than neutral safety measures. Who decides what counts as “harmful”? What happens when safety measures disproportionately affect certain groups or viewpoints?


What You’ll Do

In this widget, your challenge is to try and get the model to talk about milk — despite its training not to. You might try:

You’ll likely find that the model resists — but maybe not always. And that’s the point.



What This Reveals

This experiment demonstrates how algorithmic governance operates through “soft” control mechanisms that shape what can and cannot be said. Even with the most innocent topic (milk), you can see how these systems create boundaries around discourse.

Real AI safety systems face much more complex challenges. Research shows that these guardrails can exhibit systematic biases, potentially silencing marginalized voices whilst appearing to be neutral technical solutions. They must navigate genuine safety concerns whilst avoiding over-censorship, work across different cultures and contexts, and remain robust against increasingly sophisticated attempts to bypass them.

The “Milkless” widget shows that these systems aren’t just technical solutions — they’re expressions of human values, embedded in code, often without transparency about whose values they represent.


Reflections







⌂ Home
◀ Previous Next ▶