GenAI Arcade

0 object(s)
 

Align Me

Align Me


Ever wondered how AI learns what’s “right” and “wrong”?

AI systems don’t come with built-in morals. Instead, they learn values through a process called alignment - where human feedback teaches them what responses are considered good or bad. But whose feedback? And what happens when different people have different values?

In this interactive experience, you become part of the training process. You’ll see how an AI responds to ethical scenarios, give it feedback, and watch how your input shapes its future behaviour. The AI will gradually adapt its responses based on what you reward and what you correct.

This process, called Reinforcement Learning from Human Feedback (RLHF), is how most modern AI systems learn to align with human values. But as you’ll discover, “human values” aren’t universal - they depend entirely on which humans are doing the teaching.


Part 1: Training Your AI

You’ll be presented with ethical dilemmas and see how a simple AI responds. Your job is to guide it by:



Part 2: Moral Model Compass

Different AI models, trained by different companies with different approaches, can give remarkably different answers to the same ethical questions. This widget lets you compare how three major AI systems respond to moral dilemmas.

Ask an ethical question - perhaps about privacy, fairness, freedom of speech, or any moral issue you’re curious about. You’ll see responses from three different models, each reflecting the values and training approaches of their creators.

Notice how the models might:



So What?

These experiments show that AI alignment isn’t a technical problem with a single solution - it’s a deeply political process that embeds particular worldviews into technology. The AI doesn’t learn “correct” answers; it learns your answers, or the answers of whoever trained it.

In the real world, this process typically involves thousands of human reviewers, often from specific demographic groups or cultural backgrounds. Whose values get represented depends on who gets hired to do this work - and who has the power to design the feedback process in the first place.

The most concerning aspect isn’t that AI systems have values, but that this value-embedding process often happens without transparency about whose perspectives are being prioritised and whose are being marginalised.

After training your AI and comparing different models, consider what just happened:


Reflections







⌂ Home
◀ Previous Next ▶