One of the most powerful beliefs in the digital age is also one of the most dangerous:
Algorithms are neutral.
They are mathematical.
They are data-driven.
They are objective.
Or so the story goes.
In reality, neutrality is not a property of algorithms.
It is a story we tell about them — and that story quietly transfers authority from humans to systems we no longer question.
Artificial Intelligence as a Power System — Not a Tool

1. Why the Neutrality Myth Exists
The idea of neutral algorithms feels comforting.
If decisions are made by machines:
- Bias disappears
- Responsibility diffuses
- Conflict feels resolved
Numbers feel cleaner than people.
Code feels fairer than judgment.
Neutrality becomes a psychological shortcut — a way to avoid confronting power, values, and trade-offs.
But neutrality has never existed in decision-making.
Algorithms did not invent bias.
They formalized it.
2. Every Algorithm Is a Series of Choices
An algorithm is not a law of nature.
It is a chain of human decisions:
- What data to include
- What data to exclude
- What outcomes to optimize
- What errors are acceptable
- What trade-offs are tolerable
Each choice reflects:
- Priorities
- Assumptions
- Incentives
Neutrality would require the absence of choice.
Algorithms are nothing but choices, encoded.
3. Data Is Not Reality — It Is a Record of Power
Algorithms learn from data.
Data is often treated as a neutral reflection of the world.
It is not.
Data reflects:
- Historical inequalities
- Institutional behavior
- Cultural norms
- Economic incentives
If a system learns from the past, it inherits the past’s distortions.
Calling this “bias” understates the issue.
It is structural inheritance.

4. Optimization Is a Value Judgment
Algorithms optimize.
But optimization always answers a moral question:
What matters most?
Speed?
Accuracy?
Profit?
Engagement?
Risk reduction?
Choosing what to optimize means choosing what to sacrifice.
If an algorithm optimizes engagement, it may amplify outrage.
If it optimizes efficiency, it may marginalize edge cases.
If it optimizes profit, it may externalize harm.
Neutrality collapses the moment optimization begins.
Who Controls AI Models — Governments, Corporations, or No One?
5. Objectivity as Authority Theater
Algorithms often present outputs with confidence:
- Scores
- Rankings
- Predictions
- Recommendations
These outputs feel authoritative.
Not because they are correct —
but because they are quantified.
Numbers silence debate.
People hesitate to challenge systems that appear precise, even when:
- Inputs are flawed
- Context is missing
- Assumptions are hidden
Objectivity becomes a performance — not a guarantee.
6. Bias Is Not a Bug — It Is a Feature of Scale
At scale, systems must generalize.
Generalization requires:
- Simplification
- Categorization
- Averaging
This inevitably:
- Erases nuance
- Penalizes minorities
- Privileges the statistically “normal”
Bias is not an accident of algorithms.
It is the cost of scaling decisions.
The question is not whether algorithms are biased.
It is whose bias is scaled.
7. Neutrality as a Shield Against Accountability
Labeling algorithms as neutral has a convenient side effect:
No one is responsible.
When harm occurs:
- Developers blame data
- Companies blame models
- Institutions blame technology
The system becomes the scapegoat.
Neutrality narratives protect power by dissolving accountability.
If no one chose, no one can be blamed.
8. Why “Fixing Bias” Misses the Point
Many efforts focus on:
- Debiasing data
- Auditing models
- Improving fairness metrics
These matter.
But they often avoid the deeper issue:
Who decides what “fair” means?
Fairness is not a technical parameter.
It is a social agreement.
You cannot engineer consensus.
You can only encode priorities.
The Alignment Problem Is Not Technical — It’s Political
9. The Political Function of Neutral Algorithms
Neutral algorithms do more than process information.
They:
- Legitimize decisions
- Depoliticize conflict
- Naturalize outcomes
If an algorithm denies a loan, flags a risk, or ranks a candidate, the outcome feels inevitable.
Politics disappears behind math.
This is not neutrality.
It is governance by disguise.
Why Regulation Will Always Be Late
10. What Real Honesty Looks Like
An honest approach to algorithms would admit:
- They embed values
- They reflect power structures
- They enforce priorities
Instead of asking:
“Is this algorithm neutral?”
We should ask:
“Whose values does it enforce — and who bears the cost?”
That question cannot be answered by code alone.
Closing Thought
Algorithms are not neutral.
They never were.
They are mirrors of the systems that build them —
and amplifiers of the values those systems prefer not to defend openly.
When neutrality is claimed, power is usually hiding.
Understanding that is the first step toward real accountability in the age of artificial intelligence.