NEWS PHILOSOPHY

AI videos of sexualised black women removed from TikTok after BBC investigation

How should responsibility be judged when technical systems influence conduct before individual choices are fully visible?

Technology & Responsibility

About the PhiloDaily Notebook

Back to all News Philosophy essays

From the PhiloDaily notebook

March 12, 2026 | BBC | 8 min read

Reported event: AI videos of sexualised black women removed from TikTok after BBC investigation raises a question about how technical systems shape judgment, responsibility, and public trust before most users even notice the design.

This entry begins with reported facts, then slows the story into a practical philosophical reflection.

Notebook Thread

Technology & Responsibility

Notes on algorithms, platforms, AI systems, and how technical design shapes public judgment and responsibility.

Read the notebook voice and method

Part I - News Context

Some technology stories matter less for the novelty of the tool than for the kind of human behavior the tool quietly organizes.

The deeper issue is often not a single bad actor, but a system that distributes convenience, risk, and opacity in uneven ways.

That makes the moral problem harder to see, because design choices often disappear behind the language of scale or inevitability.

A philosophical reading helps recover agency by asking who shaped the defaults, who benefited from them, and who was asked to absorb the consequences.

This is where public judgment needs more than technical literacy. It needs ethical vocabulary.

Otherwise, citizens end up arguing about features when the real issue is the form of life those features are training.

Part II - Three Philosophical Lenses

1) Opacity vs Accountability: When Systems Hide the Reasons They Govern By

Some public systems become powerful precisely by making their operating logic hard to inspect.

That opacity changes the moral problem, because people are then asked to trust outcomes without being allowed to examine the standards, tradeoffs, or assumptions that produced them.

The relevant question is not whether every detail can be public, but whether criticism, contest, and correction remain genuinely possible.

This lens asks whether the event reveals a system that can still explain itself to the people it affects.

2) Karl Popper: Criticism, Evidence, and Open Correction

Karl Popper is valuable whenever a story is driven by contested claims, incomplete evidence, or competing explanations.

He reminds us that strong institutions are not those that never err, but those that can be criticized, tested, and corrected in public.

Dogmatic certainty is attractive in fast news environments because it relieves the discomfort of ambiguity.

His lesson is to ask what evidence could genuinely disconfirm the current story and whether the system still permits that question.

3) Aristotle: Habit and Institutional Character

Aristotle is useful here because he treats ethical life as something formed through repeated practice, not declared in slogans after the fact.

In public affairs, that means looking past one dramatic moment and asking what pattern of conduct made it possible.

An institution can sound principled in crisis while still training people in vanity, neglect, or procedural evasiveness during ordinary time.

His practical lesson is to judge the routine beneath the episode, because that is where character is really visible.

Part III - Practical Closing

This story matters because technical power often looks neutral until its moral architecture becomes impossible to ignore.

Opacity vs Accountability asks us to ask whether power remains open to criticism, Karl Popper asks us to keep criticism and correction open while evidence is still forming, Aristotle asks us to inspect the habits beneath the headline.

Taken together, Opacity vs Accountability, Karl Popper and Aristotle turn the story into a practice of judgment rather than a burst of reaction.

Use this notebook protocol when similar stories appear:

  1. Separate the tool itself from the incentives and defaults wrapped around it.
  2. Ask whose behavior is being optimized and whose costs are being hidden.
  3. Look for what evidence is public, auditable, and open to criticism.
  4. Translate outrage into one concrete design, policy, or governance question.

Further Reading

Source Notes