We are under construction, available fully functional from Q2 2026
Explainer8 min readTier 2

How AI Creates Its Own Reality—And Traps Us in It

Introduction: The Algorithm's Funhouse Mirror

In 2016, the city of Oakland deployed an AI system called PredPol to predict where crime would most likely occur. Police patrols were dispatched to these predicted hotspots, and just as the algorithm suggested, they found more crime there. The system appeared to be a success.

But there was a paradox at the heart of the operation. Police found more crime because they were sent to look for it in specific places. The AI wasn't necessarily predicting crime; it was predicting policing. This led to a feedback loop where patrol intensity in some predominantly Black neighborhoods increased by 400%, creating the very data that seemed to justify the initial prediction.

This phenomenon is called a feedback loop: a cycle where an AI's predictions change the world in a way that confirms the original prediction. These loops are not rare technical glitches; they are fundamental features of many AI systems operating today. This post will explore several surprising and impactful examples of these loops—systems that don't just find patterns, but actively create them—and what we can do to break them.

  1. When AI Turns Past Biases into Future Arrests

Predictive policing algorithms, a canonical example of feedback loops, are trained on historical crime data. Crucially, this data does not represent an objective map of all crime. Instead, it reflects historical policing decisions—where officers were already patrolling and making arrests.

This creates a self-perpetuating cycle:

  • Biased historical data teaches the model to associate certain neighborhoods, often minority communities, with crime.
  • The model predicts future crime will occur in those same neighborhoods.
  • Police are dispatched to these areas, leading to more surveillance and more arrests for minor offenses.
  • This new arrest data is fed back into the model as "proof" of its accuracy, making it even more confident in its biased prediction.

The algorithm wasn't predicting crime. It was predicting policing. And by directing more police to certain areas, it created the very patterns it claimed to discover.

The real-world impact, as documented in cities like Oakland and Chicago, includes a deterioration of community trust and the diversion of resources from community investment, all without a demonstrated reduction in serious crime. This dynamic of compounding historical bias in the streets is mirrored in the digital world, where a similar process is reshaping our inner worlds through recommendation engines.

  1. The Content Recommendation Spirals That Push You to Extremes

Recommendation algorithms on platforms like YouTube are designed for one primary goal: to keep you watching. They do this by showing you content similar to what you’ve already engaged with. While this seems harmless, this engagement optimization process naturally narrows a user's exposure over time, creating a powerful feedback loop.

Research has shown how this can create radicalization pathways. An algorithm can gradually shift a user from mainstream content to more extreme channels simply because extreme content is often highly engaging. The system isn't ideologically motivated; it's just following the data trail of user engagement to its logical, and often dangerous, conclusion—once again creating a reality (a user's belief system) rather than simply reflecting it.

The problem isn't that the algorithm is "biased" in a human sense; it's that the algorithm is ruthlessly effective at optimizing for what it can measure (engagement), and what is most engaging is not always what is good for users or society.

This process contributes directly to the formation of filter bubbles and echo chambers, eroding a shared sense of reality and amplifying societal polarization. While these spirals of belief shape our minds, another set of loops acts as gatekeepers to our financial and professional lives.

  1. How AI Gatekeepers Create Permanent Disadvantage

AI is increasingly used to make critical decisions about who gets a loan or a job interview. But when trained on historical data, these systems can create loops that lock people out of opportunities permanently.

Consider the "Thin File" Problem. A young person or recent immigrant with no credit history is often classified as risky by a scoring algorithm, simply because it has no data. This isn't a neutral starting point; it's a perpetuation of historical inequities like redlining, which systematically prevented entire communities from building a credit history. They are denied the credit they need to build one, so the initial lack of data becomes a permanent barrier—a catch-22 where the only way to prove you are creditworthy is to get credit you are algorithmically denied.

A similar loop emerges in hiring, leading to Cultural Fit Amplification. When Amazon trained an AI recruiting tool on a decade of its hiring data, the model learned that most successful applicants were men and began penalizing resumes with female-associated keywords. The system wasn't selecting for skill; it was amplifying the company's existing monoculture, encoding homogeneity as a proxy for success.

An algorithm designed to predict future success can inadvertently become a system that only rewards past privilege, effectively encoding homogeneity as the primary criterion for opportunity.

These gatekeeping loops create self-fulfilling prophecies for opportunity, but nowhere are the stakes higher than in healthcare, where a prediction can literally become a patient's destiny.

  1. The Medical Prediction That Can Make Itself True

In healthcare, a prediction from an AI can directly influence a patient's outcome, creating some of the most troubling feedback loops. The system's logic can manifest as a self-fulfilling prophecy in two distinct ways.

First is the loop of Risk Score Destiny. A patient flagged by an algorithm as "high-risk" for a complication will likely receive more intensive monitoring. This extra attention naturally leads to more complications being found and recorded, which in turn confirms and solidifies their high-risk status. Conversely, a patient deemed "low-risk" receives less monitoring. A potential complication might be missed, meaning their score stays low and reinforces the decision for less monitoring.

Second, and more dangerously, is the Resource Allocation Loop. When an algorithm predicts a patient is unlikely to benefit from an intervention, resources may be directed elsewhere. The patient receives less care and consequently has a poor outcome, which "confirms" the model's initial prediction. This is especially chilling in palliative care, where we face the problem of the unknown counterfactual—the outcome that would have occurred if a different decision had been made.

When an algorithm predicts a patient will have a poor outcome and care is reduced as a result, the prediction is "validated." But we can never know if the outcome would have been different if the AI hadn't made the prediction in the first place.

This dynamic can lead to a system where resources are systematically denied to certain patients, with the data generated from those very decisions being used to justify the denial in the future.

So, How Do We Break the Loop?

These feedback loops are a sobering feature of our algorithmic world, but they are not unbreakable. By acknowledging that AI actively shapes reality, we can design systems that are more fair and effective. Key strategies include:

  • Introduce Randomness: AI systems need to explore as well as exploit. By occasionally overriding a recommendation—for example, by approving a borderline loan application—the system can gather new data and learn from outcomes it would have otherwise missed. This is known as the "explore-exploit" tradeoff.
  • Demand Human Oversight: For high-stakes decisions, a human must remain in the loop with the authority to override the AI's suggestion. AI should serve as a tool for experts, not a replacement for them.
  • Audit for Outcomes, Not Just Accuracy: We must look beyond whether a prediction was "correct" and audit its real-world impact. This focus on fairness is why cities like Santa Cruz and New Orleans have rolled back their use of predictive policing.
  • Build in Fairness Constraints: Models can be explicitly programmed to penalize discriminatory outcomes during training. This acts as a brake, preventing the unchecked amplification of historical bias.
  • Implement Sunset Clauses: An AI model should not run forever. Requiring models to be periodically retired and rebuilt from scratch with new data and assumptions prevents them from endlessly reinforcing their own outdated logic.

Conclusion: Designing a Better Reality

AI systems are not neutral observers of our world; they are active participants in shaping it. The feedback loops they create are a critical, and often invisible, feature of modern life, influencing everything from our newsfeeds to our financial opportunities.

The systemic dangers are profound. The dynamics of the attention economy create a race to the bottom, as algorithms reward increasingly extreme content to maximize engagement. This erodes our democratic foundations by fracturing a shared factual basis for society. And when entire industries adopt the same flawed logic, we risk creating an "algorithmic monoculture" that systematically stratifies society along AI-defined lines, with no escape from its judgment.

As these systems become more powerful, the most important question isn't "Is the AI accurate?" but "Is the world it's building fair, equitable, and one we actually want to live in?"

This educational content was created with the assistance of AI tools including Claude, Gemini, and NotebookLM.