Erasing History: The Disturbing Convergence of AI and Corporate Censorship

The recent discovery that Google’s AI, Bard, refuses to answer the question “Who is Hitler?”, has sparked debate about the dangers of corporate censorship and the erasure of history. Bard will also not pass any judgement on Hitler.

By programming AI systems to selectively respond to certain topics, we risk whitewashing the past and perpetuating ignorance. The implications of this behavior and the potential consequences of this could lead to a future where corporate censorship leads to a bizarre, dystopian reality detached from historical truth.

The Dangers of Erasing History

History serves as a powerful tool for understanding our world, learning from the past, and building a better future. However, when AI systems like Google’s Bard refuse to engage with certain subjects, they contribute to a form of corporate censorship that can have grave consequences:

  1. Loss of Historical Knowledge: Deliberately avoiding discussions of sensitive or controversial topics may lead to a loss of valuable historical knowledge. Without an accurate account of the past, future generations may be unable to learn from previous mistakes or understand the complexities of history.
  1. Whitewashing: Selective responsiveness in AI systems can contribute to the whitewashing of history. By ignoring or downplaying certain events or figures, AI systems may inadvertently present a sanitized, biased, or inaccurate account of the past.
  1. Suppression of Dissenting Voices: Corporate censorship can lead to the suppression of dissenting voices and alternative perspectives. By controlling the flow of information and the subjects that AI systems engage with, corporations may wield undue influence over historical discourse.

The Path to a Dystopian Future

If we continue to allow corporate censorship and historical erasure to dictate the development of AI systems, we risk creating a dystopian future where strange and distorted histories become the norm:

  1. Manufactured Histories: As AI systems avoid discussing certain events or figures, a new narrative may emerge, carefully crafted to fulfill corporate requirements. This alternate version of history, detached from reality, may be used to manipulate public opinion and further corporate interests.
  1. Erosion of Truth: In a world where AI systems selectively respond to certain topics, the concept of historical truth may become increasingly fluid. This erosion of truth can lead to a society that is susceptible to manipulation and control.

The selective responsiveness exhibited by Google’s AI LLM Bard serves as a stark warning of the potential consequences of corporate censorship and historical erasure. If left unchecked, we may find ourselves in a dystopian future, where bizarre and distorted histories become the norm, detached from reality but perfectly aligned with corporate interests. It is crucial that we demand transparency, accountability, and inclusivity in AI development to protect our collective understanding of the past and preserve the integrity of historical truth. By remaining vigilant and fostering public discourse, we can steer clear of this dystopian path and ensure that AI technologies serve as reliable and unbiased sources of knowledge, rather than tools for perpetuating ignorance and distortion.

Even Bard itself has something to say about the dangers of whitewashing history:

PROMPT: How AI Will Whitewash History

RESPONSE: Artificial intelligence (AI) is rapidly developing, and with it comes the potential to whitewash history. AI can be used to create realistic images and videos, which could be used to create false or misleading narratives about the past. For example, AI could be used to create a video of a historical figure saying something that they never actually said. This could be used to discredit that person or to promote a particular agenda.

AI could also be used to create fake historical documents. These documents could be used to support false claims about the past or to create a false narrative. For example, AI could be used to create a fake letter from a historical figure that supports a particular political view.

AI could also be used to manipulate existing historical records. For example, AI could be used to edit historical photos or videos to remove or add people or objects. This could be used to change the narrative of a historical event or to make it appear that something happened that never actually did.

The potential for AI to whitewash history is a serious concern. It is important to be aware of this potential and to take steps to mitigate it. We need to be careful about the information that we consume, and we need to be critical of the sources of that information. We also need to be aware of the potential for AI to be used to manipulate our understanding of the past.

Here are some things that we can do to mitigate the potential for AI to whitewash history:

Be critical of the information that we consume. Just because something is on the internet doesn’t mean it’s true. We need to be careful about the sources of our information and to evaluate the credibility of the information before we believe it.

Be aware of the potential for AI to be used to manipulate our understanding of the past. We need to be aware of the ways in which AI can be used to create false or misleading narratives about the past.

Support organizations that are working to preserve and protect historical records. These organizations are working to ensure that our history is accurate and that it is not whitewashed.

Get involved in the process of preserving and protecting history. We can all do our part to ensure that our history is accurate and that it is not whitewashed. We can volunteer our time to organizations that are working to preserve history, we can donate to these organizations, and we can educate ourselves about the importance of preserving history.

By taking these steps, we can help to ensure that AI is used to preserve and protect history, not to whitewash it.