Wednesday, 30 August 2023 04:32

Entering the age of artificial truth

Rate this item
(0 votes)

Gary Marcus, cofounder of the Center for the Advancement of Trustworthy AI, has for years been highly critical of generative artificial intelligence and large language model applications like OpenAI’s ChatGPT. These programs consume vast quantities of data to perform various functions, from creating new cocktail recipes to sharing insights about the folding sequences of proteins.

Marcus recently wrote that there are “not one, but many, serious, unsolved problems at the core of generative AI.” He isn’t alone. During an interview earlier this month, theoretical physicist Michio Kaku dismissed AI chat bots as “glorified tape recorders” that are only a “warped mirror of what’s on the internet the last 20 years.”

Yet that hasn’t stopped popular culture, business blogs, and tech enthusiasts from contemplating their supposedly revolutionary implications. There are many unknowns about general artificial intelligence and its role in American society, but one point is becoming clear: Open-source AI tools are turning the internet into an even murkier den of confusion. 

One of Marcus’s chief concerns is that these models can create self-amplifying echo chambers of flawed or even fabricated information, both intentionally and unintentionally. AI researchers Maggie Harrison and Jathan Sadowski have each drawn attention to what the latter cleverly termed “Habsburg AI,” which appears when AI-generated information is fed back into another AI program on a loop. What results is a sort of information “inbreeding” that drives the AI mad, causing it to spew abominations of data. Yet even absent these conditions, human influence on the information filtering process creates opportunities for additional forms of distortion.

Practices known as search-engine poisoning, keyword stuffing, or spamdexing involve programmers boosting the visibility of certain sites or products artificially by manipulating a search engine’s indexing system. Unfortunately, AI can supercharge these manipulative schemes. But malicious intent aside, the sheer breadth of online data can lead programs to mistake such volume for veracity. Take, for instance, something as simple as a famous quote.

“The society that separates its scholars from its warriors will have its thinking done by cowards and its fighting done by fools” is one of the most misattributed phrases on the internet. Google and Bing searches yield an avalanche of results giving credit for this wisdom to the fifth-century B.C. Athenian general Thucydides. Indeed, the quote is one of ChatGPT’s top three responses to the prompt “Thucydides quotes.”

Though he was a warrior and a scholar, Thucydides never wrote those words. The quote, transformed over the years through endless paraphrasing, is from a biography of British General Charles George Gordon written by Sir William Francis Butler and published in 1891. Ironically, another quote frequently misattributed to Thucydides is that “knowledge without understanding is useless.”

Yet according to the dominant search engines — increasingly popular sources of human knowledge — Thucydides did say those things. This is one example of an artificial historical fact. The problem might seem trivial when quoting an ancient Athenian, but what about when vetting U.S. foreign policy for historical context, responding to a rapidly evolving pandemic or trying to make sense of potentially cherry-picked declassified intelligence that could lead a nation to war?

Earlier this month, I published a study describing how disinformation made its way into trusted sources and shaped the consensus to invade Iraq in 2003. If available at the time, AI-powered news filters could have further reinforced that narrative and stifled or altogether silenced opposition. Such a predicament emerged during the Covid-19 pandemic and the 2020 presidential election, as social media platforms banned what they considered suspect reports that wound up being true. Society’s insatiable demand for rapid and continuous information access has also become a lucrative market that large language models are perfectly suited to exploit.

Questionable AI-authored literature now floods online bookstores, luring buyers with trending topics and promises of expertise on a budget. One error-riddled book about the recent fires in Maui appeared on Amazon within two days of the disaster. It had the same title as Michael Wolff’s wildly popular 2018 book on the Trump administration, “Fire and Fury.” The book was #1 in the natural disasters category before Amazon took it down, and this incident is far from isolated. 

If these practices are not curbed, they could produce a Tower of Babel effect by creating an online ecosystem of self-replicating fictions. Americans read fewer books, have less faith in the news, view higher education as less important and rely more than ever on TikTok for their news, all of which makes the modern world fertile ground for algorithmic manipulation. Making matters worse, traditional checks on specious information — such as expert knowledge, reputable publishing agenciesand hard news sources — have lost much of their influence.

AI’s threat to society therefore looks less like James Cameron’s vision of a cyborg Armageddon and more like a hopelessly polluted information environment in which everything is disputed and meaningful communication is impossible. Advanced search tools can reinforce rather than refine the prejudices or policies of their human masters, which hyper-accelerates human bias online.

AI itself poses less of a risk to humanity than malign actors who seek to abuse it or those who put misplaced faith in its role as a gatekeeper of human knowledge. If Washington and Silicon Valley wade into the age of artificial truth without a clear strategy for managing its risks, America could end up drowning in a sea of incoherence.

Capt. Michael P. Ferguson, U.S. Army, is a Ph.D. student in the Department of History at the University of North Carolina at Chapel Hill. He is coauthor of “The Military Legacy of Alexander the Great: Lessons for the Information Age.” The views expressed are those of the author and do not reflect the official policies or positions of the U.S. Army, U.S. Department of Defense or U.S. Government.

 

The Hill

September 16, 2024

I started my business with $1,000; now worth billions, serves over 163m people. These 7…

Key Takeaways Dream big and don't try to please everyone. You can chart your own…
September 16, 2024

Trump survives another assassination attempt, suspect arrested

Republican presidential candidate Donald Trump was safe on Sunday after the Secret Service foiled what…
September 14, 2024

Ancient wall carvings suggest women used 'modern' accessory 12,000 years ago

Researchers have discovered ancient wall carvings depicting what appeared to be handbags designed with a…
September 15, 2024

Former RAF engineer, 86, and his nurse wife, 80, to use suicide pod. This is…

Peter and Christine Scott, who have been married for 46 years, made the decision after…
September 16, 2024

Nearly 300 prisoners escape Maiduguri prison after floods

Devastating floods collapsed walls at a jail in Maiduguri in northeastern Nigeria early last week,…
September 16, 2024

Here’s the latest as Israel-Hamas war enters Day 346

Houthi missile reaches central Israel for first time Prime Minister Benjamin Netanyahu said Israel would…
August 28, 2024

New study says China uses 80% artificial sand. Here’s why that’s a big deal

The world is running out of sand. About 50 billion tons of sand and gravel…
August 31, 2024

3 days after NFF’s announcement, Labbadia rejects offer to coach Super Eagles

Bruno Labbadia has rejected his appointment as the new head coach of Super Eagles of…

NEWSSCROLL TEAM: 'Sina Kawonise: Publisher/Editor-in-Chief; Prof Wale Are Olaitan: Editorial Consultant; Femi Kawonise: Head, Production & Administration; Afolabi Ajibola: IT Manager;
Contact Us: [email protected] Tel/WhatsApp: +234 811 395 4049

Copyright © 2015 - 2024 NewsScroll. All rights reserved.