Republicans Are Slowly 'Learing' How to Fight the Democrats
2025 Did Not End Well for These Two Brothers in the NFL
Deputy HHS Secretary to Minnesota: 'We Have Turned Off the Money Spigot'
Trump’s Christmas Present: 4 Percent Growth
Doomed?
Wrong Predictions? Never Mind
The Economists Got 2025 All Wrong
Nobody Ever Gets Punished
As Pelosi Steps Away, the Press Keeps Pampering
Lessons to Learn From the Welfare Mega-Fraud Scandal in Minnesota
The Government Controls Too Much Land in the West
Iran's Real War Is Not With the West – It Is Against Its...
Somali Daycare Fraud Uncovered by Citizens
Tim Walz Says He Takes Fraud Seriously After Keith Ellison Vowed to Fight...
Another Leftist Judge Is Blocking Trump's Deportations
OPINION

Can a Trusted AI System Be Developed?

The opinions expressed by columnists are their own and do not necessarily represent the views of Townhall.com.
AP Photo/Michael Dwyer

Artificial intelligence is changing everything from how we search for information to how we make decisions. But with that transformation comes a troubling question: Can we trust the answers and conclusions AI offers?

Advertisement

There are good reasons to suggest we cannot. Earlier this year, major newspapers published a list of recommended “beach reads” for the summer, complete with catchy summaries and clever author bios. The only problem? Every single book on the list was made up.

A few months before that, a New York lawyer was sanctioned for filing a legal brief full of citations to court cases that did not exist. The culprit wasn’t laziness or malice, but rather it was ChatGPT, which had simply invented case law out of thin air.

These weren’t isolated incidents. From fake health advice to imaginary sources, AI systems have a growing reputation for making things up while sounding perfectly confident. This phenomenon, known as a “hallucination,” is more than an embarrassment; it’s a threat to credibility, law, and even democracy itself.

The problem begins with data. Artificial intelligence systems don’t “know” facts the way humans do; they detect patterns in the massive amounts of text, images, and numbers they’re fed. If the data is biased, outdated, or false, the system simply amplifies the problem at machine speed.

Tech companies like to call this an engineering challenge, but it’s really a civic one. As AI becomes part of journalism, education, government, and law, misinformation produced by algorithms can distort the very record we rely on to make collective decisions.

A handful of data reformers are trying to change the conversation from one about smarter algorithms to one that relies on verifiable information. Dr. Jose M. Plehn, who founded the company BrightQuery, is devoted to building a“factual infrastructure” for AI.

Advertisement

A “factual infrastructure” makes AI accountable by relying on real-world evidence. For example, BrightQuery collects legal, regulatory, and tax filings from more than 100,000 jurisdictions worldwide, then converts them into structured, searchable data. The result is an enormous economic map that ties millions of companies, individuals, and locations to verified public records.

This creates a system based on a reliable database of verified information. “Every claim made by a model should be traceable to a verifiable record,” says Plehn, who works with the federal National Secure Data Service (NSDS). U.S. government data is based on more accurate, transparent, and usable information across agencies.

Key to this approach is something called data provenance or knowing exactly where data originates, how it was processed, and under what authority it was published. Without it, even the most advanced AI operates “in epistemic fog.”

This concept is gaining traction in global standards like the FAIR principles (Findable, Accessible, Interoperable, Reusable) and in efforts by OASIS Open to establish common provenance rules. The goal is to make every dataset auditable and every AI-generated statement traceable to a credible source. “If you can’t trace the data,” Plehn explains, “you can’t trust the result.”

Plehn took his efforts even further by launching OpenData.org, a nonprofit that makes parts of BrightQuery’s massive database freely available to the public. Journalists, researchers, and policymakers can now access the same verified information that major institutions use. “Open data is just as vital to democracy as the right to free speech,” he said at the project’s debut.

Advertisement

Artificial intelligence will soon play a role in every major institution: law, finance, journalism, education, and beyond. If we don’t build systems that can prove their own accuracy, we risk a future in which truth itself becomes optional.

The challenge before us isn’t just to make AI more powerful. It’s to make it honest. And that starts not with algorithms, but with the data they depend on.

Until we demand verifiable, audit-able information like that championed by the growing factual-data movement, AI will continue to speak confidently while knowing very little. The question is whether we’ll keep listening.



Frank Salvato is an independent journalist and new media analyst who writes on public issues and emerging technology with a focus on accountability, legality, and the advancement of public knowledge.

Editor’s Note: Do you enjoy Townhall’s conservative reporting that takes on the radical left and woke media? Support our work so that we can continue to bring you the truth.

Join Townhall VIP and use the promo code FIGHT to get 60% off your VIP membership!

Join the conversation as a VIP Member

Recommended

Trending on Townhall Videos

Advertisement
Advertisement
Advertisement