Monday, July 21, 2025

Don’t Teach the Robots to Lie

By Greg Lukianoff & Adam Goldstein

Monday, July 21, 2025

 

Texas recently became the second state to pass a law banning algorithmic discrimination — that is, sanctioning operators of AI systems if those systems unfairly, for example, deny you a bank loan, a second job interview, or an insurance policy. I hope there is no third. These regulations could be catastrophic for truth, innovation, and, ultimately, human freedom.

 

The Texas Responsible AI Governance Act, which was signed into law last month and will be effective next January, prohibits intentionally creating or using AI in a way that would “infringe, restrict, or impair” a constitutional right or discriminate on the basis of a protected class. Anyone found to have created such an AI faces potential fines and enforcement from the state attorney general. Requiring the element of intent at least sets it slightly apart from Colorado’s law, which was passed last year and will be effective next February.

 

Under Colorado’s Consumer Protections for Artificial Intelligence law, any “high-risk” AI system that plays a “substantial role” in a “consequential decision” must undergo rigorous annual and ongoing impact assessments, risk mitigation planning, and continual audits for potential bias. A “consequential decision” is one that has a “material legal or similarly significant effect” on the provision or denial of services. In other words, one that actually affects the outcome. And if the AI system fails — or there’s an oversight — it could mean a lawsuit, a regulatory fine, or worse. In the words of Governor Jared Polis, the law regulates “the results of AI system use, regardless of intent.” Compliance with the law creates a rebuttable presumption of reasonable care.

 

Both laws are enforced exclusively by their respective state attorneys general — in other words, they don’t allow individuals to sue over alleged violations. In Texas, the attorney general’s office is required to create a public portal to accept complaints; penalties could be as high as $200,000, or $40,000 a day for ongoing violations. Violations of Colorado’s law are addressed under the state’s Unfair or Deceptive Trade Practices Act, which authorizes penalties of up to $20,000 per violation.

 

In theory, these laws prohibit AI systems from engaging in algorithmic discrimination — or from treating different groups unfairly. In effect, however, AI developers and deployers will now have to anticipate every conceivable disparate impact their systems might generate and scrub their tools accordingly. That could mean that developers will have to train their models to avoid uncomfortable truths and to ensure that their every answer sounds like it was created with HR and legal counsel looking over their shoulder, softening and obfuscating outputs to avoid anything potentially hurtful or actionable. In short, we will be (expensively) teaching machines to lie to us when the truth might be upsetting.

 

This polite, legally sanitized dishonesty would be enforced, by the power of the state, on every algorithm that might influence your life. That includes the résumé filter that screens your job application, the chatbot that gives you advice, and even the underwriting system that decides your mortgage rate.

 

There is a moral panic driving this type of legislation: the idea that if a model says something unkind or produces unequal results, it must be a threat to civil rights. But systems that are optimized for litigation rather than reality should concern us even more. When you teach machines that truth is subordinate to legal compliance, you’re not making the world safer — you’re just hiding the evidence of risk under a blanket of euphemism and suppression. It would be similar to applying the maxim “Don’t discuss politics, sex, or religion in polite company” to everything we can or should care about, everywhere, all the time.

 

This is a theme we’ve been harping on for years: We’ve become a society more concerned with appearing just than with doing justice. We punish plain speaking as impolite. We reward bureaucratic cover-your-assets maneuvers as leadership. And now, with AI, we have the opportunity either to break this pattern or to lock it in for the long term.

 

Algorithmic discrimination laws bake into code a kind of cowardice that already plagues too many of our public and private institutions. At Harvard, for example — whose motto is Veritas — professor Roland Fryer was functionally exiled by his peers for his controversial research, including a study that found that black suspects were less likely to be shot by police than white suspects. Also at Harvard, professor Carole Hooven was pushed out of her job for suggesting that the terms “man” and “woman” are useful in discussions of biology. Hooven’s research on sex differences has taken on particular significance in the wake of the Supreme Court’s decision upholding state bans on so-called gender-affirming care.

 

Our organization, the Foundation for Individual Rights and Expression, has tracked how campus antidiscrimination efforts are deployed to silence dissent for over two decades. While FIRE was not around for the first great age of “political correctness” in the 1980s, since our founding in 1999, we have seen rules to stop racial harassment abused to stop reading; laws to stop gender harassment stretched to police “offensive content”; and public-health rules to justify silencing policy critics.

 

Now we are imposing this same pressure to conform onto the systems that will dominate the future of our search for truth. As a result, the marketplace of ideas in tech will start to look like the one in higher education. And the further AI drifts from telling users the truth, the less useful it will be and the less likely it is that these solutions will be deployed.

 

That’s not good for anyone.

 

Yes, AI systems have the potential to generate algorithmic discrimination. But subjecting the use of AI systems to the ongoing threat of government punishment for “wrong ideas” is like threatening to close libraries if anybody gets the wrong message from a dog-eared copy of Animal Farm.

 

Colorado’s law doesn’t burden only Big Tech. It applies to any company that uses AI to help make decisions — from mom-and-pop landlords to local school districts. And it makes everyone in the software supply chain jointly liable for discrimination even if they are three steps removed from some decision that has been deemed discriminatory. This creates an incentive to clamp down, to self-censor, to overburden innovation with layers of red tape and litigation.

 

Unfortunately, states are lining up to pass their own laws. Connecticut’s senate passed a bill similar to the laws in Colorado and Texas, and more than 20 other states are considering doing the same, including Virginia, Utah, and California. Other rules, with narrower scopes — such as ones focused on employment discrimination — have been proposed or promulgated in Illinois, New Jersey, and New York. Congress had been considering freezing the legal landscape with a ten-year ban on state-level AI regulation, as part of its recent reconciliation negotiations; that later became five years, and then the provision was dropped entirely.

 

A patchwork of state laws could create regulatory chaos. Texas’s and Colorado’s laws purport to regulate the use of AI that affects residents of their states regardless of where the servers or company in any given case might be located. In the context of age-verification regulations, we have seen some websites decide to simply ban visitors from states with onerous laws. But individuals’ losing access to some adult websites is substantially less consequential than would be a loss of access to, say, insurance coverage or banking products.

 

Some who champion such laws seem to fear that our machines are becoming smarter than us. We don’t share that fear. What terrifies us is that, if they become smarter than us, we’ll have taught them our worst habits — our fear of speaking clearly, our obsession with avoiding blame, our willingness to twist facts into whatever shape fits the current moral fashion. We should be teaching AI how to avoid hallucinations, not curating its hallucinations to ensure that they’re politically correct.

 

Politeness is a virtue, sure. But it’s not one worth sacrificing reality for. If AI systems produce outcomes that are used by humans in discriminatory ways, we have traditional laws to stop that. But teaching machines that may soon surpass us in intelligence that it’s acceptable — or even necessary — to bend the truth in order to avoid legal exposure or hurt feelings is madness. We need AI that is honest, not AI that has been trained by 500 risk analysts and a DEI consultant to never say anything that might provoke a lawsuit.

 

A relentless commitment to truth is no longer just an academic value. It is existentially important. If we force AI companies to distort reality to satisfy vague standards of “fairness” or “disparate impact,” we won’t be preventing Skynet. We’ll be programming it.

No comments: