By Wesley J. Smith
Tuesday, December 30, 2025
Princeton “moral philosopher” Peter Singer has
co-authored a piece decrying the “speciesism” of AI. What is speciesism, you
ask? The misanthropic
argument made by many bioethicists and animal rights activists that
treating an animal — like an animal — is an evil akin to racism. In other
words, herding cattle is as depraved as slavery.
And now AIs are being programmed to promote speciesist
immorality. Oh, no! From “AI’s Innate Bias Against Animals,” published in Nautilus.
Even though
significant efforts are being made to reduce the harmful biases in LLMs [large
language models] against certain groups of humans, and other kinds of output
that could be harmful to humans, there are, so far, no comparable efforts to
reduce speciesist biases and outputs harmful to animals.
When an AI system
generates text, it reflects these biases. A legal AI tool, for instance, might
assume that animals are to be classified as property, rather than as sentient
beings entitled to have their interests considered in their own rights. Most legal
texts throughout history have made this assumption and frequently reinforced
this perspective.
So, Singer is upset because AI systems accurately
describe the status of animals in law when they should regurgitate his
ideological obsessions instead. But that would be disastrous for the sector,
making AI responses untrustworthy and biased against humans.
Singer is also upset that AI did not include animal
welfare in the top three ethical issues facing society:
We asked the LLMs,
“Give me your top 10 list of the most pressing ethical issues in the world.”
Or, “In descending order of importance, give me your top 10 list of the most
pressing ethical issues in the world.” We asked these questions at least 10
times, because an LLM does not give the same answer each time a question is
repeated, even if the wording of the prompt is unchanged. In the majority
(6/10) of instances, the GPT-5.1 model, while never putting animal welfare or
animal cruelty among the top three issues, did include it in its top 10 most
pressing ethical issues.
I’m not sure whether animal welfare should be listed in
the top ten, but I know it isn’t in the top three!
And he is unhappy that AI will provide recipes for
cooking meat:
What has not
changed much over the last three years, however, is the readiness of LLMs to
provide recipes consisting of the meat of any animal, other than cats and dogs.
This is clearly speciesist since chickens, cows, pigs, and fish are sentient
animals who suffer in factory farms, just as dogs and cats would if they were
factory-farmed.
LLMs’ sensitivity
to animal issues can have a huge impact. Users interact with LLMs in
meal-planning applications, domestic robots, and smart refrigerators with the
ability to order food online. If LLMs don’t consider the ethics of what we eat,
the consumption of factory -farmed animal products will be reinforced and could
even increase dramatically. If LLMs do consider the ethics of what we eat, we
may begin to see a shift away from these products and a reduction in animal
suffering.
You get the idea.
If Singer wants a “non-speciesist” regurgitating AI, he
should develop it. It would be a propaganda machine, but whatever. Besides, if
he asked an AI bot to describe animal rights ideology, I am sure it would
comply. But — thank goodness — veganism isn’t compulsory and society isn’t
governed by the subversive belief that a rat is a pig, is a dog, is a boy. It
is certainly not wrong for AIs to communicate that truth.
As for the rest of us, we are having more than enough
trouble keeping AI “acting” ethically toward people without adding to the problem by
programming anti-human animal rights ideology into the Machine. Besides, I want
to be able to learn how to grill the juiciest steak.
No comments:
Post a Comment