Balance scale on a wooden desk

AI ethics isn’t about whether robots should have rights. It’s about whether we’ll still have ours.

That distinction matters more than most people realize. The conversation around AI ethics has grown enormous — every conference has a panel, every company has a policy, every government has a committee. But the actual substance of the debate remains surprisingly thin. Most of it collapses into two modes: vague reassurance from the companies building AI, or abstract alarm from the academics watching them do it.

What’s missing is the middle ground where most of us live. We use AI. We benefit from it. We also feel something uncomfortable when we think about where it’s going. The question isn’t whether AI is good or bad. It’s whether the people making decisions about it — deploying it in hiring, in healthcare, in criminal sentencing, in creative work — are thinking carefully enough about what they’re doing. And whether we, as the people affected by those decisions, have enough knowledge to push back when they’re not.

These five books gave me that knowledge. They approach AI ethics from radically different angles — skepticism, philosophy, activism, safety research, and political economy — but together they form something close to a complete education. If you read all five, you’ll be better equipped than most policymakers to think about what AI should and shouldn’t be allowed to do.

I’ve ordered them by how directly useful they are to someone trying to form their own ethical framework — not just understand the debate, but actually decide where they stand.


01

AI Snake Oil: What Artificial Intelligence Can Do, What It Can’t, and How to Tell the Difference

Before you can think clearly about AI ethics, you need to know what AI actually does and what it doesn’t. That sounds obvious. It isn’t. The gap between what AI companies claim and what their products deliver is wide enough to drive policy through — and people do, constantly. Hiring algorithms that don’t predict job performance. Recidivism models that encode racial bias under the label of objectivity. Predictive systems that are, statistically, worse than a coin flip.

Narayanan is a Princeton computer scientist. Kapoor is his collaborator. Together they built a framework for sorting real AI capabilities from what they call “snake oil” — AI products that simply do not work as advertised. The framework is straightforward: some AI tasks (like language translation and image recognition) have clear benchmarks and genuine progress. Others (like predicting crime or screening resumes) have neither, and the companies selling them know it.

What makes this an ethics book and not just a tech debunking is the implication. If we’re making consequential decisions about people’s lives — their freedom, their employment, their healthcare — based on AI systems that don’t work, that isn’t a technical failure. It’s a moral one. The Princeton rigor here is exactly what the conversation needs: less hand-wringing, more evidence.

Read this if: you want the factual foundation that every ethical argument about AI should start with — and usually doesn’t.

02

The Last Skill: What AI Will Never Own

Full disclosure: I wrote this one. I’m including it because the Governance proof — Chapter 5 — is, at its core, an argument about ethics. And it’s an argument I haven’t found anywhere else.

The Last Skill identifies four proofs of human irreplaceability: Creativity (genuine novelty), Governance (choosing the value hierarchy), Decision-Making (absorbing the real downside of the cut you make), and Reputation (the externally verified trail of all three). Together they point to what the book calls “agency under consequence” — the capacity to be the one who answers for it.

Governance is the Second Proof, and it’s the one that bears most directly on ethics. The argument is specific: governance isn’t choosing what can exist. It’s choosing what should exist. It’s deciding the value hierarchy — which priorities come first when they conflict — and then absorbing the consequences of that choice. An AI can optimize within a value system you hand it. It cannot decide which values matter more than others. That decision requires something machines structurally lack: a stake in the outcome. A life that changes depending on what you choose.

This is why I think the AI ethics conversation keeps spinning without landing. We keep asking “how do we make AI ethical?” when the real question is “who bears the cost when the choice goes wrong?” Ethics requires agency under consequence. You can’t have moral responsibility without something at risk. Machines process. Humans absorb. That asymmetry is the entire foundation of ethical reasoning, and it doesn’t go away because the processing gets faster.

Read this if: you think AI ethics needs to be grounded in what humans structurally have that machines structurally don’t — and you want an argument that goes beyond “AI can’t feel.”

Available on Amazon Kindle →
03

The AI Con

Bender and Hanna are two of the most persistent critics of Big Tech’s AI narrative, and The AI Con is their full-length case against it. Their argument is that the dominant AI story — the one told by OpenAI, Google, Meta, and the rest — is a con in the classic sense: a confidence game designed to concentrate power while claiming to distribute it.

This is the angriest book on this list, and it earns its anger. Bender (a computational linguist at the University of Washington) was co-author of the “stochastic parrots” paper that contributed to the firing of Timnit Gebru from Google. Hanna left Google over the same controversy. They have skin in this fight, and it shows in the writing — which is sharp, occasionally biting, and always specific about who benefits and who pays.

Where AI Snake Oil gives you the technical receipts, The AI Con gives you the political ones. Who funds the research? Who sets the safety standards? Who decides what “aligned” means? The answers are uncomfortable, and Bender and Hanna refuse to let you look away from them.

I don’t agree with every argument here. Some of the skepticism tips into a wholesale rejection of the technology that I think throws out genuine progress with the corporate bathwater. But as a counterweight to the default optimism of Silicon Valley, this book is necessary. You need at least one voice in your head that says “who profits from you believing this?”

Read this if: you want the adversarial view — the one that treats AI hype as a power play and demands you examine who benefits.

04

Human Compatible: Artificial Intelligence and the Problem of Control

Russell co-authored Artificial Intelligence: A Modern Approach, the textbook that has trained more AI researchers than any other single document. When he writes about AI safety, he’s not a commentator. He’s one of the people who built the field.

Human Compatible starts with a provocation: the standard model of AI — build a system, give it an objective, let it optimize — is fundamentally broken. Not because the systems are too weak, but because they might become too strong. An AI that is given a fixed objective and the capability to pursue it relentlessly will, eventually, resist being turned off. Not out of malice. Out of logic. Being turned off interferes with completing the objective.

Russell’s solution is what he calls “provably beneficial AI” — systems designed to be uncertain about human preferences and willing to be corrected. The AI should want to be turned off, because its uncertainty about what humans actually want means deferring to us is always the rational move.

Published in 2019, before ChatGPT, before the current wave of capability jumps, the book reads now like a warning that arrived too early to be heard. The technical proposals may need updating, but the core ethical insight — that aligning AI with human values requires building machines that are humble about those values — remains one of the most important ideas in the field.

Read this if: you want the AI safety argument from the person most qualified to make it — rigorous, specific, and unsettlingly calm about the stakes.

05

Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence

Crawford does something none of the other books on this list attempt: she follows the supply chain. Atlas of AI traces the physical infrastructure behind artificial intelligence — the lithium mines, the data centers, the underpaid click workers labeling training data, the environmental costs of running models that consume more electricity than some countries.

The ethics of AI, Crawford argues, cannot be separated from the ethics of extraction. Every model has a material footprint. Every dataset has a labor story. Every deployment has a power dynamic. The algorithms are the visible layer. Beneath them is a planetary system of resource extraction, exploited labor, and concentrated wealth that looks less like the future and more like a very old pattern wearing new clothes.

This is the book that changed how I think about the phrase “AI ethics.” Before reading Crawford, I thought of it as a software problem — bias in data, fairness in outputs, transparency in decision-making. After reading her, I understood it as a systems problem that starts in a mine in the Democratic Republic of Congo and ends in a recommendation algorithm on your phone. You can’t fix the algorithm without seeing the mine.

Published in 2021, some of the specific numbers have shifted (the scale has only grown), but the framework is more relevant than ever. As AI models get larger and more resource-intensive, Crawford’s question — who pays the real cost? — becomes harder to avoid.

Read this if: you want to understand the material reality behind the algorithms — the labor, the resources, and the environmental costs that never appear in the product demo.


What these books have in common

Five books, five very different approaches. A Princeton debunking. A philosophical proof. A political prosecution. A safety proposal. A supply-chain investigation. But they all circle the same core question: Who is responsible?

That question — responsibility — is what separates real AI ethics from the performative kind. Corporate AI ethics tends to focus on principles: fairness, transparency, accountability. Those words appear in every AI company’s policy document. They mean almost nothing until you ask the follow-up: Who absorbs the cost when this goes wrong?

Narayanan and Kapoor show you the systems that don’t work and the people harmed by the pretense that they do. Bender and Hanna show you the power structures that benefit from keeping the con running. Russell shows you the long-term safety problem of building systems that resist correction. Crawford shows you the invisible labor and extraction that make the whole thing run. And The Last Skill argues that ethical reasoning itself — the act of choosing values and bearing consequences — is something that requires human agency in a way that can’t be automated.

Together, these five books don’t give you a neat framework. They give you something better: the ability to see the full shape of the problem. The technical dimension, the political dimension, the philosophical dimension, and the material dimension. Most people engaging with AI ethics are only looking at one of those. Read all five and you’ll be looking at all four.

Where to start

If you’re new to the topic, start with AI Snake Oil. It will give you the factual grounding to evaluate everything else. If you already have the facts and want the deeper question — what can’t be automated, even in principle — go to The Last Skill. If you’re the kind of reader who wants to be challenged and provoked, The AI Con will do that. If you want the long view, Human Compatible. If you want the wide view, Atlas of AI.

AI ethics matters because AI decisions are human decisions wearing a different mask. Every algorithm was written by someone. Every dataset was curated by someone. Every deployment was approved by someone. The machines don’t have ethics. The people behind them do — or don’t. These books help you tell the difference.

Related reading


Juan C. Guerrero is a Costa Rican founder, the publisher behind Anthropic Press, and the author of The Last Skill: What AI Will Never Own. He builds with AI and thinks about the moral weight of those choices.