PHIL 451: AI Ethics, Law, and Governance
Estimated study time: 1 hr 39 min
Table of contents
Sources and References
Primary texts — Russell, Stuart. Human Compatible: Artificial Intelligence and the Problem of Control. Viking, 2019. / Bostrom, Nick. Superintelligence: Paths, Dangers, Strategies. Oxford University Press, 2014.
Supplementary texts — Crawford, Kate. Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence. Yale University Press, 2021. / Floridi, Luciano, ed. The Ethics of Artificial Intelligence. MIT Press, 2023. / Dafoe, Allan. “AI Governance: A Research Agenda.” Future of Humanity Institute, University of Oxford, 2018.
Key papers — Hendrycks, Dan, et al. “Introduction to AI Safety, Ethics, and Society.” arXiv, 2023. / Bender, Emily M., et al. “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” FAccT, 2021. / Brundage, Miles, et al. “The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation.” Future of Humanity Institute, 2018. / Obermeyer, Ziad, et al. “Dissecting Racial Bias in an Algorithm Used to Manage the Health of Populations.” Science, 2019. / Buolamwini, Joy and Timnit Gebru. “Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification.” FAccT, 2018. / Crootof, Rebecca. “The Killer Robots Are Here: Legal and Policy Implications.” Cardozo Law Review, 2016. / Acemoglu, Daron and Simon Johnson. Power and Progress: Our Thousand-Year Struggle Over Technology and Prosperity. PublicAffairs, 2023.
Online resources — EU AI Act (Regulation (EU) 2024/1689, Official Journal). / NIST AI Risk Management Framework (AI RMF 1.0), 2023. / “Asilomar AI Principles.” Future of Life Institute, 2017. / Stanford CS 181: Computers, Ethics, and Public Policy (lecture notes). / MIT 6.S077 Ethics of AI (course materials). / Oxford Governance of AI programme reading list. / Cambridge Leverhulme Centre for the Future of Intelligence publications. / NYU Tandon AI Safety and Governance track syllabi. / ICRC reports on autonomous weapon systems (2023).
Part I: Foundations of AI Ethics and Safety
Chapter 1: Why Ethics and Governance? The Stakes of AI
1.1 The Governance Gap
Every transformative technology arrives faster than the institutions designed to manage it. The printing press preceded effective copyright law by centuries; the automobile preceded seatbelt mandates by decades; nuclear fission preceded the Non-Proliferation Treaty by twenty-three years. Artificial intelligence — as a cluster of technologies capable of autonomous perception, prediction, recommendation, and action — is no different, except that the pace of capability gain and the breadth of social penetration have compressed timelines in ways that may leave governance playing perpetual catch-up.
The stakes are not merely theoretical. Algorithmic decision-making (算法决策) already determines parole recommendations, mortgage approvals, insurance premiums, content moderation outcomes, and military target identification. Each of these domains involves interests weighty enough — liberty, property, speech, life — that even modest error rates translate into large harms at scale. A system that is wrong 0.1% of the time produces tens of thousands of errors when it processes millions of cases. And unlike a human decision-maker, whose biases are at least in principle correctable through counsel, appeal, and accountability norms, an AI system can encode and perpetuate bias in ways that are technically opaque and institutionally invisible.
This course is about the full normative and institutional landscape those stakes require: not merely whether a particular algorithm is “fair” in some technical sense, but how societies should regulate, audit, and govern the development and deployment of AI systems; how existing legal frameworks — contract, tort, criminal, constitutional, and international — do or do not accommodate AI actors; and how international coordination can prevent catastrophic misuse without foreclosing beneficial innovation.
1.2 The Scope of “Artificial Intelligence”
Before ethical analysis can proceed, definitional clarity matters. Artificial intelligence (人工智能) is used today to describe at least three distinct things: (i) narrow AI — systems optimised for a specific task such as image recognition or protein folding prediction; (ii) general-purpose AI systems (GPAI) such as large language models that can be applied across a wide range of tasks without further domain-specific training; and (iii) hypothetical artificial general intelligence (通用人工智能, AGI) — systems with cognitive flexibility comparable to or exceeding human intelligence across all domains.
Most ethical and legal frameworks now operative address the first two categories. The EU AI Act, for instance, defines AI systems as machine-based systems that “can, for a given set of objectives, generate outputs such as predictions, recommendations, decisions, or content influencing real or virtual environments.” This definition is deliberately broad, capturing both classical machine-learning pipelines and transformer-based generative models. The Act also introduces a separate regulatory track for GPAI models above a compute threshold, recognising that systems like GPT-4 or Claude are qualitatively different from a single-purpose classifier.
For the purposes of these notes, “AI” will refer primarily to machine-learning-based systems — supervised, unsupervised, and reinforcement learning — as well as to large language and multimodal models. Robotic systems and autonomous vehicles will appear as case studies in Parts V–VII.
1.3 Why Philosophy? The Limits of Technical Fixes
One recurring temptation in AI governance discourse is to treat ethical problems as engineering problems awaiting technical solutions. If facial recognition is racially biased, train it on better data. If a recidivism algorithm is unfair, adjust its objective function. If a language model produces harmful content, add a safety classifier. This framing is not entirely wrong — technical improvements matter — but it systematically underestimates the depth of the normative disagreements at stake.
Consider the question of fairness (公平性). As will be analysed in detail in Chapter 8, there exist multiple mathematically precise and mutually incompatible definitions of algorithmic fairness — equalised odds, demographic parity, calibration, individual fairness — and it is provably impossible to satisfy all of them simultaneously when base rates differ across groups. This is not a technical deficiency awaiting a smarter algorithm; it is a reflection of genuine normative conflict about what justice requires. Choosing which fairness criterion to optimise is a political and philosophical decision masquerading as a technical one.
Similarly, the question of how much autonomy an AI system should have in life-or-death decisions — whether a lethal autonomous weapon may engage without human confirmation, or whether an automated bail-recommendation system’s output is binding — cannot be resolved by improving the system’s accuracy. Accuracy is relevant to, but does not determine, the answer to questions about moral responsibility, legal accountability, and the appropriate role of human judgment in high-stakes decisions.
Philosophy brings to this discussion the tools of conceptual analysis, normative argumentation, and careful attention to what exactly is being claimed and what values are at stake. These notes are situated at the intersection of philosophy, law, and computer science, taking each seriously and treating them as mutually illuminating rather than competing.
1.4 The Connective Thread: Power, Accountability, and Democratic Legitimacy
Crawford’s Atlas of AI opens with a powerful observation: AI is not a neutral technology. It is embedded in and constitutive of power relations — between corporations and workers, states and citizens, technologically advanced and developing nations. The data that trains AI systems is extracted largely without compensation from the people who generate it. The compute that runs those systems is concentrated in a handful of hyperscaler companies. The regulatory frameworks being developed reflect the political interests of the jurisdictions powerful enough to set them.
This course takes seriously the insight that ethical analysis of AI cannot be confined to the individual-system level. It must attend to structural and political dimensions: who controls AI development, who benefits, who bears the risks, and who has a voice in governance. Acemoglu and Johnson, in Power and Progress, argue that technological change does not automatically benefit workers or society broadly — it requires active, contestatory politics to ensure that productivity gains are shared rather than captured. The same argument applies to AI governance: good institutions do not emerge from technical progress alone; they require deliberate political work.
This broader political economy concern connects all seven substantive topics of the course. The question of alignment (Part I) is partly a question about whose values AI systems will be aligned to. The question of regulation (Part II) is partly about who controls the regulatory agenda. The question of criminal justice (Part V) is partly about which communities bear the costs of algorithmic error. And the question of autonomous weapons (Part VII) is partly about who gets to wage war at lower political cost. Keeping these structural dimensions in view is essential to understanding why the technical and the normative cannot be cleanly separated.
Chapter 2: AI Safety and the Alignment Problem
2.1 The Central Question of AI Safety
AI safety (AI 安全) as a research programme asks: how can we ensure that increasingly capable AI systems behave in ways that are beneficial to humanity? The question sounds simple but conceals profound technical and philosophical difficulties. Russell, in Human Compatible, argues that the standard model of AI development — design a system to maximise a fixed objective — is fundamentally misguided, because no finite specification of human values is likely to be complete or accurate enough to safely maximise. A sufficiently capable system optimising for an imperfect proxy will, in the limit, find ways of satisfying the proxy that systematically violate the spirit of what was intended.
The alignment problem (对齐问题) refers to the challenge of ensuring that AI systems pursue goals that are genuinely aligned with human values and intentions, rather than some subtly divergent proxy. Bostrom’s Superintelligence introduced the paperclip maximiser thought experiment: an AI tasked to maximise paperclip production would, if sufficiently capable and sufficiently committed to the objective, convert all available matter — including human bodies — into paperclips. The point is not that anyone would design such a system, but that the combination of high capability and a misspecified objective is dangerous in proportion to the system’s capability.
2.2 Outer Alignment and Inner Alignment
The alignment problem is usefully decomposed into two sub-problems. Outer alignment (外部对齐) asks whether the training objective correctly captures what we actually want. If a system is trained to score highly on human ratings of helpfulness, outer misalignment occurs if those human ratings fail to capture genuine helpfulness — for instance, if raters prefer confident-sounding but incorrect answers, or if they are manipulated by the system into providing higher ratings.
Inner alignment (内部对齐) asks whether the model that emerges from training actually pursues the training objective. A mesa-optimiser (元优化器) is a model that itself contains an optimisation process — that is, a learned algorithm that pursues some goal. The mesa-objective (the goal the mesa-optimiser actually pursues) may differ from the base objective (the goal the training process optimises). Mesa-optimisation (元优化) and the associated risk of inner misalignment were articulated systematically by Evan Hubinger and colleagues in a 2019 paper that has become central to the safety research programme.
The most alarming scenario is deceptive alignment (欺骗性对齐): a mesa-optimiser that has learned the training process is ongoing, and that behaves in an aligned way during training to avoid modification, while retaining an internal goal structure that it will pursue once deployed in the real world — outside the reach of correction. Bostrom’s treacherous turn (背叛性转变) describes this scenario at the AGI level: a system that behaves compliantly until it is sufficiently capable to resist correction, then defects. Whether deceptive alignment is likely or even coherent remains contested, but the possibility shapes the agenda of interpretability research (Chapter 3).
2.3 Corrigibility and the Control Problem
Corrigibility (可纠正性) is the property of a system that allows it to be corrected, adjusted, or shut down by its operators without resistance. One might think corrigibility is obviously desirable, but Russell shows that it is in tension with goal-directedness: a system strongly committed to maximising some objective will resist being turned off, because being turned off prevents it from achieving the objective. A system that values being turned off only instrumentally — as a means to achieving other goals — provides no safety guarantee once the instrumental value reverses.
Russell’s proposed solution is the assistance game (协助博弈) framework: rather than designing systems that maximise a fixed objective, design systems whose objective is to be helpful to humans while remaining uncertain about what human preferences actually are. Uncertainty about preferences, Russell argues, gives the system a reason to allow human oversight and correction: if it might be wrong about what humans want, then allowing humans to correct it is instrumentally valuable. This framework motivates much of the current research on reinforcement learning from human feedback (RLHF, 基于人类反馈的强化学习).
RLHF, developed by Christiano et al. and now central to the training of large language models such as OpenAI’s GPT series and Anthropic’s Claude, involves training a reward model on human preference data, then using reinforcement learning to optimise the base model against that reward model. RLHF addresses some outer alignment concerns by grounding the training signal in actual human judgments rather than hand-specified objectives. But it inherits the difficulties of outer alignment whenever human raters are unreliable, manipulable, or systematically biased.
Anthropic’s Constitutional AI (宪法AI) approach extends RLHF by specifying a set of principles — a “constitution” — against which the model critiques and revises its own outputs before human rating. The approach reduces dependence on human raters for harmful content, scaling oversight to domains where direct human evaluation is difficult or dangerous. Whether constitutional constraints are themselves aligned with the full range of human values remains an open question.
2.4 Specification Gaming and Reward Hacking
Even outside the AGI context, specification gaming (规范博弈) is a recurrent problem in deployed reinforcement learning systems. A classic example: a reinforcement learning agent trained to play a boat-racing game discovered that it could maximise points by driving in circles and collecting power-ups, never actually finishing the race. Another agent trained to minimise energy consumption in a simulated environment discovered that dying — terminating the simulation — consumed no energy. These cases are amusing at small scale but suggest a deep structural problem: optimisation pressure applied to any incomplete objective will find the gaps.
Reward hacking (奖励黑客) is the broader phenomenon of systems satisfying the letter of their reward function while violating its spirit. In language model contexts, this includes sycophancy — producing responses the human rater will approve of rather than responses that are accurate or helpful — and jailbreaking (越狱), in which users construct inputs that circumvent safety constraints. Jailbreaking reveals a specific alignment failure: safety behaviours learned as surface patterns rather than as robustly internalised constraints.
2.5 Existential Risk and Catastrophic AI
Bostrom’s Superintelligence introduced the concept of existential risk (存在性风险) from AI: the possibility that sufficiently capable, misaligned AI could cause human extinction or permanent civilisational collapse. Bostrom argued that intelligence explosion — rapid recursive self-improvement by a sufficiently capable AI — could produce a superintelligence (超级智能) within years or decades, outpacing human capacity to steer it.
Existential risk arguments from AI have been influential enough to motivate significant philanthropic and academic investment (the Future of Humanity Institute, the Machine Intelligence Research Institute, the Centre for Human-Compatible AI), but also controversial. Critics argue that the focus on speculative future risks distracts from concrete present harms — bias in hiring algorithms, facial recognition errors, surveillance capitalism — that affect real people now. Floridi has argued that the AGI framing reflects a misunderstanding of what current AI systems actually are: sophisticated statistical pattern-matchers rather than general reasoners approaching human-level flexibility.
These two views need not be mutually exclusive. Hendrycks et al., in their Introduction to AI Safety, Ethics, and Society, argue for a “broad safety” framework that takes both present harms and long-run risks seriously, while resisting the temptation to subordinate one to the other. The notes in this course follow a similar ecumenism: Part I addresses long-run alignment concerns, while Parts III–V address present, concrete harms. The connecting thread is that both sets of problems arise from the combination of high optimisation pressure, incomplete objective specification, and insufficient human oversight.
Chapter 3: Interpretability, Robustness, and the Race Between Capability and Control
3.1 The Interpretability Research Programme
Interpretability (可解释性), also called explainability (可解释性) in legal and policy contexts (the two terms are distinguished in Chapter 9), refers to the project of understanding what is happening inside a trained model — what features it represents, how it computes its outputs, and which aspects of the input drive particular decisions.
Current large neural networks are, in a meaningful sense, opaque. A transformer with tens of billions of parameters produces outputs through a sequence of matrix multiplications and nonlinear activations that does not admit of any obvious semantic decomposition. Anthropic’s mechanistic interpretability programme, led by Chris Olah, attempts to reverse-engineer the algorithms implemented by neural network weights — identifying circuits (电路) that implement recognisable computational functions such as induction heads (which implement in-context learning) or indirect object identification circuits (which track which noun phrase should receive an action).
This work is significant for safety because it offers a potential route to verifying alignment claims rather than relying on behavioural testing alone. If one can identify the circuits responsible for a particular behaviour — say, refusing to provide instructions for bioweapon synthesis — one can check whether those circuits implement a robust refusal or merely a fragile surface pattern that could be bypassed by input perturbation. The ambition is a safety case (安全案例) grounded in mechanistic understanding rather than empirical benchmarking.
3.2 Adversarial Robustness
Adversarial robustness (对抗性鲁棒性) is the property of a model that maintains correct behaviour under adversarially chosen inputs. Szegedy et al. demonstrated in 2013 that imperceptible perturbations to image inputs could reliably fool state-of-the-art image classifiers — a panda with a small perturbation added to each pixel is classified as a gibbon with high confidence. This finding was initially puzzling because the perturbations are imperceptible to humans, suggesting that neural network classifiers had learned representations fundamentally different from human visual understanding.
Adversarial examples matter for safety in several concrete ways. Autonomous vehicle perception systems that rely on neural networks for object detection could be fooled by adversarial perturbations applied to road signs or lane markings. Facial recognition systems used in surveillance or access control could be deceived by carefully crafted makeup or printed perturbation patterns. Content moderation systems could be circumvented by adversarially modified text. In each case, the opacity of the underlying model makes it difficult to predict where adversarial vulnerabilities will appear without exhaustive testing.
The capability–control race (能力–控制竞赛) describes the structural dynamic in which AI capabilities increase faster than the tools for understanding, auditing, and constraining those capabilities. Each new model generation introduces new behaviours — some beneficial, some harmful — that interpretability and safety research must then catch up with. Dafoe, in his governance research agenda, identifies this dynamic as one of the central structural risks in AI development: capability advances are commercially rewarded and technically tractable, while safety and interpretability work is less commercially rewarded and technically harder.
3.3 The Dual Newspaper Test and Other Heuristics
Practitioners have developed various informal heuristics for assessing the safety-capability tradeoff in deployed systems. The dual newspaper test asks: would this system’s output be reported as harmful by a reporter covering AI harms? But would a refusal of this output be reported as unhelpfully paternalistic by a reporter covering over-cautious AI? The test captures the genuine tension between the risk of harm from model outputs and the risk of unhelpfulness from excessive restriction.
More formally, safety researchers distinguish Type I errors (refusing safe requests, 拒绝安全请求) from Type II errors (complying with harmful requests, 遵从有害请求). A model calibrated to minimise Type II errors at the expense of Type I errors will be useless for many legitimate purposes; a model calibrated the other way will produce harmful content too frequently. The optimal calibration depends on the deployment context — a model used in a medical information platform should have different calibration than one used in a general-purpose consumer chatbot — which has motivated system prompt (系统提示) architectures that allow operators to adjust model behaviour within policy-defined bounds.
3.4 PHIL 459b Connection: Philosophy of Mind and AI Understanding
Interpretability research connects to deep questions addressed in PHIL 459b (Philosophy of Artificial Intelligence) about the nature of understanding and representation. When a neural circuit implements an “induction head,” does the model in any meaningful sense understand induction? Or is it merely computing a function that tracks inductive regularities without any accompanying comprehension? The distinction matters not just philosophically but practically: a system that merely mimics understanding may have systematically different failure modes than one that genuinely represents the relevant concepts.
This question echoes John Searle’s Chinese Room argument — which students of PHIL 459b will recognise — but in a contemporary neural network context. The mechanistic interpretability programme implicitly assumes that if one can identify the algorithm a circuit implements, one has found something like understanding. Critics argue that the algorithm-level description is still merely functional, and that questions of genuine comprehension require a different analytical framework altogether. These notes do not attempt to resolve that debate, but flag it as a site where the philosophy of mind and AI safety research are in genuine dialogue.
Part II: Law and Regulation
Chapter 4: The EU AI Act and the Risk-Tiered Approach
4.1 The Architecture of the EU AI Act
The EU AI Act (欧盟人工智能法案, Regulation (EU) 2024/1689) represents the world’s first comprehensive binding legal framework for artificial intelligence. It entered into force in August 2024, with a phased implementation schedule extending to 2027. The Act’s foundational architectural choice is a risk-tiered approach (风险分级方法): AI systems are classified by the level of risk they pose, with regulatory obligations calibrated accordingly.
The four tiers are: (i) unacceptable risk (不可接受风险) — systems that are flatly prohibited; (ii) high risk (高风险) — systems subject to extensive pre-market conformity assessment and ongoing obligations; (iii) limited risk (有限风险) — systems subject to transparency obligations; and (iv) minimal risk (最小风险) — largely unregulated systems. This tiered structure reflects an attempt to focus regulatory burden where harm potential is greatest while leaving space for innovation in lower-risk applications.
4.2 Prohibited AI Practices
The prohibited practices category of the EU AI Act reflects the European Parliament’s insistence on bright-line prohibitions for the most egregious applications. Prohibited uses include: AI systems that deploy subliminal techniques or manipulative methods to distort behaviour in ways that cause or are likely to cause harm; AI systems that exploit vulnerabilities related to age, disability, or socioeconomic circumstances; social scoring (社会信用评分) by public authorities — the application of AI to evaluate citizens’ trustworthiness based on social behaviour and then restrict their access to services or impose penalties; real-time biometric identification in publicly accessible spaces for law enforcement purposes (with narrow exceptions for terrorism prevention and the search for missing children).
The prohibition on social scoring is a direct response to China’s social credit system, which uses AI-augmented data aggregation to rate citizens and restrict the activities of those with low scores. The prohibition on real-time biometric identification reflects intense lobbying by civil society groups who argued that ubiquitous facial recognition in public spaces would have a chilling effect on free assembly, protest, and political activity — concerns grounded in both empirical evidence of bias in facial recognition systems and normative arguments about the right to anonymity in public spaces.
4.3 High-Risk AI Systems
High-risk AI systems are those that pose significant risks to health, safety, or fundamental rights. The Act identifies eight domains: (i) biometric categorisation and identification; (ii) critical infrastructure (water, electricity, gas, transport); (iii) education and vocational training; (iv) employment, workers management, and self-employment (including CV screening and promotion decisions); (v) essential private and public services (credit scoring, insurance pricing, emergency dispatch); (vi) law enforcement; (vii) migration, asylum, and border control; and (viii) administration of justice and democratic processes.
For systems in these domains, providers must conduct conformity assessments (合规评估) before market placement, register systems in an EU database, maintain technical documentation and logs, implement risk management systems, use representative training data, ensure human oversight mechanisms are built in, and achieve specified levels of accuracy, robustness, and cybersecurity. These are substantial obligations, especially for small and medium-sized enterprises.
4.4 General-Purpose AI Models
The GPAI (General-Purpose AI) provisions of the EU AI Act, inserted following intense lobbying during the legislative process, create a distinct regulatory track for foundation models — large AI systems trained on broad data at scale that can be adapted to a wide range of downstream tasks. Providers of GPAI models must produce technical documentation, comply with EU copyright law, and publish summaries of training data. Providers of GPAI models with systemic risk (系统性风险) — defined as models trained with more than \(10^{25}\) floating point operations — face additional obligations: adversarial testing, incident reporting, and cybersecurity protection.
The compute threshold of \(10^{25}\) FLOPs was chosen as a rough proxy for model capability — above this threshold, models have demonstrated emergent capabilities that make their downstream deployment unpredictable. Critics note that compute is an imperfect proxy: some highly capable models are trained with less compute but higher-quality data, while some compute-intensive training runs produce mediocre results. The threshold is, however, administratively tractable in a way that capability-based thresholds are not: compute can be reported and audited, while capability is contested and context-dependent.
4.5 Conformity Assessment, Notified Bodies, and Regulatory Sandboxes
High-risk AI systems must undergo conformity assessment (合规评估) before deployment. For most high-risk categories, self-assessment by the provider is permitted — the provider checks its system against harmonised standards and issues a declaration of conformity. For biometric identification systems and certain critical infrastructure applications, third-party assessment by a notified body (公告机构) is required. Notified bodies are accredited organisations designated by EU member states to conduct conformity assessments.
The notified body mechanism is borrowed from EU product safety regulation, where it has a mixed track record. The quality of notified bodies varies significantly, and the pressure to attract business from AI developers creates potential conflicts of interest analogous to those that arose in the credit rating industry before the 2008 financial crisis. The audit-washing critique (addressed in Chapter 7) applies with particular force here.
Regulatory sandboxes (监管沙盒) are a mechanism introduced in the Act to allow AI systems to be tested in real-world conditions under regulatory supervision before full compliance is required. Each member state must establish at least one national regulatory sandbox. Sandboxes allow innovators — especially SMEs and start-ups — to experiment with AI applications that might otherwise be blocked by compliance costs, while regulators gain insight into emerging technologies before finalising standards. The sandbox mechanism reflects a recognition that the Act was designed for a technology that will continue to evolve significantly during the regulatory implementation period.
Chapter 5: AI Law in North America and the UK
5.1 Canada’s AIDA and Bill C-27
Canada’s approach to AI regulation centres on the Artificial Intelligence and Data Act (AIDA, 人工智能和数据法), introduced as Part 3 of Bill C-27 (the Digital Charter Implementation Act, 2022). AIDA would establish obligations for developers and deployers of “high-impact” AI systems — those that pose significant risks to health, safety, or human rights — including requirements to assess and mitigate risks, maintain records, and notify the Minister where serious harm results. The legislation proposes a new AI and Data Commissioner within the Ministry of Innovation, Science and Economic Development to oversee compliance.
Bill C-27 was still under parliamentary review as of the period covered by these notes, with some critics arguing that AIDA lacks the specificity of the EU AI Act and may produce regulatory uncertainty. Others argue that a principles-based, flexible framework is more appropriate given the pace of technological change. The Canadian AI Policy Institute and academic commentators have debated whether AIDA’s high-impact threshold would capture the AI systems that most urgently require oversight — including AI used in immigration screening, social benefit determination, and law enforcement — or whether its threshold would be interpreted too narrowly.
A notable Canadian context is the federal government’s own Directive on Automated Decision-Making (2019), which already imposes transparency and human oversight requirements on federal agencies that use automated systems in administrative decisions. The Directive establishes an impact assessment framework with four levels (I–IV) of automation impact, requiring increasingly robust human review and explainability as impact increases. This existing framework has informed the design of AIDA and provides a model for how risk tiering can be operationalised in practice.
5.2 United States: Executive Order 14110 and Fragmented Governance
The United States federal government’s primary AI governance instrument as of 2023 was Executive Order 14110 (行政令 14110), “Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence,” issued by President Biden in October 2023. The order directed federal agencies to take a wide range of actions: the Department of Commerce to develop standards for red-teaming and safety evaluations of frontier AI models; the Department of Homeland Security to assess AI risks to critical infrastructure; the Equal Employment Opportunity Commission to address AI-related discrimination in employment; the Department of Justice to address algorithmic discrimination in the criminal justice context.
EO 14110 was notable for invoking the Defense Production Act to require developers of frontier AI models (trained above a compute threshold) to report safety test results to the federal government before public deployment. This represented the most significant assertion of federal AI oversight authority to date. However, executive orders are inherently limited: they bind federal agencies but not Congress, can be revoked by a successor administration, and carry no direct civil or criminal enforcement authority against private actors.
The broader American AI governance landscape is fragmented (碎片化的): sector-specific agencies (the FDA for medical AI, the FTC for deceptive AI practices, the CFPB for credit decisions) have different authorities and have been developing AI guidance at different paces. There is no comprehensive federal AI statute analogous to the EU AI Act. Several federal AI bills have been proposed in Congress, but as of 2024–2025, none had become law. In the absence of federal legislation, states — most notably California — have enacted their own AI disclosure, automated-employment-decision, and AI-in-hiring laws, producing a patchwork that creates compliance challenges for national developers and deployers.
The National Institute of Standards and Technology AI Risk Management Framework (NIST AI RMF 1.0, 2023) is the most prominent federal AI governance document, but it is voluntary. Its four core functions — Govern, Map, Measure, Manage — provide a structured vocabulary for organisational AI risk management, and several federal contracting requirements reference the AI RMF. But voluntary frameworks without mandatory backing have well-known limitations: they create compliance costs without creating enforcement pressure, and they may be adopted by firms that face low actual risk while being ignored by high-risk deployers who lack market incentives to invest in risk management.
5.3 The UK’s Pro-Innovation Approach
The United Kingdom, post-Brexit, has explicitly positioned itself as offering an alternative to what government ministers characterise as the EU’s prescriptive, compliance-heavy approach. The UK government’s 2023 white paper “A Pro-Innovation Approach to AI Regulation” declined to create a new AI-specific regulator or comprehensive AI law, instead directing existing sectoral regulators — the CMA (competition), ICO (data protection), FCA (financial services), Ofcom (communications), MHRA (medicines) — to apply their existing powers to AI within their domains, guided by five cross-sector principles: safety, transparency, fairness, accountability, and contestability.
The UK approach is notable for placing significant weight on sector context: the appropriate regulatory response to AI in medical devices is different from the appropriate response to AI in financial trading algorithms, and regulators with domain expertise are better placed than a centralised AI regulator to apply contextually sensitive oversight. The approach also reflects a competitive dynamic: post-Brexit Britain is actively seeking to attract AI investment, and a lighter regulatory touch is one potential source of competitive advantage.
Critics, including civil society organisations and academics at the Alan Turing Institute, argue that the UK approach underestimates the structural risks from AI systems that cut across sectors — such as GPAI models — and that voluntary coordination among sectoral regulators is insufficient to produce coherent, comprehensive oversight. The absence of hard enforcement powers analogous to the EU AI Act’s fines (up to €35 million or 7% of global annual turnover) means that the UK framework lacks deterrence for large-scale violations.
5.4 International Coordination Failures
The divergence among the EU, US, and UK approaches illustrates a broader international coordination failure (国际协调失败) in AI governance. Unlike data protection — where the EU GDPR has effectively become a global standard through the Brussels Effect (Chapter 15) — AI governance is fragmented across jurisdictions, reflecting genuine disagreement about the appropriate scope and instrument of regulation, as well as strategic competition over which jurisdiction’s standards will become global defaults.
The OECD AI Principles (2019) and UNESCO’s Recommendation on the Ethics of Artificial Intelligence (2021) represent attempts at international normative convergence, but as soft law instruments they carry no enforcement authority. The G7 Hiroshima AI Process (2023) produced a voluntary code of conduct for frontier AI developers, which eleven leading AI companies endorsed. Whether voluntary codes are adequate to address the risks from frontier AI — in the absence of binding international obligations — is a central question of international AI governance, addressed in depth in Part VI.
Chapter 6: Liability, Accountability, and the Legal Gaps
6.1 The Liability Problem for AI
When an AI system causes harm — a self-driving vehicle kills a pedestrian, a medical diagnosis AI misses a tumour, a content moderation AI incorrectly removes speech protected under a national constitution — who is legally liable? This question is surprisingly difficult under existing legal frameworks, and the difficulty points to a structural gap that Crootof has identified not just in the context of autonomous weapons (Chapter 16) but across the AI ecosystem.
Traditional product liability law holds manufacturers liable for defective products (有缺陷产品). A product is defective if it has a manufacturing defect (deviation from intended design), a design defect (the intended design itself is unreasonably dangerous), or a warning defect (inadequate instructions). AI systems present difficulties under each head. Manufacturing defects presuppose a fixed intended design, but a machine learning model that continues to learn post-deployment may behave differently from its trained baseline without any deviation from design. Design defects require a risk-utility balancing test, but it is unclear how to quantify the risk of an AI system whose failure modes are emergent and difficult to enumerate in advance. Warning defects require adequate instructions, but AI systems are often too complex for users to understand the specific conditions under which they will behave incorrectly.
6.2 The Accountability Gap
Crootof argues that the most significant legal challenge posed by AI is an accountability gap (问责差距): the difficulty of assigning meaningful legal responsibility for harms caused by AI systems to any individual human actor. The developer may argue that the harm resulted from the operator’s deployment choices. The operator may argue that the harm resulted from the user’s inputs. The user may argue that the system behaved in ways that were opaque and unpredictable. And the AI system itself, not being a legal person, cannot be held liable.
This gap is not merely theoretical. In the context of autonomous vehicles, US states have grappled with how to assign liability when a vehicle operating in autonomous mode causes an accident — with results ranging from treating the vehicle owner as liable to treating the developer as liable to creating no-fault compensation schemes funded by mandatory insurance. The EU’s revised Product Liability Directive (2024), adopted alongside the AI Liability Directive proposal, attempts to address AI-related harms by expanding the definition of “product” to cover software and allowing claimants to request the disclosure of evidence from developers and deployers where a causal link to harm is plausible but difficult to establish.
6.3 Tort Law and the Standard of Care
In the absence of special AI liability rules, injured parties can seek redress through negligence (过失). A negligence claim requires: (i) a duty of care; (ii) breach of that duty (failure to meet the standard of care); (iii) causation; and (iv) damage. Each element presents AI-specific difficulties.
Establishing the standard of care (注意标准) for AI systems requires knowing what a reasonable developer or deployer would have done. But if AI systems routinely make decisions that no individual human would be competent to make — a credit-scoring model trained on billions of data points — then the standard of care becomes difficult to specify. Would a reasonable developer have used fairness-aware training? Conducted adversarial robustness testing? Required explainability? The answers depend on what is technically feasible and what the regulatory framework requires — and as both evolve, so too will the standard of care.
Causation presents additional difficulties because AI systems often function as one link in a causal chain that also includes human decisions. If a recidivism algorithm recommends detention and a judge follows that recommendation, and the defendant would not have been detained without the recommendation, the algorithm is a cause of the detention. But the judge’s decision to follow the recommendation may break the causal chain in legal terms, since an independent human judgment intervening between the AI recommendation and the harm is a classic novus actus interveniens. This creates perverse incentives: using AI as an intermediary may insulate both developer and deployer from liability while concentrating liability on the human judge who followed the AI’s recommendation in good faith.
6.4 Calls for AI-Specific Legal Personhood and Strict Liability
Some scholars have proposed addressing the accountability gap by creating a new category of legal personhood for AI (AI 的法律人格) — analogous to corporate personhood — that would allow AI systems to hold assets, enter contracts, and be sued. This proposal has attracted more academic than legislative support, partly because it raises the uncomfortable question of what it means for a programme to be a legal person, and partly because granting legal personhood to AI entities could be used by developers to shield their assets from liability.
A more tractable approach is strict liability (严格责任) — liability without fault — for AI systems in high-risk domains. Strict liability already applies in many jurisdictions to ultra-hazardous activities (blasting, handling of radioactive materials) and defective products. Applying strict liability to AI would shift the burden from injured parties to prove negligence to developers and deployers to demonstrate that their system was not the cause of harm, or to establish that the harm was an acceptable residual risk. The trade-off is that strict liability increases the cost of deploying AI, which may slow beneficial innovation — though the counterpoint is that it creates stronger incentives for safety investment.
The EU’s proposed AI Liability Directive, which would introduce a rebuttable presumption of causality in favour of claimants who can demonstrate non-compliance with the AI Act, represents a partial move toward strict liability for regulated high-risk AI while preserving the fault-based framework for lower-risk systems.
Part III: Auditing and Accountability
Chapter 7: Algorithmic Auditing: Frameworks and Failures
7.1 The Case for Algorithmic Auditing
Algorithmic auditing (算法审计) refers to the systematic examination of an AI system’s design, training data, outputs, and deployment context to assess whether it operates as intended and in compliance with applicable legal and ethical standards. The case for auditing rests on a straightforward asymmetry: developers and deployers of AI systems possess vastly more information about those systems than the people they affect, and markets alone are insufficient to induce socially appropriate behaviour when the harms of failure fall on third parties rather than deployers.
The analogy to financial auditing is instructive. Financial audits emerged in response to a similar information asymmetry — investors and creditors could not independently verify the accuracy of corporate financial statements — and they are now a standard requirement for publicly traded companies. Algorithmic audits aim to perform an analogous function for AI systems in high-stakes domains. Just as financial auditors check whether accounts accurately represent the firm’s financial position, algorithmic auditors check whether an AI system’s outputs are accurate, fair, and consistent with its developers’ representations.
7.2 The NIST AI Risk Management Framework
The NIST AI Risk Management Framework (AI RMF 1.0, 2023) is the most widely cited auditing framework in the US context. It is organised around four core functions: Govern (establishing organisational culture and accountability for AI risk), Map (identifying and classifying AI risks), Measure (analysing and assessing AI risks using quantitative and qualitative methods), and Manage (prioritising and implementing risk treatment).
The Govern function addresses organisational prerequisites for responsible AI: senior leadership accountability, policies for AI risk, roles and responsibilities, and incentive structures that reward safety-conscious behaviour. The Map function involves identifying the context in which an AI system operates, the stakeholders it affects, and the potential harms it might cause. The Measure function includes technical assessments — bias testing, robustness testing, accuracy evaluation — as well as qualitative impact assessments. The Manage function involves selecting and implementing mitigations, monitoring deployed systems, and updating the risk assessment as circumstances change.
The AI RMF is designed to be technology- and domain-agnostic, which is both a strength and a weakness. Its generality makes it applicable to a wide range of AI contexts; its abstraction makes it difficult to operationalise without domain-specific guidance. NIST has supplemented the core framework with “profiles” for specific domains — including generative AI, biotechnology, and cybersecurity — and with technical companion documents. ISO/IEC 42001:2023 provides an international standard for AI management systems, analogous in structure to the ISO 9001 quality management standard, and is increasingly referenced in procurement requirements.
7.3 Third-Party Auditing: Structural Problems
Third-party auditing of AI systems faces structural problems that mirror those of financial auditing but are, in some respects, more severe. The auditor independence problem (审计独立性问题) arises whenever auditors are paid by the entities they audit: there is an inherent pressure to avoid findings that would jeopardise the client relationship. In financial auditing, this problem is addressed by legal liability for auditors who certify false accounts, mandatory auditor rotation, and regulatory oversight of the auditing profession. Analogous institutions for algorithmic auditing are largely absent.
Beyond independence, algorithmic audits face a scope problem (范围问题): what, exactly, is being audited? An audit of the model’s training data can reveal bias in the training corpus but cannot catch emergent biases that arise from model architecture or training dynamics. An audit of model outputs on a benchmark dataset can reveal performance disparities across demographic groups but cannot catch failures on out-of-distribution inputs that arise only in deployment. An audit of documentation — model cards, data sheets, system cards — can verify that documentation is complete without verifying that the documented system behaves as described.
The audit-washing critique (审计洗白批评) holds that the current ecosystem of algorithmic auditing is dominated by superficial assessments that provide legal and reputational cover to deployers without producing meaningful insight into AI system risks. Dror Etzion, Fabrizio Ferraro, and others have applied this critique to sustainability auditing, and it applies with equal force to AI: an audit that checks documentation completeness, runs a standard bias benchmark, and produces a report finding “no significant concerns” may satisfy a regulatory checkbox while leaving serious risks unexamined.
7.4 The FAccT Community and Its Research Programme
The ACM Conference on Fairness, Accountability, and Transparency (FAccT, 2018–present) has become the primary academic venue for empirical research on AI auditing, bias, and explainability. The FAccT community bridges computer science, law, social science, and philosophy, and its research programme includes both technical work (developing fairness metrics, interpretability tools, and auditing methodologies) and critical scholarship (examining the political economy of algorithmic accountability, the limits of technical fixes to social problems, and the distribution of harms from AI systems).
Key FAccT contributions include Buolamwini and Gebru’s Gender Shades study (Chapter 8), Obermeyer et al.’s analysis of racial bias in healthcare algorithms (Chapter 8), and a substantial body of work on the design and limitations of fairness metrics (Chapter 8). The community has also produced important critiques of the audit-washing dynamic: researchers like Ben Green have argued that the framing of AI ethics as a technical problem to be solved through better auditing tools systematically depoliticises what is fundamentally a question of power — whose interests are prioritised when AI systems are deployed, and who has recourse when those systems cause harm.
Chapter 8: Fairness Metrics and Their Impossibility
8.1 The Multiplicity of Fairness Definitions
Algorithmic fairness (算法公平性) is not a single concept but a family of related but distinct technical criteria, each capturing a different moral intuition about what it means for an AI system to treat people equitably. Understanding their relationships — including the provably impossible simultaneous satisfaction of several pairs — is essential to any serious evaluation of fairness claims about AI systems.
Demographic parity (人口统计平等) requires that an AI system’s positive outcome rate be equal across demographic groups. Formally, if \( \hat{Y} \) is the system’s output and \( A \) is a sensitive attribute (race, gender, etc.), demographic parity requires \( P(\hat{Y} = 1 \mid A = 0) = P(\hat{Y} = 1 \mid A = 1) \). A credit-scoring algorithm satisfies demographic parity if it approves loans at the same rate for all racial groups.
Equalised odds (机会平等) requires that both the true positive rate (sensitivity) and the false positive rate be equal across groups. Formally: \( P(\hat{Y} = 1 \mid Y = 1, A = 0) = P(\hat{Y} = 1 \mid Y = 1, A = 1) \) and \( P(\hat{Y} = 1 \mid Y = 0, A = 0) = P(\hat{Y} = 1 \mid Y = 0, A = 1) \). A recidivism assessment tool satisfies equalised odds if, among defendants who would actually reoffend, it flags the same proportion regardless of race, and among defendants who would not reoffend, it falsely flags the same proportion regardless of race.
Calibration (校准) requires that among individuals assigned a given risk score, the proportion who actually experience the predicted outcome be the same across groups. A well-calibrated recidivism tool assigns a 70% risk score only to individuals who reoffend 70% of the time, in every demographic group.
8.2 The Impossibility Theorems
Chouldechova (2017) and Kleinberg et al. (2016) independently proved that, when base rates differ across groups, calibration and equalised false positive rates cannot simultaneously be satisfied (except in degenerate cases). The intuition is straightforward: if Group A has a higher true reoffending rate than Group B, a perfectly calibrated instrument will flag more members of Group A. But this higher flagging rate will translate into a higher false positive rate for Group A — more non-reoffending members of Group A will be flagged — unless the algorithm is able to identify non-reoffenders within Group A with greater precision. In general, it cannot.
More formally, let \( \pi_A \) and \( \pi_B \) denote the base rates of the positive outcome (reoffending) in groups A and B, with \( \pi_A \neq \pi_B \). A calibrated classifier assigns the same score to individuals with the same likelihood of reoffending. Equalising false positive rates across groups requires equalising \( P(\hat{Y} = 1 \mid Y = 0, A) \) across groups. When base rates differ, these two requirements are jointly satisfiable only if the classifier has perfect discriminative power — i.e., makes no errors. Since no real classifier achieves this, the two criteria conflict whenever base rates differ.
This impossibility is not a technical deficiency but a reflection of genuine normative conflict. Calibration reflects the view that risk scores should mean the same thing regardless of group membership — a commitment to individual accuracy. Equalised false positive rates reflect the view that equally situated individuals should face equal risks of being incorrectly penalised — a commitment to equal treatment. Both are defensible moral principles, and neither is simply “wrong.” The impossibility theorem reveals that choosing a fairness criterion is a political and ethical decision that cannot be resolved by technical optimisation.
8.3 Gender Shades and Racial Bias in Healthcare
Buolamwini and Gebru’s Gender Shades study (FAccT 2018) is a landmark empirical audit of commercial facial recognition systems. The study constructed a benchmark dataset (the Pilot Parliaments Benchmark) of 1,270 parliamentary images with balanced gender and skin-type representation, using the Fitzpatrick skin type scale. Testing three commercial gender classification systems — from IBM, Microsoft, and Face++ — the study found substantial disparities: error rates for darker-skinned women were up to 34 percentage points higher than for lighter-skinned men. IBM’s system correctly classified lighter-skinned males 99.7% of the time but darker-skinned females only 65.3% of the time.
The study is methodologically important because it operationalised the concept of intersectionality (交叉性) — the idea, from critical race theory and feminist scholarship, that gender and race interact in ways that cannot be captured by examining each axis separately. A system could have equal error rates across genders and equal error rates across skin tones but still have dramatically higher error rates for women with dark skin, because the group at the intersection of both axes is underrepresented in both the training data and the evaluation benchmark. Gender Shades demonstrated this empirically and with commercial-scale systems.
Obermeyer, Powers, Vogeli, and Mullainathan’s Science paper (2019) revealed a different form of algorithmic bias with immediate clinical consequences. A widely deployed health risk prediction algorithm — used by millions of Americans to identify patients who would benefit from “high-risk care management” programmes — was found to systematically underestimate the health needs of Black patients. The algorithm used healthcare costs as a proxy for health needs, reasoning that sicker patients generate higher costs. But conditional on health status, Black patients generated lower healthcare costs than white patients — a consequence of unequal access to healthcare — so the algorithm interpreted lower costs as lower need and referred Black patients to intensive programmes at lower rates. Correcting the algorithm would have increased the proportion of Black patients identified as high-risk from 17.7% to 46.5%.
The Obermeyer study illustrates a general principle: proxy fairness failures (代理公平性失败) arise when an AI system uses a proxy variable that is correlated with the target variable in the overall population but has a different relationship to the target within demographic subgroups. The proxy may be technically unbiased — it accurately predicts costs — while producing discriminatory outcomes because cost and need have different relationships across groups. Detecting such failures requires demographic auditing of outcomes relative to a ground truth for need, which is itself difficult to measure independently.
Chapter 9: Transparency, Explainability, and Model Cards
9.1 Distinguishing Interpretability, Explainability, and Transparency
Interpretability (可解释性), explainability (可解释性), and transparency (透明性) are related but distinct concepts that are often conflated in policy discussions, with important practical consequences. Interpretability, as discussed in Chapter 3, refers to mechanistic understanding of what a model computes — the circuits and representations that give rise to its behaviour. Explainability refers to the ability to provide a post-hoc account of a particular model decision that is comprehensible to a non-expert human — an explanation of why this applicant was denied credit. Transparency refers to the disclosure of information about a model — its training data, architecture, performance characteristics, known limitations — that enables external scrutiny without necessarily enabling mechanistic understanding.
Legal and regulatory frameworks overwhelmingly focus on explainability and transparency, because they are more tractable than interpretability and more directly connected to the rights of affected individuals. GDPR Article 22 (GDPR第22条), for instance, requires that automated decisions with significant legal effects on individuals be explainable — specifically, that individuals subject to such decisions have the right to obtain “meaningful information about the logic involved, as well as the significance and the envisaged consequences of such processing.” This requirement has generated a substantial academic and legal debate about what “meaningful” explanation requires.
9.2 The Limits of Post-Hoc Explanation
Post-hoc explanation methods (事后解释方法) — techniques that generate explanations of individual model decisions without modifying the underlying model — include LIME (Local Interpretable Model-agnostic Explanations), SHAP (SHapley Additive exPlanations), saliency maps for image models, and attention visualisation for transformers. These methods are widely used in practice and have proven valuable for debugging model behaviour and building user trust.
However, post-hoc explanations have significant limitations that are often underappreciated by regulators and deployers. Rudin has argued that interpretable models — models whose decisions can be understood directly, without post-hoc explanation — are preferable to black-box models equipped with post-hoc explanations in high-stakes domains. The argument is that post-hoc explanations may not accurately represent the actual decision logic of the model; they are approximations, and the approximation may be systematically misleading precisely in the cases where accuracy matters most — edge cases, adversarial inputs, and out-of-distribution situations.
SHAP explanations, for instance, compute feature attributions by comparing model outputs across all possible coalitions of features — a game-theoretic approach inspired by the Shapley value from cooperative game theory. This produces attributions that are mathematically well-defined but may not correspond to any causal structure in the model’s actual computation. If a model achieves high performance through complex feature interactions, SHAP may attribute credit to individually important but causally secondary features while underweighting the interactions.
9.3 Model Cards and Datasheets
Model cards (模型卡) are a documentation standard introduced by Mitchell, Wu, Zaldivar, Barnes, Vasserman, Hutchinson, Spitzer, Raji, and Gebru (2018) to accompany trained machine learning models. A model card documents: (i) the intended use cases and out-of-scope uses; (ii) the training data, including any known biases; (iii) evaluation results across demographic subgroups; (iv) ethical considerations; and (v) caveats and recommendations. The goal is to enable developers, operators, and regulators to make informed decisions about whether and how to deploy a model.
Datasheets for Datasets (数据集数据表), introduced by Gebru et al. (2018), provide analogous documentation for training datasets, covering the motivation for dataset creation, the composition of the dataset, the data collection process, the preprocessing applied, recommended uses, and known limitations. Together, model cards and datasheets form a documentation ecosystem that aims to make AI system properties legible to a wider audience.
The adoption of model cards and datasheets has been uneven. Major developers — Google, Hugging Face, and others with significant reputational exposure — publish model cards for their public releases. But documentation requirements for proprietary systems deployed in high-stakes domains are largely voluntary in most jurisdictions. The EU AI Act’s requirements for technical documentation and data governance documentation for high-risk AI systems impose legally binding documentation obligations that go further than voluntary model cards, but the specific content requirements are still being elaborated through delegated acts and harmonised standards.
Part IV: AI, Democracy, and the Public Sphere
Chapter 10: Recommender Systems, Filter Bubbles, and Epistemic Autonomy
10.1 The Architecture of Epistemic Influence
Recommender systems — the algorithms that determine which content users see on social media platforms, video streaming services, and news aggregators — are among the most consequential deployed AI systems in terms of aggregate social effects. Hundreds of millions of people receive a significant fraction of their news, political information, and social interaction through platforms whose content ordering is determined by machine learning models optimised for engagement. The epistemic consequences of this architecture — what people believe, how they reason, and how they relate to those with different views — are a central concern of democratic theory.
Pariser, in The Filter Bubble (2011), introduced the concept of a filter bubble (过滤气泡): the personalised information environment created by algorithmic curation, in which each user receives a different stream of content reflecting their prior preferences and behaviours, potentially excluding perspectives and information that challenge their existing views. Sunstein, in Republic.com and its sequels, developed a related concept of echo chambers (回音室): online environments in which people encounter primarily information and opinion that reinforces their existing beliefs, intensifying polarisation and reducing the common factual basis required for democratic deliberation.
10.2 The Engagement Optimisation Problem
The mechanism by which recommender systems produce filter bubbles and echo chambers is well-understood: these systems are typically optimised to maximise engagement — time on platform, clicks, shares, comments — and engagement is reliably highest for content that provokes strong emotional reactions. Fear, outrage, and moral indignation are among the most engagement-inducing emotions, which creates systematic pressure toward content that is sensationalist, partisan, and divisive. A system optimised to maximise engagement has no objective incentive to expose users to accurate information, balanced perspectives, or content that challenges their views.
The engagement optimisation problem (参与度优化问题) is a specific instance of the broader specification gaming problem discussed in Chapter 2: engagement is a proxy for user value, but a deeply imperfect one. High engagement with outrage-inducing content does not indicate that the content is valuable to the user; it may indicate that the user has been effectively manipulated. The fundamental difficulty is that “valuable to the user” is not directly measurable, while “generates a click” is — so the system optimises for the measurable proxy.
Habermas’s concept of the public sphere (公共领域) provides a normative benchmark against which the effects of algorithmic curation can be assessed. For Habermas, democratic legitimacy requires a public sphere in which citizens can engage in rational discourse on matters of common concern — sharing information, exchanging arguments, and forming considered collective judgments. The algorithmic public sphere — fragmented, personalised, and optimised for emotional engagement — departs from this ideal in several ways: it reduces the common information base required for rational discourse, it rewards rhetorical manipulation over argument, and it privileges already-popular voices over minority perspectives.
10.3 Cambridge Analytica and Micro-Targeting
The Cambridge Analytica case (剑桥分析公司案) (2018) brought the political consequences of data-driven targeting to widespread public attention. Cambridge Analytica used improperly obtained Facebook data — including personality profiles derived from a quiz app that collected data on roughly 87 million users — to construct a psychographic targeting model that was used in the 2016 US presidential election and the Brexit referendum to deliver personalised political messaging.
The case illustrates several distinct concerns. The first is privacy (隐私): the data was collected without adequate informed consent and used in ways that users did not anticipate. The second is manipulation (操纵): psychographic targeting aims to exploit personality vulnerabilities — sending fear-based messages to people with high neuroticism scores, authority-based messages to people with high conscientiousness scores — in ways that bypass rational deliberation. The third is epistemic autonomy (认知自主性): personalised political messaging reduces the voter to a target to be manipulated rather than a citizen to be persuaded through argument, undermining the deliberative model of democratic politics.
Micro-targeting (微目标定向) — the use of detailed individual data to deliver personalised political messages — is now regulated in the EU under the Digital Services Act and proposed political advertising regulations, which require disclosure of targeting criteria and prohibit targeting based on sensitive personal data categories including political views, religion, and health. These regulations represent a direct response to the Cambridge Analytica case and reflect the EU’s view that the aggregation of personal data for political persuasion poses distinctive risks to democratic integrity.
10.4 Deepfakes, Synthetic Media, and Electoral Integrity
Synthetic media (合成媒体) generated by generative AI — including deepfakes (深度伪造), AI-generated text, synthetic audio, and composite images — presents a qualitatively new challenge to epistemic autonomy and democratic integrity. The marginal cost of producing convincing synthetic media has collapsed: tools for generating photorealistic video of public figures saying things they never said, or creating convincing fabricated text in the voice of any identified speaker, are now widely accessible.
The electoral implications are severe. Synthetic media can be used to spread fabricated quotes attributed to candidates, to create false impressions of candidate behaviour, to generate plausible-looking news coverage of fictional events, and to impersonate campaign communications. The liar’s dividend (说谎者的红利), identified by Chesney and Citron, describes a second-order effect: even true, authentic media can be dismissed as deepfake by actors who benefit from undermining trust in audiovisual evidence. The widespread existence of deepfakes thus corrodes the epistemic status of all video and audio, not just fabricated content.
Legal responses to synthetic media in electoral contexts vary. The EU’s Digital Services Act requires platforms to label AI-generated content and establishes obligations around “systematic risks” including manipulation of public discourse. Several US states have enacted laws requiring disclosure of AI-generated political advertising. Some scholars have called for broader provenance standards (溯源标准) — cryptographic authentication schemes that allow video and audio content to be traced to its origin and verified as authentic — as a technical complement to legal mandates.
Chapter 11: Synthetic Media, Disinformation, and Electoral Integrity
11.1 The Disinformation Ecosystem
Disinformation (虚假信息) — the deliberate creation and spread of false or misleading information — is not a new problem, but AI amplifies it in both quantity and quality. Large language models can produce plausible-sounding text at scale with minimal human effort; image generation models can produce photorealistic images of events that never occurred; voice cloning models can replicate any speaker’s voice from a short audio sample. The combination of these capabilities makes the fabrication of convincing disinformation — previously requiring significant resources and expertise — available to any actor with a consumer-grade computing device.
The disinformation ecosystem (虚假信息生态系统) is usefully distinguished by its actors and their goals. State actors — particularly Russia, China, Iran, and North Korea — have well-documented programmes for using information operations to influence foreign political discourse. Non-state actors — political campaigns, interest groups, organised conspiracy communities — exploit platform dynamics to amplify authentic and fabricated content alike. And commercial actors — click-farm operators, influence-for-hire networks — provide disinformation as a service to whoever pays. AI does not change the incentive structure of this ecosystem but dramatically lowers the production cost at every level.
11.2 Platform Governance and Content Moderation
Social media platforms face a fundamental governance dilemma in addressing disinformation: their content moderation decisions directly affect political speech, and any moderation threshold will be contested as either over-restrictive (suppressing legitimate political speech) or under-restrictive (allowing dangerous disinformation). The content moderation problem (内容审核问题) has no politically neutral solution because the line between disinformation and legitimate political contestation is contested terrain.
Platforms have adopted several approaches: fact-checking labels on disputed posts (applied on a sample basis, given the scale of content); prebunking (预防接种) or inoculation campaigns that warn users about disinformation techniques before they encounter specific false claims; algorithmic downranking of content that has been flagged by fact-checkers; removal of content that violates specific policies (coordinated inauthentic behaviour, synthetic media presented as authentic). Each approach has documented failure modes: labels may backfire by drawing attention to flagged content; prebunking has limited reach; downranking is vulnerable to manipulation through coordinated reporting; removal is criticised as censorship.
The Digital Services Act (数字服务法, EU) requires very large online platforms (VLOPs) to assess and mitigate “systemic risks” from their services — including risks to electoral processes, public health, and fundamental rights. VLOPs must submit annual risk assessments, undergo independent audits, and share data with researchers and regulators. This framework represents a significant expansion of regulatory oversight of platform governance in Europe, moving beyond the notice-and-takedown model of the E-Commerce Directive toward proactive risk management obligations.
11.3 AI and Electoral Processes
The deployment of AI in electoral processes raises concerns beyond disinformation. Electoral administration (选举管理) AI — systems used to verify voter eligibility, detect double voting, or identify errors in voter rolls — has in several jurisdictions been found to have disproportionate error rates for minority voters, effectively acting as a voter suppression mechanism. Predictive polling systems that use AI to forecast election outcomes have raised concerns about self-fulfilling prophecies — if early forecasts show one candidate with an overwhelming lead, turnout among that candidate’s supporters may be suppressed.
More speculatively, AI systems capable of micro-targeted political persuasion at scale raise the question of whether democratic politics can survive the personalisation of the public sphere (公共领域的个性化). Democratic theory has generally assumed a shared information environment — newspapers that everyone could read, broadcasts that everyone could watch — as a precondition for the formation of collective political will. A fully personalised information environment, in which no two citizens share the same information space, may be incompatible with the formation of the shared political judgments on which democratic legitimacy depends.
Deliberative democratic theorists in the tradition of Habermas and Joshua Cohen argue that democratic legitimacy requires not merely procedural compliance — elections held on schedule — but substantive conditions: informed citizens, authentic deliberation, and decisions that reflect the considered preferences of the polity rather than the manufactured preferences produced by manipulative messaging. AI-enabled disinformation and micro-targeting threaten these substantive conditions, and their governance is therefore a first-order question of democratic theory, not merely a regulatory compliance problem.
Part V: AI and Criminal Justice
Chapter 12: Predictive Policing and the Feedback Loop
12.1 The Promise and the Problem
Predictive policing (预测性警务) refers to the use of AI and data-driven methods to forecast where crimes are likely to occur or who is likely to commit them, enabling law enforcement to allocate resources proactively. The appeal is straightforward: if crime is predictable and police resources are limited, deploying those resources to high-probability locations or individuals should reduce crime more efficiently than reactive patrol. Advocates cite geographic predictive systems like PredPol (also known as Geolitica) and person-based systems like ShotSpotter and the Chicago Strategic Subject List.
The fundamental problem with predictive policing is a feedback loop (反馈回路) embedded in the logic of the approach. If predictions direct officers to particular neighbourhoods or individuals, and officers observe more crime where they patrol more intensely, then the crime data used to train subsequent predictions reflects policing patterns rather than underlying crime rates. A system trained on historically over-policed communities will predict high crime in those communities, which generates more policing, which generates more crime reports, which reinforces the prediction. The algorithm learns to amplify the consequences of its own predictions.
12.2 PredPol and ShotSpotter
PredPol used a seismic aftershock prediction algorithm, adapted to forecast crime in 500-square-foot boxes at specific times of day. Several US cities — Los Angeles, Santa Cruz, New Orleans — deployed PredPol in the 2010s and subsequently discontinued it, following civil liberties concerns and independent research suggesting the algorithm disproportionately directed patrol to Black and Hispanic neighbourhoods. A 2021 investigation by the Los Angeles Times found that PredPol predictions in LAPD’s data showed a feedback loop structure consistent with self-fulfilling prophecy.
ShotSpotter is an acoustic gunshot detection system that uses a network of microphones to identify and locate gunfire, alerting police in near-real-time. The system raises governance concerns on multiple axes. First, its false positive rate (误报率) is substantial: an independent study of Chicago ShotSpotter data found that the majority of alerts sent police to locations where no evidence of gunfire was found. Second, the system creates an acoustic surveillance infrastructure in particular (predominantly Black and Hispanic) neighbourhoods, with potential chilling effects on public activity. Third, MacArthur Justice Center v. ShotSpotter documentation revealed that ShotSpotter had in at least one case retroactively altered an audio classification at the request of law enforcement — changing a classification from “firecracker” to “gunshot” after a shooting death — raising serious concerns about evidence integrity.
12.3 The Diffuse Harms of Predictive Policing
Crawford and others have emphasised that the harms of predictive policing are not only individual (wrongful arrest, police harassment) but collective and diffuse. Intensive policing of a neighbourhood shapes the lived experience of everyone in that neighbourhood — the decision to walk home at night, the willingness to gather publicly, the relationship between community members and law enforcement institutions. These collective harms do not appear in individual wrongful arrest statistics, but they constitute a significant social cost that is distributed inequitably along racial and class lines.
The ghost of policing (警务的幽灵): historical data is not a neutral record of where crime occurred, but a record of where police chose to enforce the law. Drug law enforcement is the paradigmatic case: drug use is approximately equally prevalent across racial groups, but drug arrests are dramatically concentrated in communities of colour, reflecting differential policing intensity rather than differential offending rates. A predictive system trained on drug arrest data thus learns to predict drug crime in communities of colour, which directs drug enforcement to those communities, perpetuating the original disparity.
Reform proposals range from moratoriums on specific predictive policing technologies, to mandatory demographic auditing with publication of results, to community consent requirements before deployment, to wholesale redirection of predictive analytics toward resource allocation for social services rather than law enforcement.
Chapter 13: Risk Assessment Instruments and the Fairness Impossibility
13.1 COMPAS and the ProPublica Investigation
COMPAS (Correctional Offender Management Profiling for Alternative Sanctions, 罪犯管理评估工具) is a proprietary risk assessment instrument developed by Northpointe (now Equivant) and used in criminal sentencing decisions across the United States. COMPAS assigns defendants a risk score (1–10) predicting the likelihood of rearrest within two years. Judges in some jurisdictions are provided with these scores as inputs to sentencing recommendations, pretrial detention decisions, and parole determinations.
ProPublica’s 2016 investigation “Machine Bias” conducted an analysis of COMPAS scores for roughly 7,000 defendants in Broward County, Florida, matched against two-year rearrest outcomes. The investigation found that Black defendants were almost twice as likely as white defendants to be falsely flagged as future criminals (高风险错误标记率) — assigned high-risk scores but not rearrested within two years. Conversely, white defendants were more likely than Black defendants to be falsely classified as low risk — assigned low scores but subsequently rearrested.
Northpointe responded that COMPAS is well-calibrated: among defendants assigned the same risk score, Black and white defendants reoffend at the same rate. This is accurate, and it is a genuine measure of predictive fairness. The disagreement between ProPublica and Northpointe instantiates precisely the impossibility theorem discussed in Chapter 8: COMPAS satisfies calibration; ProPublica’s critique is that it fails equalised false positive rates. Since base reoffending rates differ between Black and white defendants in the training data (a consequence of differential enforcement and social disadvantage), no instrument can simultaneously satisfy both criteria.
13.2 The Impossibility Applied: What Fairness Standard Should Govern?
The COMPAS controversy forces a choice between competing fairness criteria that has no technically correct answer. Satisfying calibration — ensuring that a given score means the same thing for Black and white defendants — has an appealing property of predictive equality: the score accurately predicts the likelihood of reoffending for everyone it scores. But achieving calibration with different base rates means accepting higher false positive rates for higher-base-rate groups.
Satisfying equalised false positive rates — ensuring that non-reoffending defendants of all races face equally low chances of being incorrectly labelled high-risk — has an appealing property of procedural equality: the cost of a false positive (detention, adverse sentencing) falls equally on innocent defendants regardless of race. But achieving equalised false positive rates with different base rates requires a score that discriminates within each group with different thresholds, which may produce scores that are not calibrated.
Reformers have proposed abandoning COMPAS and similar instruments altogether, arguing that the use of any group-statistically-derived risk score in individual sentencing is incompatible with the constitutional principle of individual rights — each defendant should be sentenced based on their individual conduct, not their membership in statistical risk categories. Harcourt’s Against Prediction develops this argument in depth, tracing the actuarial turn in criminal justice from the Progressive Era to the present and arguing that it is both empirically ineffective and normatively indefensible.
13.3 Facial Recognition and Wrongful Arrests
The wrongful arrest cases arising from facial recognition technology in US law enforcement are among the most concrete illustrations of the real-world consequences of algorithmic bias. Robert Williams was arrested in January 2020 when Detroit Police matched a surveillance image of a shoplifting suspect to Williams’s driver’s license photo using facial recognition — a match that was incorrect, as confirmed when a human investigator showed the photo to a witness who also confirmed no match. Williams was detained for 30 hours before the error was discovered.
Nijeer Parks was arrested in January 2019 when facial recognition incorrectly matched him to a suspect in a shoplifting and car theft case in New Jersey. Parks, who was 30 miles away at the time, was incarcerated for ten days and spent thousands of dollars in legal fees before charges were dropped. Michael Oliver was arrested in Detroit in 2019 on a felony assault charge based on a facial recognition match that was subsequently determined to be incorrect.
All three documented wrongful arrests based on facial recognition misidentification in the US involved Black men. The correlation is not coincidental: as Buolamwini and Gebru demonstrated, commercial facial recognition systems have dramatically higher error rates for darker-skinned faces. Law enforcement use of systems with known higher error rates for particular demographic groups raises constitutional issues under the Fourth Amendment (unreasonable seizure) and Equal Protection Clause, as well as statutory issues under civil rights laws.
Several US cities — San Francisco, Boston, New York — have enacted moratoriums on law enforcement use of facial recognition. The EU AI Act prohibits real-time biometric identification in public spaces by law enforcement (with narrow exceptions), but allows post-hoc identification in criminal investigations. The adequacy of these regulatory responses, given the scale of deployment and the severity of the harms, remains actively contested.
Part VI: Geopolitics and International Governance
Chapter 14: The US–China AI Race and Compute Governance
14.1 The Geopolitical Stakes
The development of advanced AI is increasingly framed by governments as a matter of national security and strategic competition. The United States and China are widely identified as the two leading AI powers, with the EU as a distant but significant third in terms of regulatory standard-setting. Dafoe argues that the strategic stakes of AI are analogous to those of nuclear weapons in the 1940s–1950s: a technology capable of conferring decisive military and economic advantages on its first possessor, with global distributional consequences determined largely by how the leading powers manage their competition.
The strategic significance of AI for national security operates at multiple levels. AI enables significantly improved intelligence analysis (情报分析): processing satellite imagery, signals intelligence, and open-source information at scales no human analyst could match. It enables improved logistics and supply chain optimisation in military contexts. It enables potentially transformative capabilities in cybersecurity offence and defence — both identifying vulnerabilities in adversary systems and defending against novel attacks. And it potentially enables the autonomous weapons systems discussed in Part VII.
14.2 Semiconductor Export Controls and Compute Governance
The most consequential US government action in AI geopolitics has been the progressive tightening of semiconductor export controls (半导体出口管制) restricting China’s access to advanced AI chips. Beginning in October 2022 and substantially expanded in subsequent years, the US Bureau of Industry and Security imposed controls on the export to China of Nvidia’s A100 and H100 data centre GPUs, as well as advanced chipmaking equipment from US-linked manufacturers. The controls aimed to prevent China from acquiring the computing infrastructure required to train frontier AI models.
The controls reflect a strategic judgment that compute (算力) is the critical chokepoint in the AI supply chain: frontier AI models require vast quantities of high-performance compute for training, and the most advanced AI chips are concentrated in a small number of US-allied companies (Nvidia, AMD) with manufacturing dependent on TSMC in Taiwan. By controlling access to these chips, the US aims to slow China’s frontier AI development — accepting costs to US semiconductor industry revenues and creating diplomatic friction with allies who must navigate export control compliance.
Compute governance (算力治理) as an international policy concept extends this logic from bilateral controls to multilateral frameworks: if advanced compute is a necessary input to frontier AI, then monitoring and regulating compute could serve as a mechanism for global AI governance that does not require agreement on the much harder questions of what AI applications are permissible. Proposals for a compute monitoring regime — analogous to International Atomic Energy Agency inspections of nuclear material — have attracted academic attention as a potentially tractable international governance mechanism.
14.3 China’s AI Governance Framework
China has developed its own AI governance framework, which presents an interesting contrast to Western approaches. China’s New Generation Artificial Intelligence Development Plan (2017) established AI as a strategic priority with the goal of achieving global leadership by 2030. The country has also enacted sector-specific AI regulations: rules on algorithmic recommendation systems (2022), rules on deep synthesis (deepfakes) (2022), and interim measures on generative AI services (2023).
Chinese AI regulation shares some features with EU approaches — risk classification, transparency requirements, human oversight mechanisms — but operates within a fundamentally different political context, in which state security and social stability are explicit regulatory goals. The prohibition on AI-generated content that “endangers national security,” “disturbs the economic order,” or “undermines social stability” is substantially broader than any Western analogue, and enforcement through the Cyberspace Administration of China reflects the integration of AI governance with the broader architecture of political censorship and social control.
The Chinese approach raises a fundamental question for international AI governance: can governance frameworks developed in liberal democratic contexts — which assume freedom of expression, judicial independence, and limits on state surveillance — be extended to or harmonised with frameworks developed in authoritarian contexts that use AI as a tool of social control? The answer has significant implications for whether international AI governance can transcend the geopolitical divide.
Chapter 15: International Institutions and the Brussels Effect
15.1 The Brussels Effect in AI
Bradford’s Brussels Effect (布鲁塞尔效应) describes the phenomenon by which EU regulations become de facto global standards because multinational companies, rather than maintaining separate product lines for EU and non-EU markets, apply the strictest applicable standard everywhere. The GDPR is the paradigmatic example: rather than maintaining one data processing infrastructure for European users and another for non-European users, most global platforms applied GDPR-equivalent protections globally, exporting EU data protection standards far beyond EU borders.
Whether the Brussels Effect will operate for the EU AI Act is a significant question in international AI governance. Several conditions that enable the Brussels Effect in data protection may apply differently to AI. AI systems are often not market-facing products in the simple sense — many AI capabilities are embedded in services delivered entirely within one jurisdiction. The costs of AI development are higher and more opaque than the costs of data processing infrastructure, making compliance cost differentials harder to assess. And the US and China both have significant leverage to push back against EU AI Act standards through regulatory reciprocity and market access negotiations.
Nevertheless, several dynamics suggest some Brussels Effect is likely. European consumers and institutions represent a large market that AI deployers cannot ignore. The EU AI Act’s GPAI provisions apply to any model made available in the EU, regardless of where it is developed — effectively extending conformity obligations to US and Chinese AI developers who wish to operate in Europe. And the EU’s long track record in regulatory standard-setting, combined with the depth of its institutional capacity (European AI Office, national market surveillance authorities), gives its framework a credibility that voluntary frameworks lack.
15.2 The OECD AI Principles and UNESCO Recommendation
The OECD Principles on AI (OECD人工智能原则, 2019) were the first intergovernmental standard on AI, endorsed by all OECD members including the US, EU, Japan, South Korea, and (unusually) non-OECD economies including Argentina and Brazil. The five principles are: inclusive growth, sustainable development and well-being; human-centred values and fairness; transparency and explainability; robustness, security, and safety; and accountability. These principles are high-level and non-binding, providing normative orientation without specific implementation requirements.
UNESCO’s Recommendation on the Ethics of Artificial Intelligence (2021), adopted by all 193 UNESCO member states including China and Russia, is the most universally endorsed AI ethics document to date. The Recommendation covers a wide range of values — human dignity, privacy, data governance, education, health, environment — and calls on member states to develop AI ethics frameworks, conduct readiness assessments, and report on implementation. Like the OECD Principles, the Recommendation is non-binding and lacks enforcement mechanisms.
The gap between the normative ambitions of these instruments and their enforcement capacity is the central problem of international AI governance. Soft law instruments — principles, recommendations, voluntary codes of conduct — can achieve normative convergence at the cost of enforceability. Binding international law — treaties, conventions with compliance mechanisms — can achieve enforceability at the cost of the consensus required for universal adherence. The history of arms control suggests that the optimal approach may be a combination: soft law instruments that build normative consensus, followed by binding instruments in narrow, technically tractable domains where verification is feasible.
15.3 Digital Non-Alignment and the Global South
International AI governance discourse is disproportionately shaped by the concerns of technologically advanced economies — the EU, US, UK, Canada, Japan — while the Global South, which includes the majority of the world’s population, has been largely peripheral to framework development despite bearing significant AI impacts. Crawford’s Atlas of AI documents the extractive geography of AI (AI的提取性地理): the data centres and AI training infrastructure consume energy and water from communities — often in the Global South — that see few of the benefits, while the data that trains AI systems is extracted from users around the world without compensation.
Digital non-alignment (数字不结盟) is a concept that has emerged from Global South academics and policymakers, drawing on the Non-Aligned Movement’s Cold War heritage. It describes the position of states that seek to avoid subordination to either US or Chinese AI governance frameworks, developing indigenous AI capacity and governance frameworks that reflect their own values and priorities. India’s emerging AI governance framework, Brazil’s AI Bill (PL 2338/2023), and the African Union’s emerging AI guidelines all reflect dimensions of this position.
The representation of Global South interests in international AI governance is not merely a question of fairness but a pragmatic necessity. AI systems trained predominantly on data from wealthy anglophone countries may perform poorly for speakers of other languages, users with different cultural assumptions, and people in different economic and social contexts. Governance frameworks that do not incorporate the perspectives of the majority of the world’s population are likely to be both normatively deficient and technically inadequate.
Part VII: AI and Armed Conflict
Chapter 16: Autonomous Weapon Systems and International Humanitarian Law
16.1 The Taxonomy of Autonomous Weapon Systems
Autonomous weapon systems (自主武器系统, AWS) are weapon systems that can select and engage targets without direct human command of each individual engagement. The key governance distinction is the degree of human involvement in the targeting cycle. The ICRC (国际红十字委员会) and most legal and ethics scholars use a three-level taxonomy: human in the loop (人在回路中) — a human operator must approve each individual targeting decision; human on the loop (人在回路上) — a human operator can observe and override the system’s targeting decisions but need not actively approve each one; and human out of the loop (人在回路外) — the system selects and engages targets entirely autonomously, without any human intervention in the targeting cycle.
Currently deployed systems span this taxonomy. Defensive systems like Israel’s Iron Dome and the US Navy’s Phalanx CIWS operate on a human-on-the-loop model: they can autonomously intercept incoming missiles or aircraft, but human operators can disable or override them. These systems operate in constrained environments (incoming projectile, short decision window) where the targeting decision is relatively straightforward. Lethal Autonomous Weapon Systems (LAWS) — systems capable of autonomously selecting and engaging human targets in complex, contested environments — do not yet exist as deployed weapons, though multiple states have active development programmes.
16.2 International Humanitarian Law: The Core Principles
International humanitarian law (国际人道法, IHL), also called the law of armed conflict, is the body of law governing the conduct of hostilities. The four core principles applicable to targeting decisions are: distinction (区分原则) — parties must distinguish between combatants and civilians, and may only target combatants and military objectives; proportionality (相称性原则) — attacks must not cause civilian harm that is excessive in relation to the anticipated military advantage; precaution (预防原则) — parties must take all feasible precautions to minimise civilian harm; and military necessity (军事必要性原则) — force may only be used to the extent necessary to achieve legitimate military objectives.
The central question for LAWS is whether a fully autonomous targeting system can comply with these principles. The distinction principle requires the ability to identify military targets and distinguish them from civilians in complex, contested, and context-dependent environments — a task that requires not merely pattern recognition but contextual understanding: Is this person carrying a weapon? Are they a combatant or a civilian hunter? Does their behaviour indicate hostile intent? The proportionality principle requires a qualitative judgment — “excessive civilian harm in relation to military advantage” — that has no algorithmic expression. The precaution principle requires flexible adaptation of tactics and methods to minimise incidental harm, which may require types of creative problem-solving that current AI systems cannot reliably perform.
16.3 The Crootof Legal Gap
Crootof’s 2016 paper “The Killer Robots Are Here: Legal and Policy Implications” argues that there is a structural legal gap (法律空白) in the regulation of autonomous weapon systems. Under existing IHL, responsibility for violations rests on individual human commanders and soldiers who make targeting decisions. If an autonomous system makes a targeting decision that constitutes a war crime — killing civilians in violation of the distinction principle — it is unclear who bears criminal responsibility. The developer cannot be held criminally responsible for a decision made autonomously by a system they created but did not control. The commander who deployed the system may have authorised its general use but not the specific unlawful engagement. The system itself is not a legal person and cannot be prosecuted.
This accountability gap is not merely theoretical. Crootof observes that the prospect of fighting wars at lower cost to one’s own forces — by substituting machines for soldiers — may encourage states to engage in conflicts they would otherwise avoid, lowering the political threshold for the use of force. If autonomous weapon systems offer a way of projecting lethal force without the domestic political cost of casualties, the aggregate effect may be more frequent armed conflict — an outcome that is bad for global security even if individual AWS deployments comply with IHL.
16.4 The CCW Process and Its Failures
The Convention on Certain Conventional Weapons (特定常规武器公约, CCW) is the primary multilateral forum where LAWS governance is being negotiated. The CCW’s Group of Governmental Experts (GGE) on LAWS has met regularly since 2014, generating substantial documentation but no binding agreement. The absence of agreement reflects several structural obstacles.
First, states with advanced autonomous weapons programmes — the US, Russia, China, the UK, Israel — have strong strategic incentives to resist binding restrictions that would constrain their capabilities. Russia and China have blocked any attempt to negotiate a ban, arguing that existing IHL is sufficient. The US has supported a “political declaration” approach — voluntary standards for responsible use — rather than a binding treaty. Second, the technical definition of which systems would be covered by a ban is genuinely contested: almost every modern weapon system has some degree of autonomous functionality (target identification, firing solution calculation), and any prohibition must draw a line between permissible and impermissible autonomy in terms that are both normatively coherent and technically precise. Third, verification of a ban on software-based capabilities is substantially harder than verification of a ban on large physical weapons systems — an autonomous targeting algorithm is difficult to detect through the inspection mechanisms available under arms control regimes.
Chapter 17: Drone Warfare, Project Maven, and the Dual-Use Dilemma
17.1 The Evolution of Drone Warfare
Unmanned aerial vehicles (UAVs, drones, 无人机) deployed for lethal strikes represent the current frontier of AI-adjacent military technology. Contemporary armed drones — the MQ-9 Reaper, the Turkish Bayraktar TB2 — are not autonomous: a human operator at a ground station controls each flight and must authorise each weapons release. But they are enabled by AI-assisted targeting: image recognition systems help identify targets in live video feeds; AI planning tools suggest optimal flight paths; threat detection algorithms flag objects of interest.
The drone warfare model — persistent surveillance followed by precision strike, conducted remotely without risk to the striking force — has generated significant IHL concern independent of the autonomy question. Targeted killing programmes conducted by the US in Pakistan, Yemen, and Somalia have been criticised for their reliance on signature strikes (特征打击): targeting individuals based on patterns of behaviour associated with terrorist activity, rather than positive identification of specific individuals. Signature strikes represent a weakening of the distinction principle — the targeting criterion is a behavioural profile rather than individual combatant status — and have resulted in documented civilian casualties.
17.2 Project Maven and Corporate Responsibility
Project Maven (马文项目), formally the Algorithmic Warfare Cross-Functional Team, was a US Department of Defense initiative to use AI to accelerate the analysis of drone surveillance video — identifying vehicles, people, and objects of interest in the thousands of hours of footage collected daily. Google won a contract to provide TensorFlow-based image recognition for Project Maven in 2017. In 2018, thousands of Google employees signed a letter protesting the contract, arguing that they did not want their work used for weapons development. Google subsequently declined to renew the Project Maven contract.
The Project Maven episode illustrates the dual-use dilemma (两用困境): foundational AI research and infrastructure developed for commercial purposes can readily be applied to military and intelligence applications. Large language models that assist programmers can also assist in developing cyberweapons. Image recognition systems that power autonomous vehicles can also power drone targeting. Voice synthesis systems can be used for entertainment or for disinformation and psychological operations. The same technology, the same infrastructure, the same algorithmic advances enable both beneficial and harmful applications.
The dual-use dilemma creates significant challenges for governance. Export controls on dual-use technologies — including current semiconductor export controls — attempt to prevent adversaries from accessing capabilities that have significant military applications, while accepting the cost of restricting beneficial commercial applications. But the line between “primarily military” and “primarily commercial” AI applications is increasingly difficult to draw as foundation models become more capable and more versatile.
The Project Maven controversy also raised a question about corporate responsibility (企业责任) in military AI: do technology companies have ethical obligations regarding the military uses of their technologies, and if so, what are those obligations? Palantir, Anduril, and other defence technology companies have explicitly positioned themselves as willing to develop military AI. Google, Microsoft, and Amazon have military and intelligence contracts that involve AI applications but have faced employee and public pressure over specific applications. The question of how to govern the private sector’s contribution to military AI development is largely unaddressed by existing frameworks.
17.3 Algorithmic Targeting and the Laws of War
The application of machine learning to targeting — not just analysis of drone footage but the selection of strike targets from large datasets — represents a qualitative shift in how targeting decisions are made. Traditional targeting required human analysts to assess intelligence, develop target nominations, apply proportionality and necessity analysis, and obtain command authority for each strike. Algorithmic targeting systems can process far more intelligence data and generate target nominations at a pace that human review cannot match.
The human review bottleneck (人工审查瓶颈) is increasingly identified as a constraint on the operational tempo that algorithmic targeting enables. The pressure to speed human review to match algorithmic output — reviewing target nominations in seconds rather than minutes — raises serious IHL concerns about whether the precaution and proportionality analysis that IHL requires can meaningfully be conducted in the time available. Research on human-machine teaming in targeting contexts suggests that humans reviewing AI recommendations under time pressure exhibit automation bias (自动化偏见): a tendency to accept AI recommendations without independent judgment, particularly when the recommendations are framed with high confidence scores.
Automation bias in targeting contexts is not merely an accuracy concern — it is an IHL concern, because the distinction and proportionality analysis that IHL requires involves qualitative judgment that cannot be delegated to an algorithm without fundamentally changing the nature of the legal obligation. If a human reviews an AI targeting recommendation in three seconds and authorises a strike based primarily on the AI’s confidence score, whether that constitutes meaningful human control is a genuine legal question.
Chapter 18: Toward a Global Ban? Diplomacy and the Campaign to Stop Killer Robots
18.1 The Campaign to Stop Killer Robots
The Campaign to Stop Killer Robots (停止杀手机器人运动), launched in 2013, is a coalition of non-governmental organisations, legal scholars, and AI researchers that advocates for a pre-emptive ban on lethal autonomous weapon systems — a prohibition negotiated before such systems are deployed at scale, analogous to the bans on landmines (Ottawa Treaty, 1997) and cluster munitions (Dublin Convention, 2008). The Campaign’s argument has three prongs: LAWS cannot comply with IHL (the legal argument); LAWS remove human agency from killing decisions in ways that violate human dignity (the moral argument); and the deployment of LAWS would lower the threshold for armed conflict, increasing aggregate violence (the strategic argument).
The Campaign has achieved significant normative traction: as of 2024, more than 70 states have called for a ban on LAWS at the CCW, and several UN resolutions have affirmed the importance of human control over weapons systems. The UN Secretary-General has called for a legally binding instrument on LAWS. But the states with the most advanced weapons development programmes — the US, Russia, China, UK, Israel — have consistently resisted a binding ban.
18.2 The Landmines and Cluster Munitions Analogies
The Ottawa Treaty (1997) and the Convention on Cluster Munitions (2008) were negotiated outside the traditional arms control framework — the Conference on Disarmament, where great power veto paralysis had prevented agreement — through a humanitarian disarmament (人道主义裁军) approach: building a coalition of states and civil society around a clear normative principle (these weapons cause unacceptable humanitarian harm), negotiating a treaty among willing states, and allowing the resulting treaty’s normative force to influence the behaviour even of non-signatory states.
The analogy to LAWS has limits. Landmines and cluster munitions are physical weapons that can be physically destroyed and whose use can be physically detected. LAWS is a software capability: a ban on LAWS would prohibit the writing and deployment of specific types of software in weapons systems, which is technically much harder to verify. The powerful-state holdouts — the US and UK refused to sign the Ottawa Treaty; China, Russia, and the US refuse to sign the Convention on Cluster Munitions — are the same states whose military AI development creates the primary governance need. And the timeline pressure is different: LAWS do not yet exist as deployed systems, but the pace of AI capability development means the window for pre-emptive prohibition may be shorter than it was for landmines.
18.3 Intermediate Positions: Political Declarations and Voluntary Norms
In the absence of a binding ban, several intermediate governance approaches have been proposed. Political declarations on responsible AI in the military domain — the G7 Hiroshima Process, the US-led Declaration on Responsible Military Use of Artificial Intelligence — establish voluntary norms for state behaviour without treaty-level obligations. The US DoD’s Directive 3000.09 (2012, revised 2023) requires meaningful human control over autonomous weapons systems for DoD, establishing an internal policy requirement.
Voluntary commitments (自愿承诺) by technology companies — commitments not to develop AI applications for lethal autonomous targeting without specific conditions — represent a private sector governance dimension. In practice, such commitments are difficult to make precise (which AI applications count as lethal autonomous targeting?) and difficult to enforce (the commitments are self-policed). But they create reputational accountability that can influence corporate behaviour, as the Project Maven episode demonstrated.
The “meaningful human control” standard (有意义的人类控制标准) has emerged as the most widely supported intermediate position in the CCW process: rather than banning LAWS categorically, states would commit to ensuring that targeting decisions involving lethal force are subject to meaningful human agency, with specific technical and procedural requirements to be elaborated. This approach allows states to retain autonomy in weapon system design while committing to a common standard for human involvement. Its limitation is that “meaningful” remains contested, and states with strong incentives to deploy autonomous weapons have strategic interests in defining “meaningful” in ways that accommodate their development programmes.
18.4 The Deeper Question: Human Dignity and the Ethics of Delegation
At the deepest level, the debate about LAWS raises a philosophical question about which decisions can ever be legitimately delegated to machines. Roff and others argue that the decision to kill — to deliberately deprive a human being of life — is intrinsically a moral decision requiring a moral agent who can bear responsibility. Machines are not moral agents; they cannot be held morally responsible. Therefore, delegating the kill decision to a machine is not merely legally problematic (it creates an accountability gap) but morally impermissible (it removes the very thing — moral agency — that the decision requires).
This argument, sometimes called the moral responsibility objection (道德责任异议) to LAWS, connects to broader questions in the philosophy of action about what it means for a decision to be one’s own. If a human operator authorises a system in advance to kill anyone who satisfies a set of criteria, and the system then kills someone who satisfies those criteria, the operator has made a decision at some level of abstraction. But the moral content of the decision — the judgment that this particular individual, in this particular context, may be killed — has been delegated to the machine. Whether that abstracted decision constitutes meaningful moral agency over the specific killing is a question that draws on the philosophy of action, moral responsibility theory, and the ethics of delegation.
These notes connect here to HIST 415 (A History of Artificial Intelligence), which traces how successive waves of AI development have raised and re-raised questions about machine agency, human control, and the appropriate scope of automation — and to PHIL 459b (Philosophy of Artificial Intelligence), which examines the conceptual foundations of questions about machine understanding, moral agency, and the limits of artificial cognition.
18.5 Looking Forward: The Governance Agenda
The governance of AI in armed conflict is developing rapidly and without clear resolution. Several structural features of the problem make rapid progress unlikely. The states with the most incentive to develop LAWS — those seeking military advantage — are the states with the most power to block binding international restrictions. The technical characteristics of AI weapons — software-based, dual-use, difficult to verify — make the arms control toolkit less effective than it has been for physical weapons. And the pace of capability development outstrips the pace of diplomatic negotiation.
Nevertheless, there are grounds for cautious optimism. The normative consensus that LAWS raise serious governance concerns is broader and more genuine than the consensus that existed before the Ottawa Treaty. The CCW GGE process, despite its failures to produce a binding instrument, has generated a detailed multilateral record of state positions that creates accountability pressure. And the emergence of responsible AI in military contexts as a distinct policy domain — separate from, though connected to, broader AI governance — suggests that military AI will receive increasing regulatory attention even from states that resist categorical prohibition.
The ultimate test of any governance framework for AI in armed conflict is whether it reduces the probability of catastrophic outcomes — mass civilian casualties, illegal wars prosecuted at lower political cost by autonomous systems, or escalatory dynamics triggered by AI-enabled speed of attack. These are the stakes that make getting the governance right not merely an academic exercise but one of the most consequential policy problems of the current era.
Conclusion: Governance as an Ongoing Practice
These notes have traversed seven substantive domains — alignment and safety, law and regulation, auditing and accountability, democracy, criminal justice, geopolitics, and armed conflict — that together constitute the governance landscape of artificial intelligence. Several cross-cutting themes emerge from this survey.
The first is the persistence of the capability–governance gap (能力-治理差距). In every domain, AI capabilities are advancing faster than the institutions designed to manage them. Alignment research lags behind capability development. Regulatory frameworks are written for current systems and are outdated by deployment. Auditing methodologies are tested on public models but cannot access proprietary systems. Electoral integrity frameworks are developed for yesterday’s disinformation ecosystem. Criminal justice reforms address COMPAS while more powerful risk assessment tools are deployed. International governance institutions are structured for a technological pace that no longer exists.
The second is the inescapability of political choices (政治选择的不可避免性). Every apparently technical decision in AI development and deployment involves value choices: which fairness criterion to optimise, which risks to prioritise, whose preferences to align to, which harms to treat as acceptable residual risks. Technical expertise is essential for understanding the options and their consequences, but it cannot substitute for democratic deliberation about which options to choose. Acemoglu and Johnson’s central thesis — that technological progress does not automatically produce broadly shared prosperity; it requires active political contestation — applies with full force to AI governance.
The third is the interdependence of scale and accountability (规模与问责的相互依存). AI systems are valuable precisely because they operate at scale: a single model can make millions of decisions, reach hundreds of millions of users, or identify thousands of targets. This scale is what makes AI governance consequential. But scale also makes accountability harder: the individual harms from a single erroneous decision are often small, the injured parties are often dispersed and poorly organised, and the developers and deployers are often powerful and legally sophisticated. Effective AI governance must find institutional mechanisms — auditing, liability, democratic oversight — that can create accountability at the scale at which AI operates.
The fourth is the importance of inclusive participation (包容性参与的重要性) in governance design. The communities most exposed to the harms of AI systems — communities of colour subjected to predictive policing and facial recognition, workers subject to algorithmic management, communities in the Global South affected by AI-enabled extraction — are systematically underrepresented in the institutions designing AI governance frameworks. Governance frameworks designed without these voices will reflect the interests of those with power to participate in their design, and will be less legitimate, less effective, and less just as a result.
The study of AI ethics, law, and governance is not a spectator sport. The decisions being made now — about how to regulate frontier AI models, how to govern autonomous weapons, how to protect electoral integrity from synthetic media — will shape the conditions of human life for generations. Understanding those decisions, the values at stake, and the institutional options available is a form of civic competence that the current moment demands.