Big: Long Complex
The algocratic tightrope will not be walked by any single institution. It will be walked by millions of small decisions: a researcher choosing to publish safety benchmarks, a company refusing a contract, a regulator updating a benchmark, a citizen insisting on transparency. That is not a solution. It is, perhaps, the only thing that has ever been. Word count: ~1,800 (abridged from full-length target). Full-length version would include case studies (Tay, Zillow, COMPAS, Clearview), economic models (compute thresholds as Pigouvian taxes), and extended legal analysis (First Amendment vs. algorithmic speech).
Example: In 2022, a major AI company certified that its recommendation algorithm was “fair” under a state law, using a proprietary metric. An independent audit later found that the metric ignored exactly the kinds of disparate impact the law was designed to prevent. The company was legally compliant and dangerously unfair. If a country imposes strict AI safety rules, frontier development will move elsewhere. This is not speculation—it is history. When the US tightened biotech regulations in the 1970s, research moved to the UK. When the EU enforced strict data localization, cloud providers opened data centers in Ireland. Today, if the US bans training runs above a certain FLOP threshold, a Chinese or Middle Eastern state-funded lab will simply ignore it. The risk does not disappear; it relocates to jurisdictions with weaker institutions, less transparency, and potentially fewer scruples. BIG LONG COMPLEX
I. Introduction: The New Leviathan In 2023, over 1,000 tech leaders and researchers signed an open letter comparing the risks of artificial intelligence to those of pandemics and nuclear war. That same year, the European Union passed the world’s first comprehensive AI Act—a 400-page document classifying AI systems by risk level. Within months, ChatGPT, the poster child of generative AI, was banned in Italy, reinstated, and then faced 13 separate complaints across EU member states. Meanwhile, in the United States, the White House secured voluntary commitments from seven AI companies, while China implemented mandatory security reviews for “generative AI services with public opinion characteristics.” The algocratic tightrope will not be walked by
Thus, the case for regulation is compelling. But compelling does not mean feasible. A. The Opacity of Black Boxes Regulation requires measurement. Measurement requires interpretability. Modern deep learning models are famously inscrutable. A neural network with hundreds of billions of parameters does not have “rules” an inspector can audit. It has weights—floating-point numbers that correlate with no human-understandable concept. When the EU AI Act demands transparency for “high-risk systems,” it assumes that a developer can explain why a model made a particular decision. For transformer architectures, this is often false. Explainability methods (LIME, SHAP, attention visualization) are post-hoc approximations, not ground truth. As one MIT researcher put it: “Asking why a neural network made a decision is like asking why a cloud looks like a rabbit. You can always find a story, but it’s not causation.” B. Regulatory Lag and AI Speed The typical regulatory cycle—problem identification, study, stakeholder comment, rule drafting, legal challenge, implementation, enforcement—takes 5–10 years. AI model generations take 3–6 months. GPT-3 to GPT-4 was 24 months. GPT-4 to GPT-5 is estimated at 12–18 months. By the time a law takes effect, the technology it governs no longer exists. This is the Red Queen problem: you have to run twice as fast just to stay in place. It is, perhaps, the only thing that has ever been
The 2024 US Executive Order on AI attempts to address this via export controls on AI chips. But chips are physical; models are not. A company can train a model in a regulated jurisdiction, then copy the weights to an unregulated one. Once released, the model is immortal. No border patrol can stop mathematics. A. The Centralization Trap Most proposed regulations (compute thresholds, licensing requirements, mandatory reporting) disproportionately affect smaller players. A compliance burden that is trivial for Google or Microsoft is fatal for a university lab or a startup. The result is a regulatory moat: incumbents capture the state, and the state reinforces incumbents. This reduces the diversity of AI development, which is precisely what safety advocates want to avoid—diverse actors are harder to coordinate, but they also produce more innovation in safety techniques. Centralization creates monoculture, and monocultures are fragile. B. The Safety-Washing Loophole Regulation incentivizes box-checking, not risk reduction. When the EU AI Act requires “risk management systems,” companies will hire armies of compliance consultants to produce documents that look like safety. But genuine safety research—adversarial robustness, mechanistic interpretability, formal verification—is expensive and slow. Regulation creates a market for the appearance of safety, not safety itself. This is known as Goodhart’s law: when a measure becomes a target, it ceases to be a good measure.
The most dangerous AI is not the one developed in San Francisco. It is the one developed in a country with no media, no civil society, and no rule of law. If traditional regulation is too slow, too blunt, and too easily gamed, what remains? Several unconventional approaches are emerging. A. Differentiated Responsibility Instead of regulating the model, regulate the deployment context . A model that controls a power grid requires different oversight than a model that summarizes emails. This shifts the burden from developers to deployers, who are often easier to identify and sanction. It also aligns incentives: the company selling an AI for autonomous driving is better positioned to test for safety than the company that trained the base model. The base model is a toolkit; the deployment is a weapon. B. Dynamic Safety Licensing Rather than static laws, create a regulatory API. The UK’s proposed AI Safety Institute would operate as a technical body, not a legislative one. It would publish real-time safety benchmarks, red-team frontier models, and issue “safety passes” that expire after six months. Regulators then enforce against the absence of a pass, not against specific technical features. This turns the problem from “predict every risk” to “verify continuous compliance.” It is faster, more adaptive, and harder to game—because the benchmark can change without a new law. C. Liability Without Regulation The common law tradition offers a lighter touch: keep existing rules (negligence, product liability, nuisance) and apply them to AI. If an AI system causes harm, the deployer pays damages. This creates a financial incentive for safety without prior restraint. The drawback: liability requires a harm to occur first. For existential risks, that is too late. But for most AI risks—bias, fraud, physical injury—tort law is surprisingly adequate. D. Technical Countermeasures Over Legal Ones Finally, we must acknowledge that the most effective constraints on AI may not be legal at all. Cryptographic model signing, zero-knowledge proofs for model provenance, watermarking of synthetic content, and decentralized auditing protocols—these are tools that work at machine speed, not legislative speed. They do not require consent; they require code. The EU’s Digital Services Act already hints at this, requiring platforms to label AI-generated images. But the next step is automated enforcement: AI systems that detect other AI systems, without human intermediaries.