Superagency cover

Superagency

by Reid Hoffman And Greg Beato

An overview of the pluses and minuses of artificial intelligence.

A Techno-Humanist Path to Superagency

How do you grow human freedom as AI power rises? In this book, Reid Hoffman and collaborators argue that the most reliable way to expand liberty is to expand agency—your capacity to set goals and act effectively—and to do it through a pragmatic method they call iterative deployment. Rather than ban or blindly accelerate, you ship early, listen widely, and adapt fast. The thesis is techno-humanist: give people hands-on access to AI, build guardrails through testing and governance, and align the technology to human purposes through use, not theory.

The authors contend that agency is the core lens for every major AI debate—from job disruption and misinformation to national strategy. They frame today’s large language models (LLMs) as powerful but probabilistic assistants whose strengths (fluency, speed, breadth) and weaknesses (hallucination, bias, opacity) must be understood clearly. They also reframe data platforms as private commons—privately run but publicly valuable—arguing for governance that protects creators and privacy while preserving the fertile “data agriculture” that yields massive consumer surplus.

Why iteration beats abstraction

You can’t forecast all emergent behaviors of complex systems. Iterative deployment replaces speculative risk-modeling with real feedback from millions of diverse users. The OpenAI rollout of ChatGPT in November 2022 is the canonical case: a research release with disclaimers that invited society into the lab. User reports of hallucinations, bias, jailbreaks, and novel uses informed rapid improvements to GPT-4 and beyond. This mirrors earlier technologies: cars spurred traffic laws and safer designs; GPS leapt in utility when civilian access expanded and President Clinton ended Selective Availability in 2000 (dropping costs from devices like Magellan’s Nav 1000 to cheap chips).

The authors contrast this approach with blanket pauses (e.g., the Future of Life Institute’s 2023 call). For software-centered systems where updates are fast and harms are often reversible, they argue that public iteration is safer and more democratic than precautionary lock-downs (Note: they explicitly acknowledge exceptions—contexts with irreversible physical risks warrant tighter preclearance).

Core throughline

“Most concerns about AI are concerns about human agency.” The book’s remedy: widen access, measure relentlessly, and adapt in public view.

How LLMs empower—and why they fail

LLMs are predictive engines that generate the next likely token based on vast training corpora. They don’t “know” facts the way you do; they estimate patterns. That’s why hallucination is structural: outputs can be fluent yet false. The book demystifies this and shows how you can still gain “superagency”—faster learning, better writing, real-time translation—if you validate critical claims, ask for sources, and provide context. Studies cited show novice users gain most: MIT researchers found 37% faster completion on certain writing tasks; call-center agents saw 14% productivity gains. Multimodal features extend access: tools can convert legalese to plain language, turn PDFs into narrated podcasts, or offer situational help for deaf or vision-impaired users.

The authors emphasize promptcraft and persona-setting as ways to unlock “latent expertise” in models. Treat LLMs like informational GPS: you’ll navigate better if you give coordinates—your goal, constraints, and preferred format. But unlike GPS, there’s no single ground truth; language is contested and outputs are statistical (Note: this is a useful corrective to anthropomorphic hype).

Governance by testing and participation

Progress happens through measurement. Benchmarks like SuperGLUE, TruthfulQA, RealToxicityPrompts, and BLEU/WER act as communal scoreboards that reveal strengths and weaknesses across models (OpenAI’s GPT-4, Anthropic’s Claude 2, and others). Yet the authors warn against “teaching to the test” and data contamination; hence the rise of broader, user-led platforms like Chatbot Arena, where people compare anonymous model outputs and generate crowd-sourced rankings. This is “internet-style governance”: iterative, transparent, and participatory, complemented by formal law (GDPR, CCPA, and prospective AI rules) to establish rights and remedies.

From private commons to public goods—and sovereignty

Digital platforms are recast as private commons that generate outsized public value—think search, Wikipedia, YouTube, and LinkedIn’s professional graphs (Brynjolfsson and Collis estimate large consumer surplus, e.g., thousands per year from search). The challenge is governance: how to pay creators, respect privacy, and still cultivate data-rich ecosystems that make AI useful. Lawsuits (New York Times, Getty Images, authors) and privacy regimes are real frictions the book treats as design inputs, not reasons to freeze progress. At scale, networks become infrastructure—like the Interstate Highway System or GPS—that multiplies individual autonomy while demanding coordination and consent (NIST pegs GPS public-sector benefits at over a trillion dollars).

Finally, there’s geopolitics. “Sovereign AI” captures why countries invest in domestic compute, data, and models (France’s cultural datasets, Singapore’s regional norms, the U.S. CHIPS Act). Democracies that stall risk ceding capability to authoritarian rivals. The book’s advice threads the needle: pursue sovereign capacity and open participation, pair iterative deployment with legal safeguards, and anchor every decision to the question, “Does this increase human agency?”


Iterative Deployment as Compass

Hoffman presents iterative deployment as the practical north star for AI progress. You release early, monitor widely, and update often—technically, socially, and regulatorily. It is not a blank check; it is a governance method that invites users, researchers, and policymakers into a continual cycle of feedback and improvement. The model rejects absolutisms: neither a precautionary freeze nor heedless acceleration produces the learning you need to make systems useful and safe.

How the method works

Technically, you stage rollouts, run A/B tests, publish disclaimers, and push patches quickly. Socially, you cultivate a culture of reporting—bug bounties for bias and jailbreaks, and channels for user red-teaming. Regulatorily, you prefer adaptive rules that can ratchet up as you learn. OpenAI’s ChatGPT launch in November 2022 exemplified this pattern: a research preview, not a “sealed” product. Public experiments like Chatbot Arena extend this logic at scale by crowdsourcing comparative judgments that inform fine-tuning and safety settings.

The book leans on historical parallels. Early automobiles caused chaos; then cities invented traffic lights, driver’s licenses, and road standards, while manufacturers iterated on brakes, glass, and seatbelts. Over decades, fatalities per mile fell dramatically. Similarly, GPS started as a military system. With President Reagan’s pledge for civilian access (after the 1983 KAL 007 incident) and Clinton’s 2000 end to Selective Availability, civilian receivers became cheap and ubiquitous, unleashing maps, logistics, and precision agriculture.

Why it beats “pause-first”

Millions of real users surface edge cases no lab can simulate. Iteration transforms unknown unknowns into known problems you can fix.

Where restraint still applies

The authors don’t claim iteration is universal. When interventions can cause irreversible harm in the physical world—nuclear plants, medical devices, or fully autonomous weapons—precautionary licensing and ex ante testing matter more. But for most AI software and information systems, where updates are rapid and harms can be reversed, iteration plus monitoring tends to be safer and more equitable than waiting for theoretical certainty (Note: this echoes agile and MVP playbooks from software, adapted to public-interest stakes).

How to put it into practice

If you build AI, start with constrained capabilities, narrow domains, and explicit guardrails. Instrument your system to log failures, user corrections, and prompts that trigger unsafe outputs. Publish model cards, incident reports, and changelogs. If you’re a policymaker, design sunset clauses, sandboxes, and tiered obligations that scale with real risk evidence. If you’re a user, expect disclosure: what’s experimental, what data is collected, and how to report harm.

The authors also counsel humility about “grand master plans.” Complex systems reward flexibility. Iterative deployment is not a moral abdication; it is a commitment to learning in public, where evidence—not vibes—drives updates. That stance, they argue, is the surest way to align fast-moving AI with human values while delivering benefits sooner to people who can use them now.


Human Agency and Superagency

Agency sits at the heart of the book: your ability to set aims, choose actions, and influence outcomes. Technologies have always reshaped agency—books extended memory, steam amplified muscle, computers scaled cognition. AI levels this up by supplying synthetic intelligence that assists or acts on your behalf. The authors’ normative aim is “superagency”: ensure AI broadens the set of things you can do, not narrows it.

Two modes: assistant and agent

As an assistant, AI helps you plan trips, draft briefs, debug code, or translate leases to plain language. As an agent, it can execute goals with autonomy—optimizing your home’s energy use or triaging your inbox. Both increase capacity, but agentic systems raise sharper concerns: dependency, skill atrophy, or manipulation. The authors recommend graduated autonomy with override options, audit trails, and clear scopes of authority so you remain “the boss.”

Evidence suggests these tools particularly lift novices. MIT’s 2023 study reported 37% faster writing on specific tasks with ChatGPT, and customer-service experiments showed 14% productivity gains—largest for less-experienced agents. That’s democratization: capabilities once gated by education or income become accessible with a prompt. Multimodal features compound this—camera, text, and audio make information inclusive for dyslexia, vision impairment, or multilingual contexts.

Human-centered frame

“AI should be an extension of individual human wills and be as broadly and evenly distributed as possible.” This OpenAI founding vision guides the book’s agency lens.

Care, counsel, and the “superhumane” claim

In mental health and education, demand outstrips human supply. Apps like Woebot or Kokobot are limited, but they can deliver consistent, on-demand support at scale. The authors call this “superhumane”: AIs that never tire, never shame, and adapt to your pace (Note: this complements, not replaces, expert professionals). The same logic applies to legal literacy—LLMs can explain “quit premises” notices, outline options, and draft letters in your language.

Guarding against loss of agency

Agency can also shrink. Predictive policing and some judicial algorithms applied without consent reduce people to profiles. Private property regimes that deploy facial recognition to exclude (e.g., Madison Square Garden blocking attorneys tied to lawsuits) show how “code as gatekeeper” can erode due process. The authors therefore insist on user control, transparency, and avenues for contestation. Design systems for “explainability to the user,” logs for appeals, and human override for consequential decisions.

From personal to civic agency

Your agency compounds when your nation invests in AI. Democracies that obstruct development risk ceding capability to more authoritarian rivals. Hence “sovereign AI”: countries like France (funding French-language data and models) and Singapore (regional alignment in its National AI Strategy) aim to encode their norms in domestic tooling, while the U.S. strengthens compute and talent via the CHIPS Act. The civic corollary: promote broad access, education, and experimentation so agency increases for many, not a narrow elite.

The upshot: design for superagency. Keep the human in command; give novices a lift; make consent, transparency, and recourse non-negotiable; and scale access so individuals—and the democracies they inhabit—grow stronger together.


LLMs: Engines, Limits, and GPS

To use AI well, you need a crisp mental model of LLMs. They are neural predictors trained on vast text (and increasingly image/audio) corpora to estimate the next token. That architecture yields surprising fluency and reasoning-like behavior—but also structural limits: hallucinations, biases from training data, and opaque internal mechanisms. The authors give you an adoption playbook that embraces strengths while hedging weaknesses.

Why hallucinations happen

LLMs don’t retrieve facts by default; they sample plausible continuations. When training data is thin or prompts are underspecified, the model fills gaps with confident-sounding fiction. Systems like GPT-4 and Claude 2 reduce rates but don’t eliminate them. That’s why the book treats hallucination as a feature of probabilistic generation, not a mere bug (Note: some researchers prefer “confabulation,” but the everyday term helps non-specialists).

Bias, opacity, and provenance

Training choices shape outputs: whose text you ingest, what you filter, and how you RLHF models with human preferences. If the corpus carries historical prejudice, models can reproduce it. Opacity compounds the issue: tracing a given answer to a specific shard of training data is hard. The authors advocate better provenance (citations, verifiable references), red-team evaluations (RealToxicityPrompts), and interpretability research alongside product iteration.

LLMs as informational GPS

The GPS analogy clarifies use. GPS maps physical coordinates to turn-by-turn guidance; LLMs map linguistic terrain to step-by-step explanations. Civilian access made GPS explode in value after Clinton ended Selective Availability in 2000; likewise, broad LLM access unlocked productivity for non-experts. Yet the analogy has limits: GPS has a single ground truth, while language is plural and contested. Different models—tuned on different data and business incentives—create distinct “informational planets.” Expect variance and steer with context.

Operating principles

Treat outputs as drafts, not verdicts. Add your “coordinates” (intent, audience, constraints). Ask for sources, confidence, and counterarguments.

Practical tactics you can use today

Anchor prompts with role and format: “Act as a housing counselor; produce a two-paragraph plain-language summary and a checklist.” Provide background to reduce hallucinations and ask for citations with URLs or doc IDs. For sensitive domains (health, law), use AI to structure information and questions for a qualified human, not to replace one. For accessibility, use multimodal features: turn PDFs into narrated dialogues or use live transcription with speaker IDs for meetings.

Where the frontier is heading

Debate continues over whether sheer scale yields AGI. The book stakes a pragmatic middle: expect gains from architectural tweaks, multimodal training, tool-use, and hybrid neuro-symbolic methods, but plan for persistent uncertainty. Build systems with continuous evaluation, user feedback loops, and explicit “don’t overtrust me” signals. With that mindset, you harness LLMs as powerful navigational aids—accelerating understanding while preserving your judgment.


Testing and Internet-Style Governance

The book argues that modern AI advances are not a reckless arms race; they’re a measured contest of tests. Benchmarks and public evaluations function as decentralized governance: they reveal strengths and failures, set norms, and push models toward safer, higher-quality behavior. You, as a user or researcher, can help steer this trajectory by participating in these communal “Olympics.”

What we measure, we improve

Standard tasks like SuperGLUE and HellaSwag assess language understanding and commonsense. TruthfulQA and RealToxicityPrompts probe factuality and toxicity. BLEU and WER score translation and speech accuracy. Public leaderboards make progress legible, keep labs honest, and spotlight weaknesses for targeted fixes. Papers from OpenAI (GPT-4) and Anthropic (Claude 2) routinely report such metrics, reflecting a culture of measured claims.

Yet benchmarks alone can mislead. Models may overfit, especially if test data leaks into training (contamination). Narrow tasks can miss multi-turn failures, social harms, or jailbreak dynamics. That’s why the book elevates user-in-the-loop platforms like LMSYS’s Chatbot Arena: people compare anonymous model outputs, vote on quality, and generate a crowd-ranked leaderboard that correlates with real-world usefulness.

Layered governance

Benchmarks and public testing shape model behavior; law sets rights and remedies. Together, they create adaptive oversight.

From lab metrics to field safety

Iterative deployment turns evaluation into a living practice. Instrument your product with safety telemetry: hallucination catches, refusal rates, jailbreak triggers, and post-hoc corrections. Publish incident reports and changelogs like software teams do. Invite external audits or bug bounties that reward red teams for exposing prompt-injection pathways or unfair outputs. This is “internet-style” because it borrows from how the web evolved—permissionless innovation tempered by public scrutiny and rapid patching.

Regulation that learns

The authors do not reject formal regulation; they argue for adaptive regimes. Privacy laws (GDPR, CCPA) constrain data misuse; sector rules can mandate documentation, provenance, and human-in-the-loop for high-risk use. The key is to keep regulatory dials adjustable—sunsets, sandboxes, and risk tiers—so society can tighten or loosen as evidence warrants (Note: this mirrors financial regulation’s tiered capital requirements and aviation’s incident-driven updates).

Your role is practical: test models on your real tasks; report harms; share prompts that fail or succeed. In aggregate, these contributions become a crowd-sourced constitution for acceptable behavior—an evolving social contract that the best labs already bake into training and policy design.


Data, Infrastructure, and Sovereign AI

The book weaves data governance, public infrastructure, and geopolitics into a single question: how do we build systems that multiply agency at scale? The authors push a reframing: many digital platforms operate as private commons—privately administered but generating large public value. The goal is not to romanticize Big Tech, but to align incentives so data “agriculture” produces broad benefits without predation.

From extraction to cultivation

Unlike finite natural resources, data is nonrivalrous and regenerative when aggregated wisely. Platforms like Google Search, Wikipedia, YouTube, and LinkedIn yield outsized consumer surplus (Brynjolfsson and Collis’ experiments value search in the thousands of dollars per user per year). You contribute clicks, posts, and profiles—and you receive services that no individual could build alone. The authors acknowledge frictions: copyright lawsuits (New York Times, Getty Images, authors) and privacy regimes (GDPR, CCPA) reflect legitimate claims for compensation and control.

The policy task is balance. Licensing schemes, clearer provenance, and opt-out/opt-in mechanics can route value to creators while preserving the shared-data dynamics that improve AI. Participatory evaluation—public leaderboards, arenas, and audit trails—adds accountability. The authors want you to see cooperation, not zero-sum extraction, as the default frame for data-era value creation.

Infrastructure that expands freedom

Agency scales on top of public goods. The Interstate Highway System turned cross-country travel from a Donner Party ordeal into a safe 1,957‑mile drive; GPS unlocked over a trillion dollars in public-sector benefits (NIST estimate). AI-enhanced public services—faster benefits processing, epidemic detection, multilingual access—can deliver similar multipliers. South Korea’s Covid response illustrates “networked autonomy”: testing, AI-assisted tracing, and transparent communication enabled freer movement without sweeping lockdowns (with legal safeguards and scrutiny to retain legitimacy).

Consent for code

As Lessig noted, “code is law.” When architecture enforces rules, you need democratic consent, transparency, and recourse.

Law, code, and perfect control

Embedding enforcement in devices can save lives and also chill autonomy. Consider DADSS breath sensors, mandated for future cars by the U.S. Infrastructure Act: they can block drunk driving, but false positives raise due-process concerns. Smart contracts enforce terms automatically but struggle with the ambiguity of machine-learning inputs. Private venues using facial recognition (e.g., Madison Square Garden) to bar critics test the edges of property rights and public accommodation. To stay humane, the authors argue, keep space for judgment: appeal mechanisms, logs, and human review for consequential denials.

Sovereign AI without isolation

Nations now treat AI like strategic infrastructure. France funds French datasets and models to encode local culture; Singapore tailors AI to regional norms; the U.S. advances compute sovereignty via the CHIPS Act. The risk of delay is real: brain drain, lost competitiveness, and dependence on foreign stacks. Still, sovereignty should pair with global cooperation on safety, standards, and research openness. The authors champion a techno-humanist middle path: build domestic capacity; keep systems interoperable; and anchor choices to the question, “Does this broaden people’s agency?”

In practice, that means you support policies that pay creators fairly, protect privacy, and keep data commons vibrant. You demand transparency when code becomes law. And you treat AI as civic infrastructure—worthy of investment and debate—so individuals and societies both move with more freedom, not less.

Dig Deeper

Get personalized prompts to apply these lessons to your life and deepen your understanding.

Go Deeper

Get the Full Experience

Download Insight Books for AI-powered reflections, quizzes, and more.