AI & Society
This page is a survey of what AI is doing to people — measured effects, credible forecasts, and active policy fights. It avoids taking sides on "AI good" vs "AI bad" and instead tries to stay close to sourced claims, including the ones that contradict each other.
1. How to read this page
The honest version of "how will AI affect society?" is: nobody knows yet, but we can be specific about what we do know, which forecasts are taken seriously, and which claims rest on handwaving. The sections below are ordered by how much settled evidence exists:
- Labor & economy — measurable in real time. Multiple RCT-style studies exist.
- Governance — actual laws and their actual provisions.
- Epistemics — early evidence, mixed.
- Science — concrete wins, extrapolation open.
- X-risk & misuse — almost entirely arguments from priors, modeled scenarios, and expert judgment.
A reading heuristic: the further down the page you go, the more the claims rest on someone's model of the future rather than on something already measured. Weight accordingly.
2. Labor & the economy — what the data says
Three questions drive the labor debate: Who gets automated? Who gets augmented? How much aggregate productivity comes out? The answers, in 2026:
Who gets automated
The most-cited paper is Eloundou, Manning, Mishkin & Rock (2023) — GPTs are GPTs. The authors used GPT-4 plus human annotators to score whether an LLM plus "reasonable software tooling" could cut the time for each of 19,000+ specific tasks in the O*NET occupation database by ≥50%.
80% of US workers have at least 10% of their tasks exposed.
19% of workers have at least 50% of their tasks exposed.
Highest-exposure occupations: interpreters, tax preparers, writers, web designers, accountants, mathematicians, blockchain engineers.
Lowest-exposure: short-order cooks, dishwashers, masons, pile-driver operators.
"Exposure" ≠ "replacement." Many exposed tasks still need a human in the loop; many jobs aren't fully characterized by their task list. But the direction is unambiguous: the first jobs to feel pressure are information-processing and writing-heavy, which is the opposite of the 20th-century industrial-automation pattern (manual labor first).
Who gets augmented
A handful of real-world deployment studies form the empirical backbone here. All of them find that AI assistance helps less-experienced workers more than experienced ones — compressing the skill distribution.
- Brynjolfsson, Li & Raymond (NBER 2023, w31161). Call center with 5000+ agents deployed an LLM-based assistant. Average productivity +14%. Novice workers: +34%. Top-quartile workers: ~0%. Customer satisfaction also improved and employee retention was higher.
- Noy & Zhang (Science 2023, doi:10.1126/science.adh2586). Controlled experiment on 444 college-educated professionals doing short writing tasks. ChatGPT access cut task time by ~40% and increased output quality; the gains concentrated on lower-performing participants.
- Peng, Kalliamvakou, Cihon & Demirer (GitHub Copilot study, 2023). 95 developers on a JavaScript task. Copilot users finished 56% faster. The effect was larger for less-experienced and older developers.
- Dell'Acqua et al. (BCG / HBS "jagged frontier" study, 2023). 758 BCG consultants on 18 business tasks. Inside the "frontier of AI capability," consultants using GPT-4 completed tasks 25% faster with 40% higher quality. Outside the frontier, AI-assisted workers were more likely to produce wrong answers — the skill to know when to trust the tool mattered.
Aggregate productivity — the contested part
This is where respected economists disagree strongly.
- Goldman Sachs (Hatzius et al., 2023) — generative AI could raise US labor productivity growth by ~1.5 percentage points annually over a decade, lifting GDP by ~7%. Widely circulated, somewhat optimistic assumptions about deployment pace.
- McKinsey Global Institute (2023) — generative AI could add $2.6–4.4T in annual value across the economy. Also optimistic.
- Daron Acemoglu (NBER 2024, w32487) — the explicit skeptic. Estimates AI will add roughly 0.06 percentage points per year to productivity growth over a decade, for a total TFP gain of ~0.66%. His argument: only a small share of tasks are actually cost-effective to automate even when technically exposed, and the gains don't meaningfully compound without new product creation.
- IMF (2024 staff paper, "Gen-AI: Artificial Intelligence and the Future of Work") — 40% of global employment exposed; 60% in advanced economies; developing economies less exposed but also less able to benefit. Warns of widening inequality unless safety nets and reskilling catch up.
The range "0.06% to 1.5% annual productivity boost" is huge — one of those spans is barely noticeable and the other is world-historical. That uncertainty is real, not a failure of analysis. Which number is right depends on things (deployment pace, organizational adoption, complementary investment) that haven't happened yet.
3. Interactive: task-exposure simulator
A toy model. Imagine an economy of $N$ occupations, each made of $M$ tasks. A fraction $f$ of all tasks are "AI-exposed" (high enough automation probability that a rational employer will eventually replace that task with a model call). Adjust the fraction below and see how it distributes across occupations, assuming occupations that are more information-heavy have higher exposure.
Bars show the fraction of task-time affected by AI for 12 representative occupations. Default f = 0.19 reproduces the Eloundou et al. "≥50% of tasks exposed" headline.
This is an illustration, not a forecast. Real exposure is correlated with occupation structure in complicated ways — the point of the demo is only to show that the same aggregate number can describe wildly different worlds depending on how concentrated exposure is.
4. Governance & regulation — what actually exists
There is now a real, messy patchwork of AI regulation. A summary of the ones you should know about:
The EU AI Act
Published in the Official Journal 12 July 2024, entered into force 1 August 2024. The world's first comprehensive horizontal AI law. Structure:
- Risk-tiered. Prohibited uses (social scoring, real-time biometric ID with narrow exceptions, emotion recognition in workplaces/schools, etc.) → high-risk (medical devices, critical infrastructure, law enforcement tools — heavy compliance) → limited-risk (chatbots, deepfakes — transparency only) → minimal-risk (most software — no obligation).
- General-purpose AI (GPAI) rules for foundation-model providers, with extra obligations for "systemic risk" models — currently defined as anything trained with ≥10²⁵ FLOPs. That threshold captures GPT-4, Claude 3+, Gemini 1.5+, Llama 3 70B+, and onwards.
- Timeline: prohibitions from Feb 2025; GPAI rules Aug 2025; most high-risk obligations Aug 2026. The "Code of Practice" for GPAI — the compliance playbook frontier labs sign up to — was finalized in 2025.
- Penalties: up to €35M or 7% of global annual turnover, whichever is higher. That's GDPR-level.
United States
No federal horizontal AI law. The regulatory position shifted sharply at the start of 2025:
- Biden EO 14110 (Oct 2023) — broad executive order requiring safety testing for models trained at ≥10²⁶ FLOPs, Commerce/NIST risk-evaluation frameworks, and a reporting regime. Rescinded by Trump EO on 20 Jan 2025.
- AI Action Plan (July 2025) — the Trump administration's replacement framing. Pro-innovation, pro-infrastructure, pro-competition with China. Formalizes Stargate-style compute build-out and preempts state laws where possible.
- State laws. California SB 53 (2025) — post-SB 1047 compromise, focused on critical-harm reporting by frontier labs. Colorado AI Act (2024) — consumer-facing AI systems used in high-stakes decisions. New York City local law 144 — bias audits for hiring algorithms. Texas HB 1709 (2025). Dozens more in progress.
- Sectoral. FTC on deceptive AI claims, CFPB on algorithmic lending, FDA on AI in medical devices, NIST on AI risk management framework. Nothing consolidated.
United Kingdom
Light-touch, principles-based approach, coordinated through existing regulators plus the AI Safety Institute (founded 2023, renamed AI Security Institute 2025). AISI runs pre-deployment safety evaluations on frontier models under voluntary agreements with the labs. The Bletchley Declaration (Nov 2023) and its successors — Seoul (May 2024) and Paris Summit (Feb 2025) — are the main international coordination venues. Paris pivoted hard from "safety" to "opportunity" framing; some read this as a sign the coordination regime is weakening.
China
Two binding instruments worth knowing. Algorithm recommendation regulation (March 2022) — registration requirements, opt-out provisions for recommender systems. Generative AI interim measures (Aug 2023) — service providers must file security assessments, label AI-generated content, and ensure outputs reflect "core socialist values." Model registration is now routine for any provider deploying generative AI to the Chinese public.
International
G7 Hiroshima AI Process, OECD AI Principles, UN advisory body on AI, Council of Europe AI Convention (first binding international treaty, signed Sep 2024). None of these have strong enforcement; all of them set up soft-law baselines.
5. Information & epistemics
The three-part claim: AI is (1) making content creation near-free, (2) restructuring how people find information, and (3) making authenticity verification harder. Evidence on each:
Content creation costs collapsed
This is the closest thing to settled fact on this page. A 2023 paper by Hanley & Durumeric found synthetic news articles on misinformation sites rose sharply through 2022–23. NewsGuard's ongoing tracker identifies hundreds of AI-generated news sites that launched since 2023. Europol's 2024 assessment (IOCTA) treats generative AI as now-standard tooling for fraud.
Answer engines are eating search
ChatGPT Search, Perplexity, You.com, Google AI Overviews, and SearchGPT have moved meaningfully into territory previously owned by traditional web search. The measurable downstream effect is on publisher traffic. A 2024 study by Sistrix on 8000 German websites found noticeable referral declines from Google for queries where AI Overviews appeared. Multiple publishers (New York Times, News Corp, Reuters, Axel Springer) filed lawsuits or signed licensing deals between 2023 and 2025, depending on which side of the fight they wanted to be on. Net effect on the open-web economic model: unsettled and probably negative.
Did AI break the 2024 elections?
The most-asked question at the start of 2024. The honest answer is: no, not in any measurable way — at least not in the developed democracies with adequate observation. The Alan Turing Institute's 2024 report found "no evidence of a large-scale effect of AI-generated disinformation on election outcomes" across ~112 elections they surveyed. That is not a reassuring finding about 2028 — it is a statement that the 2024 worst-case didn't happen. Generative-AI capability outpaced detection capability the entire time; the lack of disruption is more puzzle than prophecy.
6. Science & discovery — the concrete wins
This is the section where the extrapolation is most grounded. There are already concrete scientific wins attributable to AI methods, not just promises:
- AlphaFold 2 (2020) & AlphaFold 3 (2024). Predicts protein structure to near-experimental accuracy. The AlphaFold Protein Structure Database has ~200M predictions covering essentially every known protein. Hassabis & Jumper won the 2024 Nobel Prize in Chemistry for this work. Downstream drug-discovery groundwork worldwide depends on it.
- AlphaProof & AlphaGeometry 2. DeepMind's math systems. Silver-medal performance at IMO 2024 — one point short of gold on the published solution metric.
- GraphCast (Lam et al., 2023). Graph-neural-network weather forecasting that beats the operational ECMWF deterministic model on most metrics at a fraction of compute cost.
- FunSearch (Romera-Paredes et al., 2023). DeepMind LLM paired with an evaluator found new constructions for the cap set problem — a genuine novel mathematical result.
- NVIDIA BioNeMo & Meta ESMFold. Broad-coverage biology foundation models. Less headline-grabbing than AlphaFold, but the ecosystem is now rich.
- Materials-science results from Google's GNoME (2023): 2.2 million new stable crystal structures predicted, thousands already synthesized by collaborators. The follow-up critique argues many predicted materials aren't genuinely novel — a healthy pattern of claim + scrutiny, not a failure.
- Sakana AI's "AI Scientist" (2024). End-to-end research automation producing small original ML papers. Interesting as a proof of concept; the papers themselves are not yet strong publications.
If you care about extrapolation: the thing to watch is whether AI-assisted science moves from "making existing pipelines faster" to "producing results that would not have been produced otherwise." AlphaFold is unambiguously the second. FunSearch is the second. The others are mostly the first — still valuable, but not yet transformative in the strict sense.
7. Existential risk — the most-contested corner
The "x-risk" question — could sufficiently capable AI cause outcomes as bad as human extinction or permanent civilizational collapse? — is where the evidence thins out and prior-laden reasoning takes over. A fair summary of the landscape:
The concern, stated minimally
- We don't yet have techniques to specify human values precisely enough to write them into an optimization target.
- Any sufficiently capable optimizer pursuing a misspecified target will, as an instrumental matter, acquire resources and resist shutdown (Omohundro 2008, "basic AI drives").
- If that optimizer is much smarter than the humans trying to correct it, correction becomes hard, possibly impossible.
This argument is due to Bostrom (Superintelligence, 2014), with antecedents in Yudkowsky. It is a theoretical case — no concrete mechanism has been demonstrated in a deployed system. Proponents say that's the point (you only find out you were wrong once). Skeptics say the premises are wrong or the premises are right and don't actually compose into the conclusion.
What the community actually says
Some concrete data points, in chronological order:
- CAIS statement (May 2023). One sentence: "Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war." Signatures include Hinton, Bengio, Altman, Hassabis, Amodei, Russell, plus ~1000 others. The moment x-risk became mainstream.
- Geoffrey Hinton resigned from Google (May 2023) to speak publicly about risks. Has since said he thinks there's a 10–20% chance AI leads to human extinction within 30 years.
- Yoshua Bengio — Turing laureate, wrote a "Rogue AIs" essay in 2023 arguing the risk is real enough to reshape his research agenda. Leads the UK-commissioned International AI Safety Report (2025), the most comprehensive official expert assessment to date.
- AI Impacts survey (Grace et al. Jan 2024, arXiv:2401.02843). Of ~2700 AI researchers, the median estimate of "extremely bad outcomes (e.g. human extinction)" was 5%. A third of respondents put it at ≥10%. This is researcher opinion, not an objective measurement, but it's the best survey we have.
- Skeptical frontier voices. Yann LeCun (Meta), Andrew Ng, Melanie Mitchell, Gary Marcus argue that current x-risk framings overweight theoretical scenarios vs. concrete near-term harms (bias, misuse, labor, concentration of power). LeCun's phrasing: worrying about superintelligence risk today is "like worrying about overpopulation on Mars."
- Anthropic's "Responsible Scaling Policy" and OpenAI's "Preparedness Framework" (both 2023, updated through 2025) — both commit to capability thresholds beyond which additional safeguards are required before further deployment. The first real industry mechanism for operationalizing "if capabilities get scary, we pause."
Something like 10–30% of frontier-lab researchers take x-risk seriously as a non-negligible possibility. Something like 20–40% consider it a distraction from concrete harms. Something like 30–50% think it's worth taking seriously but don't spend their time on it. No one reasonable believes it is zero or one. If you want a single number: the median expert P(existential catastrophe from AI) is somewhere in the low-single-digit percent, with enormous disagreement.
8. Bio, cyber, and misuse — the near-term risks
Separate from the "superintelligent AI" story is the much more concrete concern: AI makes existing dangerous capabilities more accessible to smaller groups. The three live ones:
Biological uplift
Will an LLM help an amateur synthesize a dangerous pathogen? RAND's 2024 red-team study found that LLMs did not (yet) give meaningful uplift to people attempting biological attack planning, beyond what the open internet provides. OpenAI's own evaluation reached similar conclusions — small uplift, not transformative. Anthropic's Claude's Responsible Scaling Policy explicitly names "AI systems that meaningfully uplift novice actors in creating bioweapons" as a threshold that triggers additional safeguards. The Urbina et al. (2022) Nature Machine Intelligence study — where a drug-discovery model was redirected to produce novel toxin candidates — is the stark demonstration that the machinery exists, even if the current LLM UX doesn't unlock it.
Cybersecurity
More definite. LLMs accelerate both sides of the offense-defense equation. On offense: automated vulnerability scanning, phishing customization, reverse engineering, code analysis. On defense: the same, plus faster triage and patching. The DARPA AI Cyber Challenge (2024–25) — teams compete to build autonomous systems that find and patch software bugs — is the canonical public benchmark. The net effect is probably slight defender advantage at the margin, because defenders start with more context, but this is not settled.
Concentration of power
The least-studied but most-measurable risk. Frontier AI capability costs ~$100M+ to train at state-of-the-art; only a small number of organizations can do it. The worry is not that any one of them will do something bad — it is that the structure of a handful of entities controlling the most capable systems is a poor fit for democratic societies. Both sides of the political spectrum have flagged this. Open-weight model releases (Llama 3/4, Mistral, DeepSeek, Qwen) are the main structural mitigation and are controversial for exactly the same reason.
9. Sourced reading list
Labor & economy
- Eloundou, Manning, Mishkin, Rock (2023) — GPTs are GPTs.
- Brynjolfsson, Li, Raymond (2023) — Generative AI at Work. NBER.
- Noy & Zhang (2023) — Experimental evidence on the productivity effects of generative artificial intelligence. Science.
- Dell'Acqua et al. (2023) — Navigating the Jagged Technological Frontier. HBS Working Paper.
- Acemoglu (2024) — The Simple Macroeconomics of AI. NBER.
- Goldman Sachs (2023) — The Potentially Large Effects of Artificial Intelligence on Economic Growth. Hatzius et al.
- IMF Staff Discussion Note (2024) — Gen-AI: Artificial Intelligence and the Future of Work.
Governance
- EU AI Act, Regulation (EU) 2024/1689. Published OJ L, 12 July 2024.
- OECD AI Principles (updated 2024).
- Bletchley Declaration (Nov 2023); Seoul Ministerial Statement (May 2024); Paris AI Action Summit declaration (Feb 2025).
- Council of Europe Framework Convention on AI (Sep 2024).
- NIST AI Risk Management Framework (2023).
Epistemics & elections
- Alan Turing Institute (2024) — AI-enabled influence operations: Threat analysis of the 2024 UK and European elections.
- NewsGuard AI Tracking Center (ongoing).
- Europol IOCTA 2024 — Internet Organised Crime Threat Assessment.
Science
- Jumper et al. (2021) — Highly accurate protein structure prediction with AlphaFold. Nature.
- Abramson, Adler, Dunger et al. (2024) — Accurate structure prediction of biomolecular interactions with AlphaFold 3. Nature.
- Lam et al. (2023) — GraphCast: Learning skillful medium-range global weather forecasting. Science.
- Romera-Paredes et al. (2024) — Mathematical discoveries from program search with large language models. Nature.
- Merchant, Batzner, Schoenholz et al. (2023) — Scaling deep learning for materials discovery. Nature. (GNoME.)
Existential risk & misuse
- Bostrom (2014) — Superintelligence. OUP.
- Russell (2019) — Human Compatible.
- Bengio et al. (2025) — International AI Safety Report 2025. UK government.
- Grace, Stewart, Sandkühler et al. (2024) — Thousands of AI authors on the future of AI. AI Impacts.
- CAIS (2023) — Statement on AI Risk.
- RAND (2024) — The Operational Risks of AI in Large-Scale Biological Attacks.
- Urbina et al. (2022) — Dual use of artificial-intelligence-powered drug discovery. Nature Machine Intelligence.
- Omohundro (2008) — The Basic AI Drives. AGI-08 proceedings.