The Widening Gap Between What AI Can Do and What People Think It Can Do

Stanford's latest comprehensive analysis of artificial intelligence reveals a troubling paradox: the technology is advancing faster than ever, but public understanding and expert opinion have diverged so dramatically that they might as well be describing different worlds. The 2026 AI Index Report, published by Stanford University's Human-Centered AI Institute, documents not just the capabilities of AI systems, but the massive gap between what experts believe about AI's future and what ordinary people fear .

Why Are Experts and the Public So Divided on AI's Impact?

The numbers tell a striking story. When researchers asked AI experts whether artificial intelligence will have a positive impact on how people do their jobs, 73 percent said yes. When they asked the general public the same question, only 23 percent agreed. That is a 50-point gap .

The disagreement extends far beyond employment. On whether AI will benefit the economy overall, 69 percent of experts said yes, compared to just 21 percent of the public. On medical care, the divide was even starker: 84 percent of experts predicted positive impacts, while only 44 percent of the public did .

This disconnect reflects a fundamental mismatch in expectations. Nearly two-thirds of Americans, or 64 percent, expect AI to lead to fewer jobs over the next 20 years. Experts are less pessimistic; only 39 percent predict job losses, while 19 percent actually predict more jobs. Experts also forecast far faster adoption, expecting generative AI to assist 80 percent of U.S. work hours by 2030, compared to the public's estimate of just 10 percent .

What Are AI Models Actually Capable of Right Now?

The capabilities of frontier AI models have reached remarkable milestones. Several models now meet or exceed human performance on PhD-level science questions, multimodal reasoning tasks, and competition mathematics. On a key coding benchmark called SWE-bench Verified, performance jumped from 60 percent to nearly 100 percent of the human baseline in a single year .

In medicine, AI tools that automatically generate clinical notes from patient visits saw substantial adoption in 2025. Physicians reported spending up to 83 percent less time writing notes and experienced significant reductions in burnout . In chemistry, frontier models now outperform human chemists on average on ChemBench, a specialized benchmark for chemical reasoning .

Yet these same models fail at tasks that seem trivially simple. The top AI model reads analog clocks correctly only 50.1 percent of the time, barely better than random guessing. Anthropic's Claude Opus 4.6 reads the time correctly with just 8.9 percent accuracy, despite scoring exceptionally well on other benchmarks .

"There is a research thread that shows that when systems are asked questions about combinations of language with other modalities, such as images or audio, the language component carries a surprisingly large part of the burden, even to the extent of ignoring non-language information completely," noted Ray Perrault, co-director of the AI Index steering committee.

Ray Perrault, Co-Director of the AI Index Steering Committee, Stanford University

This uneven capability profile, which researchers call the "jagged frontier" of AI, matters enormously for anyone building with or relying on these systems. Knowing what a model does well is only half the picture; understanding its blind spots is equally critical .

How to Assess AI's Real-World Impact on Your Work

Productivity gains from AI are real, but they are not uniform across all types of work. The report documents measurable improvements in specific categories:

  • Customer Support: AI tools delivered 14 to 15 percent productivity gains in customer service roles, where outputs are easy to measure and monitor.
  • Software Development: Developers saw 26 percent productivity improvements, with AI handling routine coding tasks and bug fixes.
  • Marketing Output: Marketing teams experienced a 50 percent gain in output, one of the highest documented across any sector, as AI assists with content generation and campaign planning.
  • Complex Reasoning Tasks: Gains are smaller and sometimes negative in work requiring deeper judgment or creative problem-solving, suggesting AI is better at augmenting structured work than replacing human expertise.

There is also emerging evidence that heavy reliance on AI may carry long-term learning penalties that slow skill development over time, a concern that has not yet been fully studied .

Where Is AI Investment Flowing, and What Does It Mean?

The financial picture shows explosive growth. Global corporate AI investment reached $581.69 billion in 2025, a 129.9 percent increase from the previous year. Private investment alone grew 127.5 percent to $344.7 billion .

The United States remains the dominant force, with $285.9 billion in private AI investment. That is 23 times more than China's $12.4 billion and 48 times more than the United Kingdom's $5.9 billion. In generative AI specifically, U.S. investment exceeded the combined total of China and Europe by a wide margin .

Several landmark funding events underscore the scale of this investment. OpenAI raised $40 billion at a $300 billion valuation. Anthropic raised $13 billion at a $183 billion valuation. The Stargate Project, a joint venture between OpenAI, SoftBank, Oracle, and MGX, announced plans to invest between $100 billion and $500 billion in AI data centers in the United States by 2029 .

World AI compute capacity has grown 3.3 times every year since 2022, with total AI compute increasing 30-fold since 2021. Nvidia's GPUs account for over 60 percent of the total AI compute capacity globally, though Amazon and Google, which design their own hardware, rank second and third .

Is the U.S. Really Ahead of China in AI Development?

The answer is more complicated than headlines suggest. While the United States still leads in private investment and produced more top-tier AI models in 2025, the performance gap between U.S. and Chinese models has effectively closed at the model level. U.S. and Chinese models have traded the lead multiple times since early 2025 .

In February 2025, DeepSeek-R1 briefly matched the top U.S. model. As of March 2026, Anthropic's top model leads by just 2.7 percent. Meanwhile, China leads in publication volume, citations, patent output, and industrial robot installations. South Korea stands out for innovation density, leading the world in AI patents per capita .

One important caveat: China's government has deployed an estimated $184 billion in state-backed guidance funds into AI firms since 2000, which means private investment comparisons likely understate China's total AI spending .

What About the Environmental Cost of Training These Models?

The carbon footprint of training frontier AI models has grown dramatically. The report estimates that training the latest frontier large language models, such as xAI's Grok 4, can generate over 72,000 tons of carbon-equivalent emissions. That is a huge increase from estimates in prior years. OpenAI's GPT-4 was estimated at 5,184 tons, and Meta's Llama 3.1 405B was estimated at 8,930 tons .

Emissions from AI inference, the process of running a trained model to generate responses, also continue to increase. The report estimates that carbon emissions from models with the least efficient inference are over 10 times as high as those with the most efficient inference. DeepSeek's V3 models were estimated to consume around 23 watts when responding to a medium-length prompt, while Claude 4 Opus was estimated to consume about 5 watts .

However, these figures come with important caveats. Ray Perrault noted that "these estimates should be interpreted with caution. In the case of Grok, they rely heavily on inferred inputs drawn from public reporting, xAI statements, and other non-verifiable sources, introducing a degree of uncertainty." Epoch AI independently estimates Grok 4's emissions to be significantly higher at approximately 140,000 tons of carbon dioxide .

Ray Perrault

How Quickly Has Generative AI Been Adopted Compared to Past Technologies?

Generative AI reached mass adoption faster than almost any technology in history. It hit approximately 53 percent population-level adoption within three years of its mass-market introduction .

The value consumers are getting from these tools is also growing rapidly. Estimated U.S. consumer surplus from generative AI tools reached $172 billion annually by early 2026, up from $112 billion a year earlier. The median value per user tripled over that same period. Most of these tools remain free or close to it, meaning users are capturing enormous value without paying directly .

Organizational adoption has also reached remarkable levels. Eighty-eight percent of surveyed companies reported using AI in some capacity. Four in five university students now use generative AI regularly . Globally, 58 percent of employees reported using AI on a semiregular or regular basis in 2025 .

The Stanford AI Index Report reveals a technology in explosive growth, with capabilities advancing faster than public understanding can keep pace. The real challenge ahead is not whether AI will transform work and society, but whether institutions can manage that transformation thoughtfully enough to bridge the gap between expert optimism and public concern.