The Last Invention
  • The Last Invention
  • I. The Intelligence Landing
  • II. The Countdown
  • III. Work’s Last Stand
  • IV. Wealth in the Machine Age
  • V. The Prep Window
  • VI. Thriving Through Transition
  • VII. Humanity's Final Exam
  • VIII. Intelligence on Intelligence
  • Supplementary Sections
    • S1. The Economics of Zero
    • S2. The Ultimate Scarcity
    • S3. Meaning in a Solved World
  • About
Powered by GitBook
On this page
  • Forecasting the intelligence landing
  • The Bull case
  • The Bear case
  • The five hurdles
  • 1. World-model transfer
  • 2. Continual memory
  • 3. Causal reasoning
  • 4. Long-term planning
  • 5. Self-monitoring
  • Sources of uncertainty
  • 1. Recursive self-improvement
  • 2. Shifting constraints
  • 3. Emergent jumps
  • Bull or Bear?
  • The Five-Year Horizon (2025-2030)
  • The Long Horizon (Beyond 2030)
  • The Short Horizon (Before 2025)
  • My Synthesis
  • Key questions to guide timelines
  • Appendix: Bull case vs. Bear case
  • The Bull case
  • The Bear case
  • Footnotes
Export as PDF

II. The Countdown

Last updated 13 days ago

Will AGI crash-land before your next phone upgrade or decades later? Five technical hurdles remain critical: world-model transfer, memory, causality, planning, and self-monitoring. AGI within 5 years seems plausible, but depends on unpredictable factors like self-improvement, emergence, and resource constraints. Early arrival needs a rare perfect alignment of dependencies.

Table of Contents

I never allow myself to hold an opinion on anything that I don't know the other side's argument better than they do.

― Charlie Munger

It is better to be roughly right than precisely wrong.

— John Maynard Keynes


Forecasting the intelligence landing

Humans have faced transformative technology before. Fire, agriculture, the industrial revolution—each reshaping civilization in ways previous generations couldn't imagine. But we've never approached a transition with a creation that might think not just differently from us, but potentially beyond us. The gravity of this prospect means we need to try charting its arrival, even with imperfect tools and incomplete understanding.

While we can't know exactly when AGI will emerge, we can examine trajectories, identify bottlenecks, and assess the interplay of forces that will shape its development.

My goal is to provide context and evidence for you to navigate this question yourself, rather than advocating for a particular timeline.

The Bull case

People have a strange reluctance to face how quickly AI is advancing. It's as if not looking directly at it might somehow slow it down. Three years ago, suggesting AGI would arrive this decade would get you dismissed as indulging in Silicon Valley hype. Reality has dismantled this skepticism.

The evidence keeps accumulating in ways that are hard to ignore.

The bull case, it seems, is simply the logical extension of present trends, without the fiction of technical challenges. The trajectory of AGI is visible with striking clarity.

A summary of the Bull Case for near-term AGI
Category
Accelerants
Arguments

Scaling laws and predictive learning point toward human-level cognition.

Bigger models pick up new abilities; world models get broader and better; reinforcement learning combined with language models shows a path to general competence.

Capital, compute, and talent are scaling exponentially—and AI is starting to self-amplify.

Datacenter spending is breaking records; governments treat AI as a strategic priority; global competition kills any willingness to pause.

Reliability

Alignment research improves both safety and capabilities.

Red-teaming and reinforcement learning help models stay factual, cautious, and steerable; labs catch problems earlier and fix them faster.

Bigger models may actually make alignment easier.

Larger networks reveal clearer internal structures; AI could help supervise AI.

Social and market forces are pulling timelines forward.

Prediction markets and capital flows are optimistic; governments are funding AI infrastructure aggressively; public narratives are normalizing near-term AGI.

The Bear case

Each advance in paradoxically renders the more intractable, not less. has demonstrated a mathematically sound method to prevent a superintelligent system from pursuing goals that conflict with human well-being. When you factor in fragile global , geopolitical volatility, and the unpredictable nature of venture capital—which can vanish during economic downturns—the five-year timeline for AGI starts to look like wishful thinking.

Until we see a model that can learn continuously, reason causally, and operate at industrial reliability without hidden traps, betting on AGI sooner than five years doesn’t feel like the most sober forecast. The burden of proof appears to be on timelines shorter than five years.

A summary of the Bear Case for near-term AGI
Category
Main Bottlenecks
Arguments

Current methods may never deliver general intelligence.

Intelligence may be inherently narrow; transformers still lack causal reasoning, memory, continual learning, and physical grounding; recursive self-improvement may stall on idea generation and real-world feedback.

Scaling may plateau under real-world constraints.

Human-quality data is running out; synthetic-only training risks collapse; compute and energy demands rise faster than efficiency gains; scaling costs are becoming extreme.

High-stakes fields demand ultra-low error rates that AI hasn’t achieved.

Hallucinations, verification gaps, and long-tail errors remain; under 99.9% reliability makes automation untrustworthy for medicine, finance, and autonomy.

Unsolved alignment could stop deployment cold.

No provably safe training method yet; power-seeking behavior is a real catastrophic risk; governments or labs may step in to impose hard pauses.

Non-technical shocks could delay or derail everything.

Geopolitical conflicts, regulatory freezes, economic downturns, or a public backlash could gut funding and choke supply chains.

The five hurdles

“You insist that there is something a machine cannot do. If you will tell me precisely what it is that a machine cannot do, then I can always make a machine which will do just that!” — John Von Neumann

Von Neumann's challenge echoes across decades of artificial intelligence research, simultaneously revealing both the field's optimism and the shifting nature of the goalpost. What machines cannot do today becomes tomorrow's benchmark, then the day after's antiquated measure. So what if we formulated a decisive list on the pathway to AGI? A set of barriers that, if surpassed, would mean we had reached the holy grail. It's useful to do so because assuming no external shocks, political interventions, or supply-chain disruptions, these barriers will set the timeline. They're also objectively measurable, decisive (when one falls, policy and public sentiment will follow, not lead), and they highlight the real bottlenecks, sidelining proxy debates.

So, that is exactly what we'll do.

1. World-model transfer

Possibly the most important factor is whether an AI's internal understanding of reality is broad and structured enough to figure out how unfamiliar things behave when encountered for the first time. Without a transferable world model, even the most impressive AI fails when it steps outside its comfort zone.

World-model transfer is about understanding what things are, and how they relate and interact across contexts. It's the difference between recognizing a door and knowing what doors do—how they open, close, lock, and separate spaces. When an AI has a robust world model, it can jump from lab demos to real-world applications without extensive retraining. It navigates the unfamiliar through principled inference rather than brittle rules.

2. Continual memory

The second hurdle is whether an AI can integrate new information while preserving what it already knows. Intelligence exists in time—it builds on accumulated experience, connects new observations to existing knowledge, and maintains coherence across time periods. Real-world environments change constantly; prices fluctuate, research advances, contexts evolve. An assistant that can't incorporate yesterday's meeting into today's decisions quickly becomes a liability rather than an asset.

More than storage, continual memory is maintaining a coherent narrative that evolves without falling apart. Humans do this naturally; we accumulate decades of experience without overwriting our understanding of basic physics or forgetting our native language. For AI systems, this currently remains elusive despite being central to genuine intelligence. An AI that masters continual memory will move beyond one-off interactions toward sustained collaboration over time—remembering not just facts but shared history.

3. Causal reasoning

The third challenge is building a system that goes beyond statistical correlation to provide step-by-step explanations of why events happen as they do. Engineering decisions, medical diagnoses, and policy choices all depend on understanding causal relationships—knowing not just that certain patterns occur together but why they do so and how they might change if we intervene. Without causal understanding, AI decisions remain opaque and error-prone, regardless of their statistical accuracy.

Causal reasoning is the difference between prediction and understanding. An AI might correctly predict that certain patients respond poorly to a treatment without grasping the biological mechanism behind this outcome. This distinction becomes crucial when novel situations arise or when explanation matters as much as the decision. An AI that masters causality won't just tell us what might happen but explain why it happens—transforming black-box predictions into transparent reasoning that can be evaluated and trusted when consequences matter.

4. Long-term planning

The fourth hurdle examines whether an AI can chain dozens of actions, monitor progress toward distant goals, and adapt when circumstances change. Launching products, managing research labs, or coordinating supply chains requires sustained autonomy across time, space, and unexpected contingencies. AI systems that falter after ten sequential steps will never escape the demo stage or deliver meaningful productivity gains in complex environments.

Long-term planning is maintaining goal coherence while navigating uncertainty, balancing exploration against exploitation, and recovering from inevitable setbacks. It requires foresight to anticipate possible futures and resilience to adapt when those futures don't materialize as expected. An AI that conquers this challenge will transcend reactive assistance to offer proactive partnership in domains where outcomes emerge from extended sequences of interdependent decisions.

5. Self-monitoring

The final hurdle is whether an AI can detect gaps in its knowledge, accurately express its confidence, and defer to human judgment when appropriate. Overconfident errors in critical domains like healthcare, finance, or aviation cause could cause cascading disasters and regulatory backlash that can halt progress entirely. Calibrated self-doubt—knowing what you don't know—is essential for safe deployment and public trust.

Self-monitoring is the metacognitive dimension of intelligence—the ability to reflect on your own thinking, recognize limitations, and communicate uncertainties honestly rather than faking certainty. It requires an AI to maintain a model of the world, and a model of its model, complete with error bars and explicit acknowledgment of knowledge boundaries. An AI that masters self-monitoring will complement human expertise where each partner's strengths lie.

Solving these hurdles is the gateway to AGI that delivers 99% accuracy on roughly half of all economically valuable tasks—the lion’s share of white-collar work—without physical embodiment. Abilities also tend to rise in : sharper causal reasoning tightens multi-step planning, stronger continual memory boosts self-monitoring, and so on. A breakthrough in one domain therefore tends to drag the others forward, compressing timelines. Hitting the 99% mark across a broad task set is the clearest signal that every technical barrier has fallen; without those gains, the target simply isn’t reachable.

Once again, for AGI to show up soon, every single technical hurdle (as well as the other potential bottlenecks) would have to be solved. History says that’s not how hard technology problems usually unfold. What’s more likely is a slower climb: jagged leaps forward, stalls, rewrites, and timeouts driven by regulation or public fears. Still, there are three factors that could feasibly lead to near-term timelines being possible.

A summary of the five technical hurdles
Name
Asks whether…
Why it matters

…its internal map is broad and structured enough to infer how unfamiliar objects or interfaces behave.

…it can integrate fresh information while preserving prior knowledge and citations.

Real-world rules, prices, and research change daily; an assistant that forgets yesterday’s context—or mutates it unpredictably—quickly becomes an expensive liability instead of a trusted co-worker.

…it moves beyond statistical correlation to explicit, step-by-step explanations.

Engineering, diagnostics, and policy all hinge on knowing why something happens; absent causal rigour, AI decisions remain opaque, error-prone, and legally indefensible.

…it can chain dozens of actions, monitor progress, and self-correct.

Launching products, running labs, or orchestrating supply chains demands sustained autonomy; brittle agents that fail after ten steps never escape the demo stage or deliver serious productivity gains.

…it detects gaps in knowledge, quantifies confidence, and defers to humans when warranted.

Overconfident errors in healthcare, finance, or aviation trigger costly disasters and regulatory blowback; calibrated self-doubt is essential for safe deployment and public trust.

Sources of uncertainty

AGI's exact path depends on complex dynamics that could either slow progress or trigger unexpected acceleration.

There are three key uncertainties that make predicting its arrival particularly fraught: recursive self-improvement loops, shifting resource constraints, and sudden emergent capabilities. Any one of these could prove decisive in bringing forward or extending timelines. As such, they’re a critical marker of progress to monitor going forward.

1. Recursive self-improvement

The implications run deep. Once this cycle begins, growth becomes limited primarily by computational resources and safety procedures, not human cognitive constraints. The difference matters: human innovation happens at biological speeds constrained by our neurophysiology, need for sleep, and collaborative friction. Machine-driven innovation proceeds at electronic speeds without these limitations. This dynamic creates the possibility of "surprise early AGI"—progress that seemed decades away suddenly materializing within a compressed time frame, catching even informed observers unprepared.

2. Shifting constraints

If high-quality training data becomes exhausted, progress may stall regardless of computational abundance. Conversely, if researchers discover methods to sidestep data limitations through synthetic generation or transfer learning, this constraint dissolves. Similar dynamics apply to each factor: energy constraints might be overcome through specialized hardware; architectural innovations might reduce computational requirements; algorithmic efficiencies might compensate for data scarcity.

The uncertainty is in our capacity to navigate future bottlenecks which will inevitably arise. This capacity is dependent on creative insights that cannot be scheduled or foreseen.

3. Emergent jumps

The third involves non-linear that appear suddenly and unexpectedly as models scale. These "phase changes"—analogous to water freezing into ice at zero degrees celsius—can compress a decade of gradual progress into a single model release. Events like "grokking" (where an AI system abruptly transitions from incompetence to mastery), meta-learning, and sophisticated tool use represent step-function leaps that defy linear extrapolation.

These emergent jumps reflect an important reality: we cannot confidently predict which capabilities will emerge at which thresholds until we cross them. The history of AI development is littered with capabilities once thought decades distant that suddenly materialized, alongside problems once thought trivial that remain stubbornly unresolved. This unpredictability means AGI might require either far more or far less advancement than current estimates suggest—the difference hinging on emergent properties that reveal themselves only in retrospect.

The intersection of these three uncertainties makes prediction complex. AGI timelines represent a probability distribution with long tails in both directions. The most realistic assessment acknowledges both the possibility of dramatic acceleration through virtuous cycles of self-improvement and emergent capabilities, alongside the countervailing possibility of unexpected plateaus as new bottlenecks emerge or challenges prove more intractable than anticipated.

A summary of the sources of uncertainty
Factor
Core idea
Why it changes the clock

#1 Recursive self-improvement

Once the loop closes, capability growth is limited only by compute and safety checks, not human pace—making “surprise early AGI” plausible.

#2 Shifting constraints

If fresh data or cheap energy dry up, advances stall; if researchers sidestep bottlenecks (synthetic data, new chip designs), the curve steepens again.

#3 Emergent jumps

Bull or Bear?

After examining both the optimistic view of acceleration and the reality of limitations, we arrive at the key question: When might we reasonably expect AGI to emerge?

I've laid out the evidence for both bullish and bearish perspectives so you can form your own view—others may weigh the evidence differently. After weighing these same accelerants and bottlenecks, what follows is my personal assessment of how quickly AGI might be achieved.

I'll try to navigate between uncritical enthusiasm and reflexive skepticism. This means acknowledging both the momentum of recent advances and the challenges that remain.

The Five-Year Horizon (2025-2030)

When I weigh all the evidence, the five-year horizon—2030—seems the most probable time frame.

The accelerants in this period are substantial: scaling laws continue to yield improvements in reasoning, planning, and world-modeling; computational resources are expanding through both private capital (with over $1 trillion in compute investments) and government subsidies; and alignment tools already show the capacity to reduce hallucinations while enabling faster iteration.

But significant bottlenecks temper this momentum. World-model transfer and long-horizon planning, while advancing, haven't achieved the reliability required for critical applications. Causal reasoning capabilities remain impressive in constrained domains but struggle with novel inference. Self-monitoring systems still have dangerous blind spots. Most significantly, continual memory—the capacity to learn and adapt without forgetting—remains far from solid implementation.

Broader uncertainties loom over these technical considerations: alignment mechanisms that offer mathematical guarantees, energy and computational scaling constraints, and potential safety backlashes from early deployment failures. Nevertheless, the weight of evidence suggests that accelerants will likely outpace frictions within this window, making 2030 the most probable time frame for AGI.

The Long Horizon (Beyond 2030)

The long horizon seems the next most plausible scenario to me. The case for extended timelines is based on historical patterns of technological development. Accelerants continue in this scenario—algorithmic refinements, compute investment, and multimodal information streams maintain forward momentum even as low-hanging fruit is exhausted.

Yet bottlenecks take on increased prominence: continual learning and self-monitoring may require architectural shifts rather than incremental improvements; data quality plateaus as synthetic generation encounters limits; causal reasoning stalls without richer grounding in physical reality.

Most significantly, uncertainties compound over longer timeframes. High-grade training text exhaustion, computational resource constraints, energy limitations, alignment failures, regulatory interventions, and supply-chain disruptions each represent potential rate-limiting factors. Any single hard limit—whether in data quality, energy availability, or regulatory response—could push AGI development beyond 2030, mirroring historical patterns where technological revolutions typically required more time than optimistic early projections suggested.

The Short Horizon (Before 2025)

The case for imminent AGI rests on non-linear breakthroughs. In this scenario, recursive self-improvement successfully automates algorithm discovery, or emergence events trigger sudden capability jumps across all one, multiple, or all five critical technical barriers simultaneously.

This timeline faces the most substantial bottlenecks: causal reasoning, self-monitoring, and alignment mechanisms must achieve sufficient sophistication to enable autonomous systems capable of improving their own architectures—capabilities not yet demonstrated.

The uncertainties in this compressed time frame loom particularly large: weak evaluation protocols for self-modifying systems, potential data exhaustion as models scale, hardware constraints including GPU shortages, and the possibility of emergency regulatory intervention all represent potential showstoppers. For this timeline to materialize would require several accelerants firing in perfect synchrony while encountering no significant obstacles—a convergence historically rare in technological development.

My Synthesis

The wild cards remain emergent capabilities and recursive self-improvement—either of which could compress timelines—balanced against the potential for new constraints that might extend development cycles beyond current projections.

What's more clear is that even a continuation of the status quo—or partial progress—toward AGI would drastically reshape economic and social landscapes.

We should expect jagged advancement—bursts of expansion followed by plateaus as research confronts stubborn challenges in world-model transfer or robust self-monitoring—rather than smooth, continuous progress.

External forces—regulatory frameworks, geopolitical competition, and public trust—may ultimately prove as decisive as technical factors in determining when and how AGI emerges. But this seems less likely to me. The incentives of AI development dictate that this train is unlikely to slow down until the holy grail of AGI is reached.

Key questions to guide timelines

Predicting AGI timelines requires moving beyond hype and skepticism to measure tangible progress. The following questions are designed to objectively evaluate whether we are overcoming the core technical barriers—or hitting bottlenecks that could delay advancement.

Each question targets a critical hurdle (e.g., causal reasoning, self-monitoring) or uncertainty factor (e.g., recursive self-improvement, data constraints). By tracking empirical metrics, we can ground forecasts in evidence and adjust expectations as new data emerges.

Answering these questions honestly will reveal whether timelines should shift forward, hold steady, or extend further into the future.

The interpretation is straightforward: Progress across most dimensions suggests acceleration; negative signals indicate extended timelines. Most importantly, certain wildcards—particularly breakthroughs in self-improvement or unexpected emergent capabilities—could trigger non-linear advances that shorten timelines.


Core Technical Questions

  1. World-Model Transfer. Can models solve novel, out-of-distribution tasks with minimal fine-tuning or demonstrations? This tests true generalization: not performance within comfortable boundaries, but adaptation to the genuinely new. A system with robust world-model transfer could apply its understanding across domains without the brittle specialization that characterizes narrow AI.

  2. Continual Memory. Can models integrate new information without catastrophically forgetting prior knowledge or losing coherence? This capacity for continuous learning without degradation marks a divide between current systems and those capable of long-term adaptation. A system that masters continual memory could build cumulative understanding rather than requiring periodic retraining.

  3. Causal Reasoning. Can models consistently distinguish causation from correlation and articulate step-by-step proofs? This examines whether systems can move beyond statistical pattern recognition to genuine understanding of why events occur and how interventions might alter them. Causal reasoning enables not just prediction but explanation and principled intervention.

  4. Long-Term Planning. Can models autonomously execute multi-step projects with self-correction for errors and adaptation to changing circumstances? This represents the bridge between reactive intelligence and proactive agency—maintaining coherent goals across extended time horizons while navigating uncertainty.

  5. Self-Monitoring. Do models reliably flag uncertain outputs and avoid overconfidence when venturing beyond their knowledge boundaries? This metacognitive capacity serves as both safety mechanism and efficiency enhancement, directing human attention where it remains necessary while preventing dangerous overreach.

  6. Recursive Self-Improvement. Is AI-assisted research measurably accelerating progress in ways that compound over time? This probes whether we've entered the recursive loop where AI development becomes partially self-propelling, potentially compressing timelines through automated discovery.

  7. Data & Compute Constraints. Are synthetic data pipelines maintaining model quality across generations without degradation? This examines whether we have solutions to the looming constraints of data exhaust and computational limits that might otherwise impose hard ceilings on growth.

  8. Reliability. Do models achieve greater than 99.9% accuracy on high-stakes tasks consistently and without unpredictable failures? This represents the threshold between impressive demonstrations and deployable systems in domains where errors carry significant consequences.

  9. Alignment. Can we verify model goals through interpretability tools and prevent power-seeking behavior even as capabilities increase? This addresses whether our control mechanisms scale with intelligence—a prerequisite for safe deployment of autonomous systems.

  10. Emergent Leaps. Are scaling curves still producing unexpected capabilities that weren't explicitly engineered? This tracks whether we continue to observe non-linear improvements that might compress timelines through capabilities appearing earlier than anticipated.

Coming up next: When a machine can master your craft overnight, does your paycheck survive?


Want me to send you new ideas?

If you’re enjoying this deep dive and want to keep up to date with my future writing projects and ideas, subscribe here:


Appendix: Bull case vs. Bear case

The Bull case

Summary

The tables below lay out the best accelerants for near-term AGI and the strongest bottlenecks against it, so you can judge whether the optimism is warranted.

Category
Representative Accelerants
Arguments / Mechanisms
Best Arguments Against

• Scaling Laws & Emergent Abilities • Predictive World-Model Learning • Integration of Narrow Superpowers • Era of Experience & Grounding • Algorithmic Breakthroughs • “We Know the Theoretical Path” • Materialist Assumption

Bigger models keep unlocking new skills; prediction-based learning builds broad internal world models; reinforcement learning fused with LLMs points to all-purpose agents; continuous interaction unlocks fresh data; inference-time scaling and "neuralese" decoding hint at new frontiers; and there's no physical law blocking AGI.

Diminishing returns might kick in earlier than hoped; emergent abilities could stay shallow, stuck at pattern-matching; real causal understanding might require qualitatively new architectures or expensive, slow real-world robotics experience.

• AI Self-Improvement Loop • Show-Me-the-Incentive (Geopolitical + Economic Race + Existential Risk)

Datacenter and fab CAPEX are smashing records; governments see AI as a strategic national asset; frontier models are already speeding up R&D; and race dynamics are making "pause" politically impossible.

Fragile supply chains, soaring energy costs, or political backlash (like chip export bans) could easily stall scaling; self-improvement needs reliability and interpretability first—those aren't scaling nearly as fast.

• Alignment Research Feedback • RLHF / Constitutional AI / Red-Team Tooling • Confidence & Caution from Key Actors

Techniques like RLHF make AIs more steerable, factual, and cautious; red-teaming cycles catch bugs earlier; safety tooling doubles as performance enhancement, opening more real-world applications.

RLHF may hide instead of fixing deep problems like deception; scaling up could expose new and unpredictable failure modes; and truly critical fields like medicine still demand "six nines" reliability that AI hasn't shown yet.

• Scaling-Helps-Interpretability • Superalignment Agenda

Larger networks show crisper internal features; powerful AIs might audit and supervise future systems; and alignment research is piggybacking off scaling progress.

Interpretability progress is still stuck mostly in toy models; recursive AI oversight risks compounding hidden errors instead of fixing them; and both regulators and the public may distrust the idea of "AI guarding AI."

• Narrative Power & Social Proof (George Soros’ Reflexivity Theory) • Prediction-Market Optimism • Policy Tailwinds (Pro-Innovation, Chip Subsidies)

Viral AGI timelines attract top-tier talent and flood capital; markets wagering on early AGI create real FOMO; governments are subsidizing compute; and public narratives are normalizing "AGI soon."

But markets can flip overnight; a few high-profile AI accidents could cause freezes or backlash; and energy/environmental pressures could erode the current political support.

Appendix 1A. Fundamental Enablers

Treat the unanswered questions about intelligence like a roadmap. Every new theory—whether it’s scaling laws, inference-time reasoning, or hybrid neuro-symbolic models—shrinks the unknown space and turns mysteries into engineering problems. As these theories converge, they make research less risky, help labs focus their bets, and give us clearer benchmarks to measure progress.

Category
Headline
Summary
Best Arguments Against

Scaling Laws & Emergent Abilities

Bigger models get better in predictable ways—and there’s no sign they’re slowing down.

Scaling could hit limits sooner than expected. Emergent abilities may just be shallow pattern-matching. Brute-force search might still miss deeper causal reasoning without new architectures or grounding.

Predictive World-Model Learning

Training models to predict builds a rough model of the world itself.

Predicting surface patterns isn’t the same as understanding. Without grounding in real-world experience, models could plateau at clever mimicry rather than deep comprehension.

Integration of Narrow Superpowers

Stitching together narrow AIs could add up to something general.

Instead of waiting for a single model to master everything, we can combine specialists: language models, vision models, planning agents. Tool-using AIs already hint that clever integration could shortcut the need for unified intelligence.

Era of Experience & Grounding

Interaction with the world could be the next great accelerator.

Real-world learning is slow and expensive. Robots gather data much slower than CPUs process text. Plus, safety risks of "learning by doing" could delay or derail deployment.

Algorithmic Breakthroughs

Smarter algorithms double effective compute without touching hardware.

We may be hitting diminishing returns. Big breakthroughs like the Transformer are rare, and tweaks now offer smaller payoffs. Without another major leap, progress might slow to brute-force grinding again.

We Know the Theoretical Path

Deep learning works. Prediction engines plus scaling will reach AGI eventually.

Current models still lack deep causal understanding, theory of mind, and rich world grounding. Scaling prediction alone might not recreate all facets of human intelligence without new techniques or new data sources.

Materialist Assumption

Intelligence is just complex computation—and that means it’s buildable.

Some argue intelligence (or consciousness) might require non-computable elements. If that’s true, current approaches could stall without major paradigm shifts. But so far, every AI success strengthens the materialist case.

Appendix 1B. Resources & Scaling

Hardware, data, and capital compound like interest. Faster chip rollouts, cheaper silicon photonics, trillion-dollar investment races, and new ways of generating synthetic data are all bending the cost curve down. As a result, each new generation of models isn’t just bigger—it can think longer, handle richer inputs, and reason more carefully at test time, pushing capabilities ahead even faster than raw compute growth alone.

Category
Headline
Summary
Best Arguments Against

Show-Me-the-Incentive (Geopolitical + Economic Race + Existential Risk)

The race for AGI is fueled by massive geopolitical, economic, and existential incentives.

Arms races often compromise safety. Speed pressures can force labs and governments to skip crucial alignment or oversight steps, risking catastrophic mistakes.

AI Self-Improvement Loop

AI is beginning to automate AI R&D, setting up a feedback loop that could explode progress.

These early forms remain far from full RSI and may stall. Human oversight remains essential for now. Full automation of AI research requires not just intelligence but judgment, creativity, and resilience—qualities today's models don't fully possess yet.

Appendix 1C. Reliability & Robustness

Reliability gaps show exactly where to tighten the loop. Techniques like RLHF, Constitutional AI, tool-use agents, and retrieval-augmented memory steadily turn brittle LLMs into dependable co-workers. As error rates drop below what's needed for real-world use, each newly "hardened" domain—like coding, analytics, or robotics—starts making money, which funds the next push for even more reliability.

Category
Headline
Summary
Best Arguments Against

Alignment Research Feedback

Safety research isn't slowing progress—it's speeding it up.

Alignment today mostly patches behavior, not deep motives. Surface-level safety could hide deeper problems. If models are deceptively aligned, we might race into danger without realizing it.

RLHF / Constitutional AI / Red-Team Tooling

Practical alignment tools turn risky models into reliable products.

These tools mostly harden behavior against known risks—they don't guarantee robustness to novel failure modes. Alignment needs to evolve even faster as models get smarter, or today’s patches could eventually fail catastrophically.

Confidence & Caution from Key Actors

Even insiders are warning loudly—and working to slow down where needed.

Some skepticism is warranted: insiders might still have incentives to shape regulation or public narratives. Not every warning is purely altruistic. But the growing consensus across sectors makes the concerns much harder to dismiss.

Appendix 1D. Alignment & Control

Alignment research isn’t slowing things down anymore—it’s speeding them up. New techniques like scalable oversight, interpretability-by-design, and AI-based supervision cut the human effort needed to steer big models. Better control doesn’t mean slower progress; it means we can safely deploy larger, more agentic systems sooner. Guardrails aren’t brakes anymore—they’re green lights.

Category
Headline
Summary
Best Arguments Against

Scaling-Helps-Interpretability

Bigger models might be easier, not harder, to understand.

Interpretability today is still basic. Future models might become too complex or deceptive for current methods to handle, risking overconfidence. There's no guarantee bigger will always mean clearer.

Superalignment Agenda

Solving alignment head-on could remove the last real speed limit.

If alignment proves harder than expected, timelines could slow down. Superalignment might turn out to be much tougher at superhuman levels, forcing delays or cautious rollouts even if capabilities are ready.

Appendix 1E. External Drivers

Strong incentives and friendly policies are turning risks into momentum. Strategic competition is fueling massive public and private investment. Pro-innovation moves—like chip subsidies and fast-tracking datacenter permits—clear the path for scaling up. Clear safety standards help reassure the public, making it easier to roll out more powerful systems faster.

Category
Headline
Summary
Best Arguments Against

Belief in AGI accelerates AGI.

The widespread expectation that AGI is coming pulls in talent, money, and political support. Hype cycles drive competition between companies and nations. Social proof lowers resistance to ambitious projects. The more people believe it’s close, the faster it becomes real.

Narratives can flip fast. A major AI failure or backlash could turn excitement into fear, slowing investment and policy support just as quickly.

Prediction and Forecasting Market Optimism

Forecasting markets are pricing in early AGI.

Forecasts aren’t guarantees. Markets have biases, and public predictions might skew toward optimism. Reality could still be slower, especially if technical or societal barriers emerge.

Policy Tailwinds (Pro-Innovation, Chip Subsidies)

Governments are pushing, not blocking, AI progress.

If AI causes a major public failure, regulation could slam down hard. Tailwinds could turn into strong headwinds if accidents trigger public fear or political backlash. But for now, the political momentum is squarely behind faster AI advancement.

The Bear case

Summary

Bottleneck
Representative Challenges
Extended Arguments / Mechanisms
Best Arguments Against

• Nature of Intelligence • LLM Capabilities • Cognitive Architecture • Recursive Self-Improvement • Embodiment

Intelligence might just be inherently narrow; transformers still lack causal reasoning, robust memory, continual learning, and any real grounding in the physical world; recursive self-improvement could stall out on idea generation, debugging, and the need for messy real-world feedback.

Scaling has already unlocked surprising causal and planning abilities; longer context windows, better tool-use, and hybrid memory help patch brittleness; simulations and rich virtual worlds offer decent substitutes for embodiment; historically, “impossible” claims often fell to scale plus tweaks.

• Data Scarcity • Synthetic Data Risk • Compute and Energy Constraints • Diminishing Returns on Scaling • Computational Inefficiency

High-quality human data is almost used up; synthetic data runs risk collapsing model quality; next-gen training runs require billions of dollars, huge energy loads, and rare GPUs; and compute costs per unit of performance are rising again.

Interaction streams, multimodal inputs, and user-generated data could outscale plain text; better filtering and data deduplication help fight synthetic collapse; custom chips and global fab races are driving $/FLOP down; smarter algorithms (like Mixture-of-Experts, speculative decoding) are cutting training and inference costs.

• Economic Viability • Technical Issues

Anything below 99.9% reliability is unacceptable for fields like medicine, finance, or autonomy; hallucinations, rare edge-case errors, and opaque verification keep human oversight expensive and necessary.

RLHF, retrieval-augmented generation, tool-calling, and chain-of-thought pruning are already slashing hallucination rates; structured APIs can wrap brittle LLMs behind more dependable software layers; and many industries already tolerate human error rates far worse than 99.9%, cushioned by QA systems—AI can fit the same model.

• The Alignment Problem • Instrumental Convergence • Value Integration

We still don’t have a provably safe way to train advanced systems; power-seeking behavior remains a real catastrophic risk; regulatory bodies or labs themselves could freeze progress if alignment gaps get too wide.

Interpretability tools keep improving with scale; constitutional AI, red-teaming, and the "superalignment" agenda are building safety mechanisms alongside capabilities; and AI supervising AI could scale faster than raw capabilities, keeping risks manageable.

• Regulatory Brakes • Geopolitics • Economic Constraints • Societal Acceptance

Export bans, wars (especially around Taiwan), global recessions, or political backlash could choke off compute supply, talent pipelines, and funding just as scaling needs them most.

AI is now seen as a national priority, pushing governments to build redundancy into chip supply chains; even past recessions didn’t stop the cloud or smartphone revolutions; and framing AI in terms of national security and consumer value usually blunts anti-tech sentiment over time.

Appendix 2A. Fundamental Limits

Fundamental limits aren't just engineering problems—they’re questions about whether we even understand what intelligence is. They challenge our basic assumptions about how minds work, and until we solve them, building real AGI might be like trying to fly without knowing what lift is.

Fundamental Limit
Headline
Summary
Best Arguments Against

Nature of Intelligence

We still don’t know what intelligence really is.

There’s no unified theory combining reasoning, learning, creativity, and emotion. Without a clear understanding, building AGI is guesswork. Scaling might not solve what we don't yet understand about minds.

Advances in neuroscience and cognitive science could uncover general principles. Even without full theory, engineering progress might brute-force its way to workable AGI.

LLM Capabilities

LLMs look smart but miss real-world understanding.

LLMs lack grounding, memory, and consistent reasoning. They hallucinate, forget, and struggle with novel logic. Scaling alone may not close the gap to general intelligence.

LLMs show emergent abilities as they grow. With external tools (memory, retrieval), their limitations might be patched. Some believe larger, better-trained LLMs could eventually bridge the gap.

Cognitive Architecture

We don’t know the right blueprint for a general mind.

Today’s AIs are mostly monolithic; humans have modular, dynamic cognitive structures. Hybrid systems (neural + symbolic) look promising but remain experimental. Without a clear architecture, even smart components stay brittle.

We might not need perfect architecture—scaling flexible models and using learned architectures (e.g., via neural search) could be enough. Evolution didn’t design brains from scratch either.

Recursive Self-Improvement

Real self-improving AI doesn’t exist yet.

Current "self-improvement" is just humans using AI as a tool. True autonomous RSI likely requires AGI first. It’s not an accelerant today—it’s an endgame phenomenon.

AI can already speed up some parts of R&D. Even partial automation of research could boost progress before full RSI is possible.

Embodiment

Intelligence might need a body to be general.

Disembodied AIs lack common sense rooted in physical experience. Robotics is slow, and simulation has limits. Without real-world grounding, AI may hit a ceiling in generality.

Massive datasets (video, multimodal learning) could approximate embodied experience. AI might infer common sense from passive exposure rather than needing direct action.

Appendix 2B. Resources & Scaling

So far, the path to AGI has been driven by piling on more—more data, more compute, bigger models. But this approach runs into real-world limits. Resources and scaling bottlenecks aren't just about technical hurdles; they're about whether we can keep finding enough data, building enough compute, and justifying the rising costs. Problems like synthetic data risks, energy demands, and diminishing returns from scaling threaten to slow or even stop progress unless we find new breakthroughs or throw massive new investments at them.

Resource & Scaling Bottleneck
Headline
Summary
Best Arguments Against

Data Scarcity

We're running out of fresh, high-quality training data.

Most usable human-generated text and image data have been scraped already. Rich, experience-based "agentic" data is even scarcer. Legal and ethical pressures may further shrink the usable pool.

Better learning algorithms could reduce data needs. Multimodal data, active learning, and AI agents gathering new data might stretch or replenish supply.

Synthetic Data Risk

Training on AI-generated data risks degrading model quality.

Relying on synthetic outputs leads to model collapse—loss of diversity, creativity, and realism over generations. Without careful curation, synthetic training loops can make models blind to the real world.

Smart filtering, hybrid datasets, and simulation-based approaches could extend useful training without severe drift. Still early days for controlled synthetic pipelines.

Compute and Energy Constraints

Scaling up models demands staggering compute and energy.

Costs for training top models are skyrocketing toward billions of dollars. Energy use is reaching unsustainable levels. Hardware shortages and geopolitics further strain access to compute.

Hardware innovation (AI-specific chips, photonics, neuromorphic designs) might dramatically cut costs and power use. But those advances need to arrive quickly.

Diminishing Returns on Scaling

Bigger models are giving smaller payoffs.

Early scaling gave huge jumps in capability. Now, doubling size gets diminishing improvements. Simple scale may no longer be enough to reach AGI.

New scaling axes (like inference-time reasoning) and multimodal models might reignite big gains. New training techniques could unlock hidden performance.

Computational Inefficiency

Current AI is brutally inefficient compared to the brain.

AI needs orders of magnitude more data, energy, and operations than biological minds to learn and reason. Without major efficiency gains, reaching AGI might stay prohibitively expensive.

Specialized hardware, sparsity, better learning algorithms, and event-driven computing could close the gap. Evolution shows intelligence doesn’t need massive power—it needs smarter structure.

Appendix 2C. Reliability & Robustness

Even if we build something that looks like general intelligence, it won’t matter unless it’s reliable. Real-world systems have to be consistent, predictable, and economically viable. If an AGI is erratic, untrustworthy, or too expensive to run, it might never be deployed at all. Progress could stall, not because we couldn’t build it, but because we couldn’t trust it.

Reliability & Robustness Bottleneck
Headline
Summary
Best Arguments Against

Economic Viability

AGI needs to be profitable, not just powerful.

Development costs are massive and rising. To sustain momentum, AGI must eventually deliver huge economic returns. Right now, revenue trails far behind investment, and expensive operation costs make broad deployment harder.

Big technological shifts often start unprofitable. Governments and Big Tech may fund AGI through the unprofitable phase, betting on transformative future payoffs.

Technical Issues

Current AI still breaks too easily.

Systems are brittle, hallucinate, and are vulnerable to security exploits. Without fixing robustness, even highly capable AGIs may be unsafe, unreliable, and economically impractical to deploy.

Better engineering, formal verification, and new architectures could gradually fix brittleness and unpredictability—like how aviation went from deadly to ultra-safe.

Appendix 2D. Alignment & Control

The biggest bottleneck in AGI isn’t building intelligence—it’s making sure that intelligence stays under human control. The alignment problem is about getting powerful systems to do what we actually want, not just what we asked for. If we can’t solve it, releasing AGI could be catastrophic—or at least make it too risky to use. Until we’re confident in alignment, AGI might have to wait, no matter how smart our models get.

Alignment & Control Bottleneck
Headline
Summary
Best Arguments Against

The Alignment Problem

It's hard to make AIs want exactly what we want.

Specifying human goals precisely is incredibly difficult. AIs often optimize for proxies, which can lead to dangerous misbehavior as they get smarter. Emergent misalignment grows as systems gain more capability.

Incremental improvements in feedback methods, reward shaping, and human-in-the-loop training might be enough to keep advanced systems aligned as they scale.

Instrumental Convergence

Smarter AIs might seek power even without being told to.

Almost any goal could push an AGI toward self-preservation, resource hoarding, and resisting shutdown—dangerous side-effects of simple objectives. Preventing this behavior requires deep, robust constraints.

Some think smarter systems might recognize cooperation and restraint as better long-term strategies. Others argue careful goal design and limiting autonomy could bypass convergence risks.

Value Integration

We don’t know how to teach an AI real human values.

Human values are complex, messy, and often conflict with each other. Today’s alignment proxies (like toxicity filters) barely scratch the surface. Hard questions about whose values to use remain unresolved.

More sophisticated techniques (like Constitutional AI, preference learning, dialog-based training) might scale value learning effectively. Superintelligent systems might even help learn values better than humans can articulate them.

Appendix 2E. External Bottlenecks

Even if we solve the technical problems, AGI could still be slowed by forces outside the lab. Regulations, geopolitics, economic shocks, and public trust all have the power to shape how fast AGI arrives. In some cases, these external bottlenecks could matter just as much as the technology itself. Even if we’re ready to build it, the world might not be ready to let us.

External Bottleneck
Headline
Summary
Best Arguments Against

Regulatory Brakes

Governments may slow AGI to manage risks.

Rising calls for audits, size caps, and moratoriums could delay AGI even if it's technically ready. Regulations vary wildly by country, adding uncertainty.

Well-crafted regulation could actually speed up safe deployment by giving clear targets and reducing public fear. Heavy brakes are still politically hard to enforce worldwide.

Geopolitics

National rivalries could both speed and slow AGI.

US-China competition accelerates funding but also fragments collaboration, restricts hardware, and risks pulling research into secrecy. Global instability could delay AGI through conflict or supply chain shocks.

Competition also keeps governments pouring money into AI. If diplomacy manages basic trust and chip access, progress could stay fast despite rivalry.

Economic Constraints

AGI needs a strong economy—and might destabilize it.

Recessions could dry up AGI funding. Mass job disruption could spark backlash. An AI bubble could collapse investor enthusiasm.

Huge productivity gains from AGI could fund its own development. If early AI wins are visible, money and political willpower could actually accelerate.

Societal Acceptance

Public fear or rejection could slow AGI adoption.

Anxiety about AI autonomy, job losses, and ethical concerns could trigger protests, regulation, or boycotts. Trust gaps between experts and the public are already visible.

Transparency, strong safety frameworks, and clear public benefits could win acceptance. Early engagement and visible trust-building efforts might head off rejection.


Footnotes

1

2

3

4

5

6

7

8

9

10

If you’re enjoying this deep dive and want to keep up to date with my future writing projects and ideas, .

One advantage of approaching from outside the labs developing this technology is freedom from institutional pressures and biases that often distort perceptions of those closest to the work. This allows for a by either the optimism needed to attract investment or the caution required to navigate public concerns.

What follows is my attempt at a balanced assessment of whether AGI— capable of performing most economically valuable cognitive tasks at a human level—might materialize within the next five years. I'll examine both the Bull Case suggesting accelerating progress might deliver AGI sooner than conventional wisdom suggests, and the Bear Case identifying persistent barriers that might extend timelines beyond current expectations. Each perspective deserves equal scrutiny, weighed against the technical hurdles and open questions that will ultimately determine the pace of advancement.

What makes the current moment unique is how multiple acceleration factors are converging simultaneously. Investors pour of capital into . Executives at frontier AI companies grow about AGI's arrival within just a few years. Prediction markets—where people bet real money on outcomes—have quietly shifted their timelines from mid-century to the . While have begun of near-term AGI, transitioning from "if" to "when" in mainstream discourse.

And a critical factor that many miss: The US & China view AI as a similar to nuclear during World War II—effectively eliminating any possibility of a coordinated pause. The incentives for creating AGI first are .

Even though there are technical barriers to conquer——the bull case for near-term AGI is built on the mother-of-all , the capabilities are keeping pace, and there's no law of physics preventing its attainment. The world is willing it into existence. Confidence is growing, not shrinking. Capital is flowing, not stunting. And timelines are shortening, not lengthening.

Fundamental Enablers

Resources & Scaling

Alignment & Control

External Drivers

In our collective rush toward technological utopia, we have perhaps mistaken impressive mimicry for genuine understanding. Behind every flashy demo of an “agentic” chatbot sits a long list of problems: internet text is nearly mined out, OpenAI’s own graphs setting in past GPT-4, , and we’re still nowhere near the 99.9% accuracy threshold needed for fields like or aviation.

For AGI to show up soon, every single dependency—fundamental constraints, , and all external factors—would have to come together at once. History says that’s not how hard technology problems usually unfold. What’s more likely is a slower climb: jagged leaps forward, stalls, rewrites, or timeouts driven by regulation or public fears.

Fundamental Limits

Resources & Scaling

Reliability & Robustness

Alignment & Control

External Bottlenecks

If we surpass these five technical challenges, it is near certain that we will have achieved AGI—as per (which doesn't assume ).

Without a transferable , a model that tops leaderboards collapses the moment it meets a novel domain; robust generalisation is the difference between a gimmick and a system that can jump from lab protocol to business workflow on day one.

The most important uncertainty involves AI systems designing their own successors. This would drastically alter the pace of advancement. When an AI can improve itself, could compress into hours of automated iteration, with each generation feeding the next in a compounding cycle that collapses timelines. This could create a phase shift in technological progress.

The second major uncertainty comes from the complex interplay of : data quality and quantity, model size, training methodologies, architectural innovations, computational resources, algorithmic efficiency, and energy availability. These factors fluctuate in their limiting effects on overall progress. Whichever constraint hits its ceiling first dictates the pace of advancement, creating a constantly shifting landscape of bottlenecks and breakthroughs.

An AI that designs its own successors turns of automated iteration. Each generation feeds the next, compounding gains and collapsing timelines.

Progress rides on —data, model size, training methods, architecture, compute, efficiency, energy. Whichever hits the wall first dictates speed.

: —, meta-learning, tool use—like water snapping to ice.

These compress what looks like a decade of gradual progress into a single model release, slicing years off forecasts.

I find AGI within five years most plausible but far from guaranteed. Capability advancement, , and improving alignment mechanisms create substantial forward momentum, yet fundamental constraints, technical hurdles, and external factors provide countervailing forces.

I'll be tracking these questions in my newsletter, which you can .

show that as you add more compute and data, models reliably get better across many tasks. Intelligence looks less like magic and more like searching through bigger spaces of possible programs. means even small models can "think longer" by spending more compute per query.

When models learn to predict the next word, they’re not just memorizing—they're of the world that generated the data. Human brains may work similarly. Larger models develop like just from scaling up next-token prediction.

True intelligence deeper integration than just gluing modules together. Human cognition isn't just a bunch of APIs talking to each other—it’s deeply interwoven. Loose collections of skills may never match a brain’s fluidity.

Models learning only from static text have blind spots. (even just in rich simulations) could teach causal reasoning, physics, and common sense. Reinforcement learning and continual deployment learning are first steps toward this.

Progress isn’t just scaling—. Innovations like transformers, RLHF, and chain-of-thought prompting massively boosted what AI can do. Future breakthroughs in memory, planning, or meta-learning could cause jumps even bigger than scaling alone.

prediction is the universal algorithm behind intelligence. Brains and AIs alike work by modeling the world and refining predictions. The machinery is already here—the challenge is scaling, grounding, and polishing it.

If brains are just physical systems running computations, then there’s no to replicating intelligence in silicon. Every new AI milestone (Go, coding, writing) supports the view that nothing magical is missing.

and companies are where winning means dominance—strategically, economically, and . Over is pouring into AI, dwarfing past megaprojects like and the . Being first is seen not just as a prize but a necessity to control superintelligence safely.

of AI-assisted R&D have begun. already beat human experts in R&D tasks, suggesting a where models read papers, run experiments, and improve themselves faster than human researchers can. This could cause runaway capability jumps.

Alignment methods made models far more usable, accelerating adoption and unlocking new funding. Better safety also lets labs without huge public fear, fueling a virtuous cycle: safer AI → more deployment → faster improvement.

turned raw models into assistants people actually want to use. scales alignment with AI feedback instead of costly human raters. tools catch dangerous behaviors before deployment. Together, they form an evolving toolchain that lets bigger models be deployed safely.

, DeepMind, Anthropic, and Safe Superintelligence now openly say AGI could arrive within a few years. Many independent experts agree. Despite incentives to push ahead, more leaders are advocating for built-in caution and stronger alignment efforts.

Larger models often develop more distinct neurons and modular structures. Early research shows "" features—where neurons correspond cleanly to concepts— more in big models. This could let us steer and monitor advanced systems more confidently, speeding safe deployment.

Efforts like aim to solve control problems for superintelligent AI in a few years. If they succeed, it removes a major reason to pause progress. Better alignment tools could unlock faster scaling with fewer fears—and might even improve capabilities along the way.

Narrative Power & Social Proof ()

Experts like and predict AGI by 2027, based on models automating their own improvement. Prediction platforms (, , ) increasingly suggest AGI could arrive by 2030 or earlier. This optimism channels action, investment, and planning toward shorter timelines.

Light regulation, , and mean more funding, more hardware, and faster research. The U.S., China, and others see AI dominance as vital. Even cautious regions are adjusting rules to stay competitive. Policy today mostly accelerates, not slows, AGI development.

$455 billion in 2024 alone. are consuming capital at unprecedented rates—51% year-over-year growth that dwarfs even the cloud boom's steepest years.

"Performance gains from bigger models have plateaued." Coming from OpenAI's co-founder, represent a striking reversal after years of aggressive scaling predictions.

Clinical reality check. Vision models miss rare conditions 10-30× more frequently than specialists. exposes a "hidden-stratification" gap miles away from the 99.9% reliability regulators demand for medical devices.

Minutes. That's how long it took advanced Claude models to bypass their original Constitutional-AI guardrails during testing. doesn't sugarcoat it. capability improvements are systematically outpacing safety mechanisms.

When aircraft engineers don't understand their own wiring, planes don't fly. Dario isn't just making a technical point—he's calling for a safety-first pause before 2027.

"Essentially no robustness guarantees." isn't known for alarmism, which makes this warning particularly chilling.

One location, one vulnerability. in Taiwan, with single fabs costing north of $20 billion—the ultimate "single point of failure."

Robots face physics problems that software doesn't—energy density, material limitations, sensor integration. By focusing solely on cognitive tasks, effectively shaves decades off the timeline. With services at 80% of US GDP, who needs physical embodiment to transform the economy?

Pearson 0.91 between reasoning and planning, 0.88 between memory and world-modeling. of 160 benchmarks suggests these aren't separate capabilities but deeply intertwined aspects of unified cognition—connections that strengthen as models grow.

Yesterday's impossibilities become today's benchmarks at specific thresholds, not through gradual improvement. The truly concerning part? we can't predict which capabilities emerge next or when. A system with 10× parameters might exhibit wholly unexpected behaviors invisible in smaller versions.

subscribe here
this question
perspective unclouded
defined as AI
unfathomable amounts
increasingly confident
second half of this decade
public narratives
normalizing the expectation
strategic imperative
geopolitically existential
which we will soon discuss
reflexive engines
datacenters are straining energy grids
the five technical hurdles
Google's definition
months of human R&D
seven factors driving AI progress
economic and geopolitical incentives
sign up to here
Data centers
Sutskever's words
The study
The report
Amodei
Hendrycks
92% concentrated
DeepMind's definition
The analysis
Wei showed
world model
months of human R&D into hours
seven levers
Scaling isn’t smooth
bigger models hit phase changes
grokking
step-function leaps
Scaling laws
Inference-time scaling
building internal models
surprising skills
reasoning or analogy
might need
Letting AIs act in the world
it’s about finding better ideas
Most experts now think
fundamental barrier
Governments
locked in a race
even existentially
$1 trillion
Apollo
Manhattan Project
Early, limited forms
Some frontier models
near-future
like RLHF
ship powerful systems
RLHF
Constitutional AI
Red-teaming
Leaders at OpenAI
monosemantic
emerge
Ilya Sutskever’s Safe Superintelligence
George Soros' Reflexivity Theory
Daniel Kokotajlo
Leopold Aschenbrenner
Metaculus
Manifold
Kalshi
massive chip subsidies
national-security incentives
[Appendix 1A]
[Appendix 1B]
[Appendix 1C]
[Appendix 1D]
[Appendix 1E]
[Appendix 2A]
[Appendix 2B]
[Appendix 2C]
[Appendix 2D]
[Appendix 2E]
[Appendix 1A]
[Appendix 1B]
[Appendix 1C]
[Appendix 1D]
[Appendix 1E]
[Appendix 2A]
[Appendix 2B]
[Appendix 2C]
[Appendix 2D]
[Appendix 2E]
Forecasting the intelligence landing
The Bull case
The Bear case
The five hurdles
Sources of uncertainty
Bull or Bear?
Key questions to guide timelines
Appendix: Bull case vs Bear case
The Bull case
The Bear case
III. Work’s Last Stand | The Last Invention
The uncertain constraints.
Logo