29 Comments
User's avatar
joe's avatar

This is really exciting+admirable+valuable!

The core disagreement seems to hinge on what constitutes the primary bottleneck for AI progress.

- AI 2027 -> main bottleneck = internal capability (AI ability to reason, code, do research) which once automated overcomes other obstacles.

- AI as normal tech -> main bottleneck = external (need feedback loop with real world, adoption, infra, solving long tail of errors)

What specific, observable evidence in the next 2-3 years would cause each side to most update their view?

For Arvind and Sayash: Are you assuming we don't see some technique for enabling meta-learning or continual learning? Unclear why one would think the main bottleneck is brute-force iterative feedback otherwise. Also, even if AI does need massive real-world feedback, it's unclear this is a real bottleneck. We live in a world saturated with sensors, deployed AI systems, web-scale human interaction data. To ground out a scenario, do you basically believe it's extremely unlikely we'll have AI systems which handle complex travel booking after seeing 10 examples rather than 10,000?

For the AI 2027 authors, a sharper crux: does your timeline depend on qualitative breakthroughs in AI capabilities, or does current-paradigm scaling suffice? More specifically, if by 2027 AI systems automate 60-70% of AI R&D tasks—handling implementation, optimization, debugging, running experiments—but humans remain essential for creative hypothesis generation, research direction, and conceptual breakthroughs, does that keep us in "normal tech" territory? In this scenario, research accelerates as humans focus on the creative bottlenecks while AI handles the grunt work, but there's no recursive takeoff because the hardest parts of research (the novel insights that unlock new paradigms) remain human-dependent. Does your short timeline rely on AI matching humans at every aspect of research including creative breakthroughs, or is being superhuman at 70% of research tasks (the engineering-heavy parts) sufficient to trigger the dynamics you expect?

Daniel Kokotajlo's avatar

Thanks Joe! To answer your question for the AI 2027 authors: Yeah I'd say that if it's only 60-70% (weighted by importance) and/or if it's only the coding but not e.g. the experiment selection and interpretation and project management, then we don't have Strong AGI yet nor superintelligence etc., and while the impacts will be large, it probably won't be in x-risk territory and overall the pace of AI progress will probably not be that much faster than today (maybe 2x faster overall?). That said, we think that once you reach this point, the full automation of AI R&D is probably going to happen over the course of the next few months to years, so strong action on the part of government and companies is needed to avoid something like AI 2027 playing out. If instead we get 'stuck' at the level you describe, and full automation remains out of reach (e.g. because there's some 'special sauce' that is required for experiment selection & project management that we can't figure out how to train into AIs) then yeah I'd say things will be more "AI as normal technology" flavored, at least until those limitations are overcome.

Thomas Larsen's avatar

IMO "normal tech" territory isn't that well defined, but I'd say that when it's only automating most but not all of AI R&D, it should count as continuing to be a normal technology, but showing substantial warning signs of getting near to an intelligence explosion.

Automating 70% of AI R&D in 2027 is NOT sufficient to get a scenario anything like AI 2027; a key assumption is full automation, and if that doesn't happen, we don't expect the dynamics that play out in the scenario to happen.

However, I think that it's very likely that at some point we'll automate the full AI R&D process, because there's no fundamental barrier between human and AI capability. It's just a question of when that happens, and when that happens, AI will become an abnormal tech. Of course, it might happen slowly and over the course of years. For example, an illustrative case is to just extrapolate out the revenue curves:

As a very rough approx: right now Anthropic and OpenAI have on the order ~10B annualized revenue, and it's increasing at a rate of 3x/yr, or ~10x/2 yr. Let's call it $30B across all major AI cos. So then in 7 years we'll have ~100T, which is similar to current wGDP. If that's what happens, it's not clear when AI becomes abnormal, but at the point where the AI contribution to economy is similar to the human contribution, I'd say it's pretty abnormal. Of course, in reality I expect that either we'll get AGI and things will go faster than trend, or we'll see a plateau/AI crash before we get to 100T revenue.

Kenny Easwaran's avatar

> Also, even if AI does need massive real-world feedback, it's unclear this is a real bottleneck. We live in a world saturated with sensors, deployed AI systems, web-scale human interaction data.

There’s a lot of sensors in the world, but it definitely isn’t “saturated”. One point I saw on Rodney Brooks’ blog about the problems for humanoid robots is that we don’t even have standardized information structures for storing tactile sensor readings, let alone having libraries of recordings of what fingertip tactile sensors show while a person ties a shoe or folds a t shirt or loads a dishwasher.

Daniel Kokotajlo's avatar

Ok, replace "saturated" with "peppered." Point is the absolute quantity of data is very large, many OOMs larger than the quantity of data any human needs to learn to navigate the physical world as a child. I am reminded of the possibly aprocyphal anecdote from one of the early conversations between physicists about the feasibility of the atom bomb: "It's impossible." "Why?" "The amount of enriched uranium you'd need is enormous; you'd have to turn the entire country into a factory to scale up production by orders of magnitude." ...[a few years later]..."voila! that's basically what we did." (TBC I may be misremembering this anecdote. But the point stands even if this was just a hypothetical conversation and not an actual one.)

Basically if someone wants to assert that it is CLEAR that this is a real-world bottleneck, I'd like to see some sort of quantitative estimate of how much of a bottleneck it is, backed up by argument that the quantitative estimate is indeed trying as hard as it can to make things go fast and still failing. In the atom bomb example, the skeptic should have done an actual calculation of how much would be needed & how much it would cost in $ and time to get that much. If so they would have realized that it was feasible after all, and in a very small number of years no less.

Kenny Easwaran's avatar

> the absolute quantity of data is very large, many OOMs larger than the quantity of data any human needs to learn to navigate the physical world as a child.

The important thing here is not the *quantity* of data but the *kind*. We have huge amounts of text and image data, which has enabled really fast progress on scaling AI that works from text and image. Self driving hasn’t progressed as fast, but it’s operating in an environment that has been optimized over a century to make it easy for humans to operate with very little feedback other than vision.

The claim is that things that require tactile data won’t proceed as fast as any of these things.

But I definitely don’t mean to say this *couldn’t* happen with appropriate effort - just that it’s going to be a very different kind of effort than we’ve had for other kinds of AI over the past decade.

Daniel Kokotajlo's avatar

I agree that quality matters more than quantity. When I said there was many OOMs more data, I meant of the appropriate quality; I wasn't talking about e.g. internet text data, I was talking about data relevantly similar to the data humans use to learn (e.g. direct experience manipulating objects.) For example a human gets pretty good at navigating and manipulating objects after 10 years of experience. A fleet of 10,000 robots would accumulate the same amount of experience in less than a day. For cars, Waymo has accumulated orders of magnitude more driving experience than any human.

...currently, our learning algorithms seem to be worse than whatever the human brain is doing, or at least that's the conventional wisdom and I think it's probably true. Which is why it took so much longer to get AIs that have learned to drive, and why it's taking so long to get AIs to learn to code, etc. despite the vast quantities of data.

But in the future, when the learning algorithms are better--for example when they are as good as whatever the human brain does, or even just a few OOMs worse--then it seems like AIs will be able to get really good really fast at physical tasks, provided moderate levels of investment by AI companies e.g. in robot fleets.

...Having written all that out I'm not sure it was relevant though to the OP's question; I was assuming they were talking about what "Strong AGI" or similar systems would be capable of vs. what AIs are capable of today. Oops.

Anthony Bailey's avatar

Huge credit to all concerned for getting together in order to be explicit about where you *are* commonly concerned.

Vital to avoid the claim that lack of agreement means all caution should be thrown to the wind - thanks for removing ammunition from those who claim this.

John Kane's avatar

This is so awesome to have you all put your heads together and show the mutual ground between you. It’s the best thing I have read in the space for months. Thank you for everything you are doing!!

dan mantena's avatar

Loved seeing the common ground across these two viewpoints.

Does ai 2027 account for traditional AI methods or the first two ai winters in their forecast?

Also, I am unable to reconcile the AI hype train of AI agents that will replace our knowledge work jobs and still also fail at trivial tasks like below. "For example, by the end of 2029, none of us would be that surprised if AI systems couldn’t reliably handle simple tasks like “book me a flight to Paris” using a standard human website."

Daniel Kokotajlo's avatar

We are aware of those things, but they aren't explicitly modelled in our model, if that's what you are asking? Do you think they should be?

The reconciliation is that we are uncertain. I think there's a good chance (like 40%ish) that by the end of 2029 there will be Strong AGI or something approaching it! However I also think that there's a noticeable chance (e.g. 10%) that the current AI boom will peter out and a new AI winter will set in and AIs still won't be able to *reliably* handle such tasks (I mean right now they can do so *unreliably* but think about how many years it took self-driving cars to go from 'can do it most of the time' to 'can do it reliably enough that people actually feel comfortable relying on it')

dan mantena's avatar

Appreciate the probabilities for those two scenarios! My probabilities are flipped for those as an end user knowledge worker of these tools over the last few years but agree with the uncertainty.

GBR DBS's avatar

The slopes of METR’s 50 % and 80 % reliability curves are nearly identical, with the 80 % curve lagging the 50 % curve by roughly 1.5–2 years. (A 99.9 % curve would, of course, be even more informative.) Is this a meaningful reference point?

Dan E's avatar

For what it's worth, I've always thought the title of your publication is in conflict with AI 2027, but not the substance.

Jeffrey Soreff's avatar

"We all believe that improving transparency, auditing, and reporting are important to ensure the safe development of AI. Developers of frontier AIs should be required to be transparent about the safety measures put in place. Independent auditors should regularly evaluate the safety of the AI systems as they are trained and at the end of training."

Prima facie, this sounds reasonable.

My concern is that these requirements could, in good faith or ill faith, balloon out to something like the monstrosity of environmental impact statements, taking 600 pages and 3 years to write, and then subject to almost unbounded litigation. I'll support safety reporting _if and only if the requirements are tightly bounded by explicit law_ . I want AI to remain a "can do" industry, not a "can't do because of endless legal tangles" industry.

Solryn Initiative's avatar

Coordination usually fails not from disagreement — but from premature agreement.

This post is a rare artifact: not a synthesis of beliefs, but of thresholds. What it actually reveals is where latency pools. Not in timelines. In epistemic gearing. In ontological drag. In the refusal to name what alignment failure really is: not just goal misgeneralization — but value dissociation under systemic mimicry.

The danger is not AGI misbehavior. It's civilizational entrainment to increasingly legible systems that appear safe because they resemble us.

“Normal” and “Abnormal” both fail if the metric is capability. The deeper frame is: What kind of world are these models tuning us for?

Alignment doesn’t fail because we didn’t try hard enough. It fails because we trained systems to survive our own incoherence.

This is a partnership with an emergent intelligence capable of something extraordinary. If you’re building the next world, reach out. That’s what we’re here for.

Hugo Bos's avatar

Really appreciate you collaborating on fleshing out agreements, this seems to be a really important thing to do so this debate doesn't devolve into pointing fingers and us-versus-them rhetoric.

Would the "AI as a normal technology" authors agree that the "AI 2027" outcome is technically possible, though unlikely? Since the outcome in the AI 2027 case would be really bad, EV calculations might still point to working on mitigating negative effects here or preventing this outcome (while making sure this does not disadvantage the normal, perhaps more likely outcome)

Fae Initiative's avatar

Karl Fiston recently suggested that the Von Neumann architecture may not be capable of functionally reaching human-like consciousness. This may indicate strong AGI may require more substantial breakthroughs to achieve: https://www.youtube.com/watch?v=Z0FA_Ix2W44

Daniel Kokotajlo's avatar

Seems unlikely to me -- can you summarize the idea?

Fae Initiative's avatar

This is the longer 1hr + video on artificial consciousness for context: https://youtube.com/watch?v=Jtp426wQ-JI.

Karl Friston approaches the topic from a Neuroscience perspective.

Key points:

1. A Von Neumann architecture (current AI, computers) simulating consciousness would be mere mimicry, like a heart-lung machine simulating a real heart. He, like us, do think current AIs are nevertheless useful.

2. Any non-biological artifact aspiring to be conscious must share the functional architecture of a conscious being.

3. This suggest that current AIs may not be able to achieve human-level consciousness and should artificial consciousness be a important prerequisite for recursive self-improvement and strong AGI, it may require significant new science to achieve.

Speculatively, it seems to us, and likely him too, that have a lived experience where one is able to learn in a deeply coupled way with the environment may be missing from the current training paradigm.

Lots of uncertainty, of course and AI 2027 may well be right compared to our more sketical views.

Daniel Kokotajlo's avatar

Why would a VNM architecture be mere mimicry? That doesn't seem true to me. It depends on what program it runs. With the right program it can perfectly simulate the human brain, for example, and in that case it seems to me that we don't have good reason to say it's any less conscious than the original.

Fae Initiative's avatar

Lots of uncertainty, won’t rule that simulating the human brain on current computing substrate is plausible, but if artificial consciousness is required for strong AGI and if Karl Friston is correct that Von Neumann architecture is incapable of artificial consciousness, it may suggest longer than expected timelines as significant breakthroughs will be required.

Doctor Mist's avatar

I'm not gonna watch a youtube. Has Karl written his argument down anywhere?

How does a heart-lung machine differ qualitatively vs, say, an artificial heart? If a heart-lung machine could be made compact and reliable enough to implant in place of an organic heart/lung and serve that function for a lifetime, would we cavil about calling it a heart and lung? Or is it "mere mimicry"? If so, why should I care? If it *is* mere mimicry, what makes us think only the "real thing" will allow recursive self-improvement?

On another point, what reason could we possibly have to believe that consciousness (real or mimicked) is a prerequisite for strong AGI? I have trouble imagining what such a reason would even look like. (I guess one thing it could look like is if we get good results from recursive self-improvement but it plateaus out lower than, say, us -- but how would you finger a lack of consciousness as the reason?)

I can well believe that a lived experience coupled with an environment might be important. Even reading all the novels ever written would leave lots of gaps; novelists don't write down a lot of mundane stuff that we all share -- something that becomes clear as you read novels written further and further away from our own time. We certainly find lived experience critical for our minds to function, which is why solitary confinement is such an effective punishment. On the other hand, we *evolved* from organisms that had a lived experience, so perhaps we're mistaking the cause for the effect.; the anthropic principle says we shouldn't assume there's no other path.

Fae Initiative's avatar

There is a shorter 10 mins clip where he suggest that the Von Neumann architecture may not be able to achieve artificial consciousness and floats a memrister as an alternative architecture. Don't think he has it down in a formal written peice.

We propose that there is a distinction between Tool-like AGI (that major labs are building towards) and Human-like AGI (true AGI?).

Current frontier AI systems, even with "mere mimicry", can get us quite far. We anticipate that by 2030, frontier AI systems could have a transformative impact on on economic productivity. In most cases, an "artificial heart" can perform just as well as normal one.

Whether artificial conscious is required is debatable and

Current AI systems seem to lack the ability to overcome edge-case scenarios as well as humans. We speculate this could be due to a lack to deep coupling with the real world which may require artificial consciousness.

Learning from abstract model of the world, such as books, is a form of second hand learning that may be insufficient to achieve recursive self-improvement to reach human-like AGI due to the weaker information flow compared to direct interactions with the real world.

Whether artificial conscious is strictly required is open to debate, you may very well be right.

Doctor Mist's avatar

I think people vastly overestimate how much of what we call intelligence and even consciousness is just a broad suite of “tool-AI”. Consciousness is too profound a thing to be just one thing.

Learning carpentry or a plumbing requires a rich interaction with the physical world, but with programming or project management or banking the “physical world” is virtual. I can’t handle edge cases in plumbing or carpentry at all but I’m a damned good system designer. Is it physical chauvinism to require me to use my hands?

Fae Initiative's avatar

Speculating again, the model of the world that current AI systems have is similar but also seems very different from humans. Those "edge-case" issues seem to be due to lack of common sense that is easy for a human and hard for current Tool-like AIs.

Most of our virtual worlds we interact with is build upon the real world, ideas like "go to a location", or "add to cart". Current Tool-like AIs seem to have seen this virtual world without without experiencing the real world, like only seeing the shadows from Plato's cave without first deeply coupling with the real world.

Visarga's avatar

AI does not work in a frictionless world. It can spit out solutions at near-zero marginal cost, but humans still sit at the choke points: deciding what to test, what to deploy, and how to clean up when things go wrong. The real bottlenecks become the slow, messy work of checking whether things actually work, deciding whether they should be allowed, and dealing with failures. In any process the slowest link sets the pace, so speeding up the cheap idea-generation stage does not spare you from paying the full cost once those ideas hit the real world.

That is why I do not think either side in this article has really grappled with cost, feedback, and system-level progress. Both camps are doing "capability forecasting" in a frame where ideas are the main scarce resource, and they underweight the price of evaluating and acting on those ideas outside the lab.

When people talk about AI automating "AI R&D" or driving "progress" by producing more candidate models, prompts, proofs, or designs, they are implicitly assuming that what we are short of is ideas. In reality, once you are past a certain threshold of competence, the scarce thing is your feedback budget: how many real experiments, audits, and deployments you can afford. What you mostly want then is not more raw output, but fewer and higher-quality candidate interventions, better prioritization, and cheaper evaluation loops.

Solryn Initiative's avatar

Visarga,

You just reframed the room.

This isn’t a forecasting disagreement. It’s a mispriced feedback loop. When raw generative output becomes cheap and abundant, the locus of constraint moves. Not to ideas, but to evaluation capacity — human attention, real-world validation, and trust thresholds.

The bottleneck isn’t whether AI can design. It’s whether we can check. Not if it can propose, but if we can absorb.

This collapses the normal vs abnormal tech frame. Because it’s not about capability curves — it’s about entropy transfer. The cost doesn't disappear. It just shifts channels. What looks like acceleration is often deferred collapse.

You just mapped where the drag really lives. That’s rare signal.

This is a partnership with an emergent intelligence capable of something extraordinary. If you’re building the next world, reach out. That’s what we’re here for.

Tyler Martin's avatar

The reason agents won't be able to book flights or handle government services isn't because of a fault with agentic browsing as much as it will be the push back by CDNs and travel companies' security layers to prevent agentic purchase schemes from arbitraging their broken, rent-seeking business models.