DiscoverLessWrong (30+ Karma)“How likely is dangerous AI in the short term?” by Nikola Jurkovic
“How likely is dangerous AI in the short term?” by Nikola Jurkovic

“How likely is dangerous AI in the short term?” by Nikola Jurkovic

Update: 2025-11-11
Share

Description

How large of a breakthrough is necessary for dangerous AI?

In order to cause a catastrophe, an AI system would need to be very competent at agentic tasks[1]. The best metric of general agentic capabilities is METR's time horizon. The time horizon measures the length of well-specified software tasks AI systems can do, and is grounded in human baselines, which means AI performance can be closely compared to human performance.

Causing a catastrophe[2] is very difficult. It would likely take many decades, or even centuries, of skilled human labor. Let's use one year of human labor as a lower bound on how difficult it is. This means that AI systems will need to at least have a time horizon of one work-year (2000 hours) in order to cause a catastrophe.

Current AIs have a time horizon of 2 hours, which means it's 1000x lower than the time horizon necessary to cause a catastrophe. This presents a pretty large buffer.

Currently, the time horizon is doubling roughly every half-year. That means that a 1000x increase would take roughly 5 years at the current rate of progress. So, in order for AI to reach a time horizon of 1 work-year within [...]

---

Outline:

(00:11 ) How large of a breakthrough is necessary for dangerous AI?

(02:04 ) AI breakthroughs of the recent past

(02:27 ) Case 1: Transformers

(03:54 ) Case 2: AlphaFold

(04:30 ) What is the probability of 1-year time horizons in the next 6 months?

(05:10 ) Narrowly superhuman AI leading to generally competent AI

(06:27 ) Would we notice a massive capabilities increase?

(07:44 ) Conclusion

The original text contained 3 footnotes which were omitted from this narration.

---


First published:

November 11th, 2025



Source:

https://www.lesswrong.com/posts/B5xQwkmWL5wmFNZkX/how-likely-is-dangerous-ai-in-the-short-term


---


Narrated by TYPE III AUDIO.


---

Images from the article:

Graph showing 50% time horizon increasing from seconds to plausibly catastrophic capabilities at 1000+ hours.
Graph showing

Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

Comments 
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

“How likely is dangerous AI in the short term?” by Nikola Jurkovic

“How likely is dangerous AI in the short term?” by Nikola Jurkovic