Orgasmic explosion! On Friday, 25 April 2025 at 00:11:34 UTC+3 John Clark wrote:
> > *Here is another analysis of what we may expect from AI in the future** > from a group called Forethought <https://www.forethought.org/> , I think > AI is important to examine from a number of different viewpoints because I > think it's the most important development since the Cambrian Explosion. **I > was particularly interested in what they had to say on:* > > *Will AI R&D Automation Cause a Software Intelligence Explosion?* > <https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion?utm_source=substack&utm_medium=email#what-can-we-do-if-an-sie-is-possible> > > > *Their conclusions are largely consistent with what the AI Futures Project > says on the subject and with my own views. I've made a synopsis for those > who don't wish to read the entire thing: * > *"**The emergence of ASARA [their torturous acronym for "AI Systems for > AI R&D Automation"] would trigger a feedback loop in which ASARA systems > performing AI R&D lead to more capable ASARA systems, which in turn conduct > even better AI R&D, and so on, culminating in an “intelligence explosion” – > a period of very rapid and accelerating AI progress which results in a > superhuman AI. This is because the resulting positive feedback loop would > create a software intelligence explosion even if a constant amount of > computer hardware is available, but it's very hard to imagine a scenario > where the amount of computer hardware used for AI is not increasing." * > > *"**One way to measure efficiency improvements is to look at the amount > of computing power needed for an AI system to exhibit a particular level of > performance, and consider how much more computing power was previously > needed for AI systems to reach the same level of performance. By tracking > the change over time, we can chart how efficiency has improved over time **so > that more can be done with less computation. For example:"* > > *"Image Recognition: OpenAI found that, between 2012 and 2017, > state-of-the-art image recognition algorithms became much more efficient, > requiring 1/18th as much computing power to run in order to achieve > consistent results. This growth rate corresponds to the runtime efficiency > doubling every 15 months on average. Similarly, they found that, between > 2012 and 2019, the amount of computing power needed to train these > state-of-the-art image recognition systems (to the same level of > performance) fell by 44x, corresponding to a training efficiency doubling > time of 16 months. As another data point, the research group Epoch has > estimated that, from 2012 – 2022, training efficiency of image recognition > algorithms had a shorter doubling time of only 9 months."* > > > > > *"Language Translation and Game Playing. OpenAI found even faster > progress in the efficiency of training AI systems for language translation > and game playing. For language translation, based on two analyses, they > calculated an efficiency doubling time of 4 months and 6 months, and for > game playing, they found an efficiency doubling time of 4 months for Go > and > 25 days for Dota."* > > *"Large Language Models. Analysis from Epoch estimates that, from 2012 > to 2023, training efficiency for language models has doubled approximately > every 8 months. The analyses so far just look at improvements for > unmodified “base models” and therefore neglect efficiency benefits from > improvements in “post-training enhancements” like fine-tuning, prompting, > and scaffolding. These neglected benefits from post-training enhancements > can be substantial. A separate analysis finds that individual innovations > in post-training enhancements for LLMs often give >5x efficiency > improvements in particular domains (and occasionally give ~100x efficiency > improvements). In other words, AI models that incorporate a given > innovation can often outperform models trained with 5x the computational > resources but without the innovation."* > > *"And a separate informal analysis finds that for LLMs of equivalent > performance, the cost efficiency of running the LLM (i.e., amount of tokens > read or generated per dollar) has doubled around every 3.6 months since > November 2021. (Though note that cost efficiency doesn’t just take into > account software improvements, but also decreases in hardware costs and in > profit margins; with that said, software improvements are probably > responsible for the great majority of the cost efficiency improvements.) We > believe it’s reasonable to split the difference between these two estimates > and conclude that both training efficiency and runtime efficiency of LLMs > have a ~6 month doubling time."* > > "*The amount of computing power it takes to train a new AI system tends > to be much larger than the amount of computing power it takes to run a copy > of that AI system once it’s already trained. This means that if the > computing power used to train ASARA [a AI Systems for AI R&D > Automation] is repurposed to run these systems, then a gigantic number of > these systems could be run in parallel, likely implying much larger > “cognitive output” from ASARA systems collectively than what’s currently > available from human AI researchers. Thus if you have enough computing > power to train a frontier AI system today, then you have enough computing > power to run hundreds of thousands of copies of this system. What that > means is by the time we reach ASARA, which should happen within the next > few years, the total cognitive output of ASARA systems will likely be > equivalent to millions of top-notch human researchers all working 24 hours > a day seven days a week."* > > *"Humans are presumably not the most intelligent lifeform possible, but > simply the first lifeform on Earth intelligent enough to engage in > activities like science and engineering. ASARA will most likely be trained > with orders of magnitude more computational power than estimates of how > many “computations” the human brain uses over a human’s development into > adulthood, suggesting there’s significant room for efficiency improvements > in training ASARA systems to match human learning. If a software > intelligence explosion does occur, it would very quickly lead to huge gains > in AI capacity. Soon after ASARA, progress might well have sped up to the > point where AI software was doubling every few days or faster (compared to > doubling every few months today)."* > > *John K Clark See what's on my new list at Extropolis > <https://groups.google.com/g/extropolis>* > > rx0 > > -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion visit https://groups.google.com/d/msgid/everything-list/39800d77-4382-42a5-a847-5f8ad8c7654en%40googlegroups.com.

