This is an automated archive made by the Lemmit Bot.

The original was posted on /r/artificial by /u/tail-recursion on 2023-10-28 13:05:58.


I’m watching this interview with Francois Chollet where he talks about science as an example of a superhuman recursively self improving problem solving system and how we can use it to reason about what a superhuman artificial general intelligence might be like. One thing I find interesting is his claim that the amount of resources we are investing into science is exponentially increasing but we are only making linear progress. If we assume this is true, i.e. that to continue making linear progress in science we need to invest exponentially increasing resources, doesn’t it imply that eventually if we can’t keep investing the exponentially increasing required resources to keep make linear progress that eventually we will start making worse than linear progress? Does this imply that in the very long term scientific progress is likely to slow down significantly?