I mean, if you draw the scaling curves out and believe them, then sometime in the next 3-10 years, plausibly shorter, AIs will be able to achieve best-case human performance in everything able to be done with a computer and do it at 10-1000x less cost than a human, and shortly thereafter robots will be able to do something similar (though with a smaller delta in cost) for physical labor, and then shortly after that we get atomically precise manufacturing and post-scarcity. So the amount of stuff that amounts to nothing is plausibly every field of endeavor that isn't slightly advancing or delaying AI progress itself.
It is kinda a meme at this point, that there is no more "publicly available"... cough... training data. And while there have been massive breakthroughs in architecture, a lot of the progress of the last couple years has been ever more training for ever larger models.
So, at this point we either need (a) some previously "hidden" super-massive source of training data, or (b) another architectural breakthrough. Without either, this is a game of optimization, and the scaling curves are going to plateau really fast.