I have been thinking about LLMs just like everyone else, I guess! The prevailing sentiment seems to be they are going to change the world, from jobs to politics and everything in between. However, I have been wondering if we are perhaps reaching “peak” LLM and we will see a dramatic reduction in progress.<p>I am a casual observer of this space so I may be barking up the totally wrong tree, but I thought I would put these ideas out anyway to see if anyone has any thoughts.<p>Lack of quality training data:
Is there a possibility that all the good quality training data has already been used? I would argue that most of the internet doesn’t actually contain good content, and increasing the amount of training data from the internet might actually make these models produce worse output. Moreover, organisations with valuable content could begin to restrict data access, presenting further challenges for training.<p>AI generated content polluting training data:
This is somewhat related to the previous point, but is there a risk that with LLMs generating so much content that the training data becomes polluted with AI generated text. How would this effect the model output? Is it like taking a photocopy of a photocopy over and over again?<p>Compute Resources:
We continually hear that Moore’s Law is dead, are we going to start running into compute/memory issues? Without dramatic increases in both compute/memory are we going to hit scaling issues where it simply takes too long or we don’t have the memory to train larger models?<p>Architecture Limitations:
From my understanding so far adding more and more parameters to a transformer increases its performance. Are we sure that this scales? Or at some point does this performance increase stop or even perhaps go into reverse?<p>Hopefully these points are not totally off the wall!
Moore's Law is not really dead.<p>Even if you froze the datasets and software architecture in place right now, LLMs would get much better simply because compute costs are coming down. The cheaper training is, the more people you have chipping away at little changes and modifications. And you get more hyper specialized models.<p>Also, setting that aside, theres really tons of low hanging fruit to pick.