You can try to build a monster the size of GPT-4.5, but even if you could actually make the training stable and efficient at this scale, you still would suffer trying to serve it to the users.
Next generation of AI hardware should put them in reach, and I expect that model scale would grow in lockstep with new hardware becoming available.
You can try to build a monster the size of GPT-4.5, but even if you could actually make the training stable and efficient at this scale, you still would suffer trying to serve it to the users.
Next generation of AI hardware should put them in reach, and I expect that model scale would grow in lockstep with new hardware becoming available.