Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Raw model size is still pegged by the hardware.

You can try to build a monster the size of GPT-4.5, but even if you could actually make the training stable and efficient at this scale, you still would suffer trying to serve it to the users.

Next generation of AI hardware should put them in reach, and I expect that model scale would grow in lockstep with new hardware becoming available.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: