Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What about some kind of sharding, parts of the computation that could be executed in isolation for a longer period of time?


An ongoing research problem. OpenAI would certainly like being able to use smaller GPUs, instead of having to fit the entire model into one.


GPT-3 does not fit in any one GPU that exists at present. It's already spread out across multiple GPUs.




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: