Regular desktop PC with a €2500 euro graphics card.
Machine Learning
A community for posting things related to machine learning
Icon base by Lorc under CC BY 3.0 with modifications to add a gradient
A 120b parameter model is small compared to the models running in datacenters. However, this does seem like the current "Moore's Law" for AI. Finding more and more efficient ways to run larger parameter models.
Do we really need datacenters to host (load balanced) web servers and services?
This could challenge the business model of Big Tech!
The article isn't very concrete on what the substance of this Anyway System is.
The FAQ answered my question though, and it seems mostly about dynamic management of PCs as execution nodes.
I assume that means I could run the model on one of those PCs as well, despite the article claiming you can use as few as four PCs? Or does this system somehow distribute a single model prompt process into multiple execution pipelines?