this post was submitted on 09 Jan 2026
5 points (77.8% liked)
Machine Learning
596 readers
8 users here now
A community for posting things related to machine learning
Icon base by Lorc under CC BY 3.0 with modifications to add a gradient
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
A 120b parameter model is small compared to the models running in datacenters. However, this does seem like the current "Moore's Law" for AI. Finding more and more efficient ways to run larger parameter models.