this post was submitted on 10 Feb 2025
66 points (100.0% liked)

PC Gaming

10691 readers
414 users here now

For PC gaming news and discussion. PCGamingWiki

Rules:

  1. Be Respectful.
  2. No Spam or Porn.
  3. No Advertising.
  4. No Memes.
  5. No Tech Support.
  6. No questions about buying/building computers.
  7. No game suggestions, friend requests, surveys, or begging.
  8. No Let's Plays, streams, highlight reels/montages, random videos or shorts.
  9. No off-topic posts/comments, within reason.
  10. Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 9 points 1 month ago (1 children)

Most AI models need at least 24 but preferably 32.

Where are you getting this information from? Most models that are less than 16B params will run just fine with less than 24 GB of VRAM. This github discussion thread for open-webui (a frontend for Ollama) has a decent reference for VRAM requirements.

[โ€“] [email protected] -2 points 1 month ago* (last edited 1 month ago)

I should have been more specific. The home models that actually compete with paid ones in both accuracy & speed. Please don't be one of those to exaggerate & pretend it works just as good with much less. It simply doesn't.