Selfhosted
A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.
Rules:
-
Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.
-
No spam posting.
-
Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.
-
Don't duplicate the full text of your blog or github here. Just post the link for folks to click.
-
Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).
-
No trolling.
-
No low-effort posts. This is subjective and will largely be determined by the community member reports.
Resources:
- selfh.st Newsletter and index of selfhosted software and apps
- awesome-selfhosted software
- awesome-sysadmin resources
- Self-Hosted Podcast from Jupiter Broadcasting
Any issues on the community? Report it using the report flag.
Questions? DM the mods!
view the rest of the comments
When it comes to Nvidia GPUs the VRAM is the main thing to look for.
For consumer cards it is:
Entry level - RTX 5060 Ti 16GB RAM with a price point around 500-550 euro
Mid - Buying a used RTX 3090 24GB RAM with a price point around 830 euro when I look at swedish second hand markets
High - RTX 5090 32GB RAM with a price point around 3500 euro
After that you end up looking at the RTX Pro Blackwell cards:
Entry - RTX PRO 5000 Blackwell 48GB RAM ~5300 euro
Mid - RTX PRO 6000 Blackwell 96GB RAM ~10100 euro
It all depends on which models you want to run, you can definitely start playing around with Llama 3 8B and similar models with a 5060 Ti 16GB.
If you're looking at 24B-30B models you need the 24GB VRAM that RTX 3090 offers and get a larger context window if you go for the RTX 5090.
If you're looking to run Llama 3 70B then you need to go into the RTX Pro level of vram.
All of this is based on running it with Nvidia cards, there's also other setups such as Mac Studios with huge amount of RAM. They're slower but allow for much larger models at the same price point.
You could also run with AMD/Intel gpus but much software is built primarily for running CUDA (and Nvidia) gpus so it's more work and not always compatible.
I know you said no "monster rack" but I don't really know what you classify as a monster. :)
An ordinary gaming pc is also a good starter AI pc, so something like this allows you to do both:
https://pcpartpicker.com/list/sFp4qd
Thank you for the detailed writeup. Are you aware of anything small footprint? I am thinking like dgx spark size maybe a bit bigger?
Memory bus speed of the Spark is poor and that's a huge detriment.
Problem with smaller footprint is cooling and how audible it becomes.
One idea is to use fiber optic hdmi cables and a usb extender to hide the pc away in another room.
If you want smaller footprint then the keyword to use is "Unified memory", it can be reasonable fast for 30B models and a slow thinker mode for 70B ones.
edit: example of a Unified Memory Apple Mac Studio can be found here at $5499 for 96GB RAM
https://www.apple.com/shop/buy-mac/mac-studio/m3-ultra-chip-32-core-cpu-80-core-gpu-96gb-memory-2tb-storage
If you happy with 16g , nothing can beat in speed/cost of AMD RX 9070 XT.
Wouldn't an AMD RX 9060 XT with 16 GB RAM be nice as well if you're hunting for good speed/cost options?
Probably. It just not as fast as 9070 XT. I'm using 9070 XT myself and limitation for running LLMs is memory, not speed. If model fit in memory it will runs fast enough to be practical.