bazinga

joined 2 weeks ago
[–] bazinga@discuss.tchncs.de 7 points 17 hours ago* (last edited 17 hours ago)

I agree. I also think that there is nothing good in for-profit AI corporations. I can recommend the book "the empire of AI" However, I personally think self hosting and having full control of the use is a bit different.

[–] bazinga@discuss.tchncs.de 2 points 18 hours ago (1 children)

Thanks, will also ask in the other group you mentioned. I am still having a gaming rig here with rx6900xt as well but way too big to get it wife approved into the living room and have no man cave to run it 24/7. ;) But maybe good for testing what I actually need in model size, I think it is just 1 generation before all the ai hype took off but going to try now right away.

[–] bazinga@discuss.tchncs.de 2 points 18 hours ago (4 children)

Thank you for the detailed writeup. Are you aware of anything small footprint? I am thinking like dgx spark size maybe a bit bigger?

22
submitted 20 hours ago* (last edited 18 hours ago) by bazinga@discuss.tchncs.de to c/selfhosted@lemmy.world
 

I realize, I need to upgrade my little NUC to something bigger for higher inference of bigger llama models. I want something that you still can have on your living room's tv bench, so no monster rack please, but that has also the necessary muscle when needed for llama. Budget doesn't matter right now, want to understand what's good and what's out there. Thanks

EDIT: Wow, thanks for the inspiration, guess I need to look at bit for "how to stuff a huge graphics card into a mini box". To clarify a bit more what I want with it: I want to build a responsive personal assistant. I am dreaming of models bigger than 8B, good tool calling for things like memory, websearch etc., no coding, no image generation, no video generation required. Image recognition would be good but not a must. Regarding footprint, the no monster ;) Something that you can have in your livingroom, and could be wife approved - so no big gaming rig with exhaust pipes and stuff, needs to be good looking ;)

[–] bazinga@discuss.tchncs.de 2 points 21 hours ago

NUC i7, 32gb ram. Full Docker stack includes home assistant and all relevant containers for it zigbee2mqtt, esphome, vscodium, sqliteweb, rtl_433, mqtt, kokoro_tts, weaviate, zwavejs, openwebui, ollama, paperless with ai capabilities, a1111, whisper, sync thing, searxng, redis, qdrant, postgres. Runs fine, however ollama starts to max it out as I want to go for bigger models, so looking for something with serious gpu oomphbut still small footprint and low power consumption

[–] bazinga@discuss.tchncs.de 3 points 4 days ago

Wireguard runs in a different subnet at home, so the ping sensor for my phone fails on the regular WLAN address and this my ha always knows when home and when not.

[–] bazinga@discuss.tchncs.de 7 points 1 week ago

Opnsense any option for you instead of openwrt. I run an old hp with Intel Ethernet card and connected a unify AP to it. Works well since years