this post was submitted on 09 May 2026
129 points (97.1% liked)
Technology
84502 readers
5460 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I use an APK called Off Grid and load ai onto that (right now I'm using genna 4). It's all done on my phone. Nothing on the cloud. No data sent anywhere. Completely local. No entities get shit from it. The only way I'll use ai.
I highly value privacy, but the gap between local LLMs vs top of the line cloud LLMs (e.g. Claude & DeepSeek) is still too great for me to switch completely to the former.
I’ll use PWAs to sandbox LLMs from everything else (and each other) and try to create semantic distance between the user and the queries.
How about that leaked Claude source code? Is there a reliably clean version of that available anywhere yet?
Gemma 4 seems pretty legit so far.
I have my local LLM currently setup and it runs just as well as Sonnet 4.6 from a quality standpoint, and for performance it is slightly slower but it's still faster than I can respond.
This is with a Strix Halo APU with 128GB unified memory using the latest Qwen3.6 models with llama.cpp.
Can you please provide any feedback on this?