a_fancy_kiwi

joined 2 years ago
[–] a_fancy_kiwi@lemmy.world 4 points 1 day ago* (last edited 1 day ago)

That’s unacceptable and infuriating. I started going to my wife’s appointments because of shit like this. I don’t know if it’s because my wife wasn’t advocating enough for herself or if the doctor was shitty but all I do is sit there and repeat my wife’s concerns in different words when I think I need to. If I’m not satisfied with an answer, I just keep asking questions. She’s getting the care she deserves now.

This is coming from a good place and I don’t mean any disrespect but it sounds like your wife needs a mammogram. Do something about it; you only have each other.

[–] a_fancy_kiwi@lemmy.world 31 points 6 days ago (3 children)

After COVID, I forget common words in the middle of speaking now which has been fucking fantastic in work environments.

“Hey can you hand me that…um…uh….that piece of paper for coffee grounds?”

“A filter?”

“Yeah that”

[–] a_fancy_kiwi@lemmy.world 1 points 1 week ago

No, you actually caught be at the perfect time, the transfer to my temporary pool is almost done. I was just curious how inheritance worked on a pool but after giving it some thought, your recommendation makes more sense; turn it on when I know I need it vs turn it off when I know I don’t. Thanks for the advice.

[–] a_fancy_kiwi@lemmy.world 1 points 1 week ago (2 children)

Let’s say I did turn on compression on root. I can’t then turn it off on a per file system basis where it isn’t needed?

[–] a_fancy_kiwi@lemmy.world 2 points 1 week ago (1 children)

Thank you for this

30
Raid Z2 help (lemmy.world)
submitted 1 week ago* (last edited 1 week ago) by a_fancy_kiwi@lemmy.world to c/selfhosted@lemmy.world
 

tldr: I'm going to set up raid z2 with 4x8TB hard drives. I'll have photos, documents (text, pdf, etc.), movies/tv shows, and music on the pool. Are the below commands good enough? Anything extra you think I should add?

sudo zpool create mypool raidz2 -o ashift=12 /dev/disk/by-id/12345 ...

zfs set compression=lz4 mypool #maybe zstd?
zpool set autoexpand=on mypool
zpool set autoreplace=on mypool #I might keep this off. I can see myself forgetting in the future
zpool set listsnapshots=on mypool

With ai raising hard drive prices, I over spent on 3x10TB drives in order to reorganize my current pool and have 3 hard drives sitting on a shelf in the event of a failure. My current pool was built over time but it currently consists of 4x8TB drives. They are a mirrored stripe so a usable 16TB. If I understand it correctly, I can lose 1 drive for sure without losing data and maybe a second drive depending on which drive fails. Because of that, I want to move to raid z2 to ensure I can lose 2 drives without data loss. I'm going to move data from my 4x8TB drives, to the 3x10TB, reconfigure the 4x8TB, and move everything back. I run Immich, plex/jellyfin, and navidrome off the pool. All other documents are basically there for long term storage just in case. What options should I use for raid z2 when setting it up?

I know I can look this stuff up. I have been and continue to do so, I was just hoping for some advise from people that are more knowledgeable about this than me. The move from the 4x8TB drives to the 3x10TB is going to take ~3 days so I really don't want to mess this up and have to start over 😅

Edit:

After looking up each property, this is the command I will probably end up using to create the raid z2 pool, thanks Avid Amoeba:

~~sudo zpool create
-o ashift=12 -o acltype=posixacl -o xattr=sa
-o compression=lz4 -o dnodesize=auto -o relatime=on
-o normalization=formD
raidz2
mypool
/dev/disk/by-id/12345 ...~~

Edit2:

Above command didn't work on my machine. The order and uppercase "O" matters. Had to do this:

sudo zpool create \
  mypool \
  raidz2 \
  -o ashift=12 -O compression=lz4 \
  -O normalization=formD -O acltype=posixacl \
  -O xattr=sa -O dnodesize=auto \
  -O relatime=on \
  /dev/disk/by-id/12345 ...

Edit3:

And finally, after all this, I set up my tmp pool of 3x10TB disks as a raid z2 instead of raid z1. Spent a day and a half transferring before I finally saw my mistake after running out of space 🫠

[–] a_fancy_kiwi@lemmy.world 5 points 3 weeks ago (2 children)

I’ll look into it, thanks.

I’m still in the information gathering phase. Do you know if the element client works with the continuwuity server? Is it as easy as entering the domain, user, and password in the client?

[–] a_fancy_kiwi@lemmy.world 5 points 3 weeks ago

Fair criticism. I just don’t have a lot of free time. I can invest in Element but I wanted to crowd source information to see if it was worth it or if there was an easier way. It doesn’t get much easier than Docker

[–] a_fancy_kiwi@lemmy.world 2 points 3 weeks ago (4 children)

Out of curiosity, what makes it better?

A quick search says it’s a package manger for kubernetes. Besides plex, everything I selfhost is just for me. Would you say helm/kubernetes is worth looking into for a hobbyist who doesn’t work in the tech field?

 

My friends are open to leaving Discord which has finally given me a reason to look into Element/Matrix. I found the install instructions and am immediately put off. Is this it? No official docker compose? 😞

[–] a_fancy_kiwi@lemmy.world 2 points 1 month ago

I’d love to but it’s a chicken and egg thing. Regular people don’t understand bitcoin let alone monero. On top of that, you still have fees for converting from a currency to monero and again from monero to a currency, so there’s still a middle man :/

[–] a_fancy_kiwi@lemmy.world 1 points 1 month ago

thanks, I'll look into it. Much appreciated

[–] a_fancy_kiwi@lemmy.world 2 points 1 month ago* (last edited 1 month ago)

I understand your view and sympathize deeply but there is a lot wrong in the world today and if I have to divert energy somewhere to try and change something, Patreon's 10/90 split is at the bottom of the list for me at the moment. Regular working class people aren't getting that good of a deal at the jobs they work at. I'm not even talking 90/10, just getting their .01% of the profits in a 1000 person org for their contribution would probably be life changing.

You are right but our frames of reference are different.

[–] a_fancy_kiwi@lemmy.world 8 points 1 month ago (4 children)

I’ve never looked into adding GitHub releases to FreshRSS. Any tips for getting that set up? Is it pretty straight forward?

 

I recently noticed that htop displays a much lower 'memory in use' number than free -h, top, or fastfetch on my Ubuntu 25.04 server.

I am using ZFS on this server and I've read that ZFS will use a lot of RAM. I also read a forum where someone commented that htop doesn't show caching used by the kernel but I'm not sure how to confirm ZFS is what's causing the discrepancy.

I'm also running a bunch of docker containers and am concerned about stability since I don't know what number I should be looking at. I either have a usable ~22GB of available memory left, ~4GB, or ~1GB depending on what tool I'm using. Is htop the better metric to use when my concern is available memory for new docker containers or are the other tools better?

Server Memory Usage:

  • htop = 8.35G / 30.6G
  • free -h =
               total        used        free      shared  buff/cache   available
Mem:            30Gi        26Gi       1.3Gi       730Mi       4.2Gi       4.0Gi
  • top = MiB Mem : 31317.8 total, 1241.8 free, 27297.2 used, 4355.9 buff/cache
  • fastfetch = 26.54GiB / 30.6GiB

EDIT:

Answer

My Results

tldr: all the tools are showing correct numbers. Htop seems to be ignoring ZFS cache. For the purposes of ensuring there is enough RAM for more docker containers in the future, htop seems to be the tool that shows the most useful number with my setup.

 

This is a continuation of my other post

I now have homeassistant, immich, and authentik docker containers exposed to the open internet. Homeassistant has built in 2FA and authentik is being used as the authentication for immich which supports 2FA. I went ahead and blocked connections from every country except for my own via cloudlfare (I'm aware this does almost nothing but I feel better about it).

At the moment, if my machine became compromised, I wouldn't know. How do I monitor these docker containers? What's a good way to block IPs based on failed login attempts? Is there a tool that could alert me if my machine was compromised? Any recommendations?

EDIT: Oh, and if you have any recommendations for settings I should change in the cloudflare dashboard, that would be great too; there's a ton of options in there and a lot of them are defaulted to "off"

 

tldr: I'd like to set up a reverse proxy with a domain and an SSL cert so my partner and I can access a few selfhosted services on the internet but I'm not sure what the best/safest way to do it is. Asking my partner to use tailscale or wireguard is asking too much unfortunately. I was curious to know what you all recommend.

I have some services running on my LAN that I currently access via tailscale. Some of these services would see some benefit from being accessible on the internet (ex. Immich sharing via a link, switching over from Plex to Jellyfin without requiring my family to learn how to use a VPN, homeassistant voice stuff, etc.) but I'm kind of unsure what the best approach is. Hosting services on the internet has risk and I'd like to reduce that risk as much as possible.

  1. I know a reverse proxy would be beneficial here so I can put all the services on one box and access them via subdomains but where should I host that proxy? On my LAN using a dynamic DNS service? In the cloud? If in the cloud, should I avoid a plan where you share cpu resources with other users and get a dedicated box?

  2. Should I purchase a memorable domain or a domain with a random string of characters so no one could reasonably guess it? Does it matter?

  3. What's the best way to geo-restrict access? Fail2ban? Realistically, the only people that I might give access to live within a couple hundred miles of me.

  4. Any other tips or info you care to share would be greatly appreciated.

  5. Feel free to talk me out of it as well.

EDIT:

If anyone comes across this and is interested, this is what I ended up going with. It took an evening to set all this up and was surprisingly easy.

  • domain from namecheap
  • cloudflare to handle DNS
  • Nginx Proxy Manager for reverse proxy (seemed easier than Traefik and I didn't get around to looking at Caddy)
  • Cloudflare-ddns docker container to update my A records in cloudflare
  • authentik for 2 factor authentication on my immich server
 

I've been interested in building a DIY NAS out of an SBC for a while now. Not as my main NAS but as a backup I can store offsite at a friend or relative's house. I know any old x86 box will probably do better, this project is just for the fun of it.

The Orange Pi 5 looks pretty decent with its RK3588 chip and M.2 PCIe 3.0 x4 connector. I've seen some adapters that can turn that M.2 slot into a few SATA ports or even a full x16 slot which might let me use an HBA.

Anyway, my question is, assuming the CPU isn't a bottle neck, how do I figure out what kind of throughput this setup could theoretically give me?

After a few google searches:

  • PCIe Gen 3 x4 should give me 4 GB/s throughput
  • that M.2 to SATA adapter claims 6 ~~GB/s~~ Gb/s throughput
  • a single 7200rpm hard drive should give about 80-160MB/s throughput

My guess is that ultimately, I'm limited by that 4GB/s throughput on the PCIe Gen 3 x4 slot but since I'm using hard drives, I'd never get close to saturating that bandwidth. Even if I was using 4 hard drives in a RAID 0 config (which I wouldn't do), I still wouldn't come close. Am I understanding that correctly; is it really that simple?

 

PSA

After updating to TvOS 17, my Sonos Beam sound bar started making weird crackling sounds and music sounded tinny. Turns out, I had to change the audio format in the Apple TV settings from Stereo to Dolby Digital 5.1 for the issue to be fixed.

Not sure what I had that setting set to before but I’m leaning toward the idea that the update reset the audio format back to default settings. If you are having sound issues after updating, that might be the issue.

 

I occasionally find myself reinstalling home assistant and every time I do, I get stuck on two steps because I forgot the commands and didn't write them down from the last time. I'm writing them below mainly for myself but also for anyone else who may get stuck. For future reference, I'm using Ubuntu 23.04 with Virt-Manager.

Before you begin the installation of the provided qcow2 image, you might want to resize that image from 32G to whatever size you want. ex:

qemu-img resize haos_ova-10.3.qcow2 +68G

Next, you might want to make a network bridge device. Navigate to your netplan folder and backup the yaml file that's in there (your file may be named differently)

cd /etc/netplan

cp ./01-network-manager-all.yaml ./01-network-manager-all.yaml.old

Edit the yaml config.

nano ./01-network-manager-all.yaml

Change the renderer to networkd and add the bridge device (br0). Your ethernet device may not be named enp12s0, make sure to use your ethernet device name. If you are on wifi, look up a netplan wifi config and make adjustments as needed.

network:
  renderer: networkd
  ethernets:
    enp12s0:
      dhcp4: true
  version: 2
  bridges:
    br0:
      dhcp4: yes
      interfaces:
        - enp12s0
      parameters:
        stp: true

save the file. generate and apply the new netplan. WARNING - If you are hosting this on your own network, it's possible the Ubuntu host IP could change. If you were doing these steps over SSH, you might need to find the new IP and reconnect. Static IPs can be set in the netplan config but I usually just do it from my router settings afterwards which is probably why the IP changed.

netplan generate

netplan apply

Now just go through the installation process and when you select your network device, make sure you select "Bridge Device" and the device name is "br0"

Edit 12/15/23 - well, I rebuilt my server again. I used regular Ubuntu desktop this time and I for the life of me I couldn’t get networking to function properly. I ended up buying an Ethernet card and passed it through to the VM

view more: next ›