moonpiedumplings

joined 2 years ago

Wow, this looks extremely promising.

https://sylve.io/guides/node/settings/authentication/users/

I would want RBAC, and single sign on tied into that, but they do seem to be aware and working to add it.

It's complicated. There's a lot of context to this, and even the debate in general.

One big problem is that there's a lot of money in this. If you "prove" something is real, and pretend it's a novel discovery, then you can try to sell a novel product that capitalizes off of that.

For example, there used to be a big trend in education, "evidence based learning". https://en.wikipedia.org/wiki/Evidence-based_education . The idea was science would be used to discover the best ways to learn/teach.

The problem was that the method of implementation would be software, or trainings. That you buy...

This reddit thread is a snapshot of the anger and frustration from that: https://www.reddit.com/r/Teachers/comments/jj6tvx/im_done_with_evidencebased_educational_research/

And of course, much of it was debunked later. Like learning styles, for example, were debunked. Although there was some good stuff, like spaced repetition, for which there is a FOSS app called Anki.

Psychology is kinda the same. People do science to try to back products, or trainings, which are then sold.

The inability to replicate these studies is ultimately not a failure, but a success. Science is still doing it's job.

It looks real. I am interested in the way they sandbox wordpress extensions, which have been cause of a lot of vulnerabilities, but I am wondering how they sandbox extensions that want more privileged access, like those that replace the content editing and site rendering features.

Well, more like was interested. From their github.

EmDash depends on Dynamic Workers to run secure sandboxed plugins. Dynamic Workers are currently only available on paid accounts. Upgrade your account (starting at $5/mo) or comment out the worker_loaders block of your wrangler.jsonc configuration file to disable plugins.

Okay, I hath returned. Here is what I am doing with FLuxCD and it's method of installing helm charts:

https://github.com/moonpiedumplings/flux-config/blob/fe71453facc4f5eaf70738430ca7f38f098171b2/apps/apps/authentik/helmrelease.yaml

Okay, I'm cheating. :/ . I'm using Flux's method where you can have a secret that has values, and then I'm just including those.

But yeah, using an ENV var that pulls from a secret is probably better.

[–] moonpiedumplings@programming.dev 16 points 1 day ago* (last edited 1 day ago)

a grand scale with the XZ backdoor

The XZ backdoor, affected a lot less machines than you think. It did not affect:

  • Debian Stable, or Red Hat Enterprise Linux (and rebuilds) — RHEL/rebuilds
  • Linux distros that did not integrate ssh into ssytemd
  • Linux distros that do not use systemd

The malicious code never made it into RHEL or Debian. Both of those distros have a model of freezing packages at a specific version. They then only push manually reviewed security updates, ignoring feature updates or bugfixes to the programs they are packaging. This ensures maximum stability for enterprise usecases, but the way that the changes are small and reviawable also causes them to dodge supply chain attacks like xz (it also enables these distros to have stable auto update features, which I will mention later). But those distros make up a HUGE family of enterprise Linux machines, that were simply untouched by this supply chain attack.

As for linux distros that don't integrate ssh with systemd or non systemd distros being affected, that was because the malware was inactive in those scenarios. Malicious code did make it there, but it didn't activate. I wonder if that was sloppiness on the part of the maker of the malware, or intentional, having it activate less frequently as a way of avoiding detection?

Regardless, comparing the XZ backdoor to the recent NPM and other programming language specific package manager supply chain attacks is a huge false analogy. They aren't comparable at all. Enterprise Linux distros have excellent supply chain security, whereas programming language package managers have basically none. To copy from another comment of mine about them:

Debian Linux, and many other Linux distros, have extensive measures to protect their supply chain. Packages are signed and verified, by multiple developers, before being built reproducibly (I can build and verify and identical binary/package). The build system has layers, such that if only a single layer is compromised, nothing happens and nobody flinches.

Programming langauge specific package repos, have no such protections. A single developer has their key/token/account, and then they can push packages, which are often built on their own devices. There are no reproducible build to ensure the binaries are from the same source code, and no multi-party signing to ensure that multiple devs would need to be compromised in order to compromise the package.

So what happened, probably, is some developer got phished or hacked, and gave up their API key. And the package they made was popular, and frequently ran unsandboxed on devs personal devices, so when other developers downloaded the latest version of that package, they got hacked too. The attackers then used their devices to push more malicious packages to the repo, and the cycle repeats.

And that’s why supply chain attacks are now a daily occurrence.

And then this:

You should probably turn off Dependabot. In my experience, we get more problems from automatic updates than we would by staying on the old versions until needed.

Also drives me insane as well. It's a form of survivorship bias, where people only notice when automatic upgrades cause problems, but they completely ignore the way that automatic security upgrades prevent many issues. Nobody cares about some organization NOT getting ransomwared because their webserver was automatically patched. That doesn't make the news the way that auto upgrades breaking things does. To copy from yet another comment of mine

If your software updates between stable releases break, the root cause is the vendor, rather than auto updating. There exist many projects that manage to auto update without causing problems. For example, Debian doesn't even do features or bugfixes, but only updates apps with security patches for maximum compatibility.

Crowdstrike auto updating also had issues on Linux, even before the big windows bsod incident.

https://www.neowin.net/news/crowdstrike-broke-debian-and-rocky-linux-months-ago-but-no-one-noticed/

It's not the fault of the auto update process, but instead the lack of QA at crowdstrike. And it's the responsibility of the system administrators to vet their software vendors and ensure the models in use don't cause issues like this. Thousands of orgs were happily using Debian/Rocky/RHEL with autoupdates, because those distros have a model of minimal feature/bugfixes and only security patches, ensuring no fuss security auto updates for around a decade for each stable release that had already had it's software extensively tested. Stories of those breaking are few and far between.

I would rather pay attention to the success stories, than the failures. Because in a world without automatic security updates, millions of lazy organizations would be running vulnerable software unknowingly. This already happens, because not all software auto updates. But some is better than none and for all software to be vulnerable by default until a human manually touches it to update it is simply a nightmare to me.

[–] moonpiedumplings@programming.dev 2 points 2 days ago* (last edited 2 days ago)

Wikipedia itself is doing fine but they have a bunch of super interesting side projects that they don't advertise much, and aren't doing as well. Wikinews, their news site is shutting down: https://en.wikipedia.org/wiki/Wikipedia:Wikipedia_Signpost/Single/2026-03-31#News_and_notes (this is really close to april fools hopefully I didn't eat the onion. Or hopefully I did?).

My favorite is wikibooks: http://wikibooks.org/ , which are open source texbooks that can be edited wikipedia style. Their programming one's are really high quality. The idea behind those is that you can export a known good frozen version of them, as a texbook for a class. Related is also wikiversity, which is course curriculum. It's similar, but different.

But they also have a travel voyage, wikivoyage, and more: https://en.wikipedia.org/wiki/Wikipedia:Wikimedia_sister_projects

This is a message to remind myself to share my config later.

I will state that I a, using cloudnativepg for postgres.

The way forgejo actions works, is that it is not a universal thing for every repo. Each repo, can have it's own forgejo actions instance connected to it, running stuff.

The big benefit of that, is that you can make users bring their own actions servers, and not bother to deploy your own.

[–] moonpiedumplings@programming.dev 2 points 4 days ago* (last edited 4 days ago)

It has newer packages than Debian.

This is not quite true. They have overlapping release cycles. A new Debian release will ship frozen versions of the latest packages, causing it to have newer packages than most ubuntu releases. Then the new ubuntu release comes out, with and it has newer packages. Ubuntu doesn't universally newer packages than debian. The difference is that Debian ONLY does security updates, and doesn't do feature updates or even bugfixes over it's lifespan. Ubuntu, on the other hand, does ship feature updates and bug fixes, incrementing the package version as they go over the lifespan of an Ubuntu release.

Comparing the bash versions of the latest ubuntu stable version versus the current debian stable, and you'll notice that Debian has a newer bash:

[moonpie@osiris moonpiedumplings.github.io]$ podman run -it --rm debian
root@980ac170ddb4:/# bash --version
GNU bash, version 5.2.37(1)-release (x86_64-pc-linux-gnu)
Copyright (C) 2022 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>

This is free software; you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.
root@980ac170ddb4:/# exit
exit
[moonpie@osiris moonpiedumplings.github.io]$ podman run -it --rm ubuntu
Resolved "ubuntu" as an alias (/etc/containers/registries.conf.d/00-shortnames.conf)
Trying to pull docker.io/library/ubuntu:latest...
Getting image source signatures
Copying blob 817807f3c64e done   | 
Copying config f794f40ddf done   | 
Writing manifest to image destination
root@1486a1c38699:/# bash --version
GNU bash, version 5.2.21(1)-release (x86_64-pc-linux-gnu)
Copyright (C) 2022 Free Software Foundation, Inc.
License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html>

This is free software; you are free to change and redistribute it.
There is NO WARRANTY, to the extent permitted by law.

This is Ubuntu 24, the current stable release. 25/questing, the rolling version does have newer/same package versions of debian. But people don't base distros off of the rolling version of ubuntu, only the stable releases.

[–] moonpiedumplings@programming.dev 25 points 4 days ago* (last edited 4 days ago)

Debian Linux, and many other Linux distros, have extensive measures to protect their supply chain. Packages are signed and verified, by multiple developers, before being built reproducibly (I can build and verify and identical binary/package). The build system has layers, such that if only a single layer is compromised, nothing happens and nobody flinches.

Programming langauge specific package repos, have no such protections. A single developer has their key/token/account, and then they can push packages, which are often built on their own devices. There are no reproducible build to ensure the binaries are from the same source code, and no multi-party signing to ensure that multiple devs would need to be compromised in order to compromise the package.

So what happened, probably, is some developer got phished or hacked, and gave up their API key. And the package they made was popular, and frequently ran unsandboxed on devs personal devices, so when other developers downloaded the latest version of that package, they got hacked too. The attackers then used their devices to push more malicious packages to the repo, and the cycle repeats.

And that's why supply chain attacks are now a daily occurrence.

No, they're dual licensed. Canonical has users contributing signing a Contributor License agreement, in which they agree to allow Canonical to distribute alternatively licesed, or proprietary versions.

This change was somewhat controversial, and partially why Incus was forked from LXD.

 

Finally I can doomscroll books

 

As usual, phoronix is full of trolls. I was surprised to see only 17 comments, but perhaps that's because I viewed this very early. A highlight from the first page:

Everyday we stray further from GNU, POSIX, C, X11 and now SysVinit. 80s are over. Party is over. Wake up. It's 2026. Adapt or perish in irrelevance. Future is bright and is inevitable. Long live systemd, Wayland, Rust, Gnome and atomic and immutable distros.

Given the way this covers Systemd, SysV, and AI agents, and the way that I see trolling on the first page, There is a very real chance this could be one of those legendary Phoronix threads that manages to hit the 500 comment limit.

EDIT: more relevant threads: https://www.phoronix.com/linux/systemd

 

Youtube video: https://www.youtube.com/watch?v=xrIFL7wSRw4

I am excited about the changes to incus-migrate that allow for direct importation of a remote qcow2 or vmdk. Although many people distribute vmdk's zipped or in tarballs, but it's still a cool feature.

 

Sample with fibonacci:

⍥◡+9∩1 is the fibonacci in this language

 

Here are some cool examples I was looking at:

https://github.com/zardoy/minecraft-web-client — Minecraft in your browser, complete with connections to servers.

https://github.com/inolen/quakejs — quake 3 in your browser, has multiplayer as well.

Any other good examples? or good lists?

 

cross-posted from: https://programming.dev/post/45725210

I noticed in a fairly recent version of KDE, my computer would pretend to be a bluetooth sink when connected to devices like my phone.

This is a really cool feature, and I really like it, because it lets me stream audio from my phone to my computer with no fuss.

However, there is an annoying glitch where the stream stops all of a sudden. The phone keeps playing the music, but I can't hear anything. I've noticed that this seems to have something to do with CPU usage, like when I switch windows rapidly or do something that requires CPU the bluetooth process is dropped. The only reliable way to fix it is to disconnect and reconnect, or wait a minute, and then it works again. Is there any way to fix this more persistently?

I am using CachyOS + KDE right now.

 

I noticed in a fairly recent version of KDE, my computer would pretend to be a bluetooth sink when connected to devices like my phone.

This is a really cool feature, and I really like it, because it lets me stream audio from my phone to my computer with no fuss.

However, there is an annoying glitch where the stream stops all of a sudden. The phone keeps playing the music, but I can't hear anything. I've noticed that this seems to have something to do with CPU usage, like when I switch windows rapidly or do something that requires CPU the bluetooth process is dropped. The only reliable way to fix it is to disconnect and reconnect, or wait a minute, and then it works again. Is there any way to fix this more persistently?

I am using CachyOS + KDE right now.

 

0patch provides "micropatches", that replace running windows code in place, fixing security issues rapidly without requiring an update/reboot.

I really want something like them for an upcoming cybersecurity competition, specifcally patches for the zerologin and eternalblue vulnerabilities.

Unfortunately, 0patch does want a credit card for the free trial, which makes it unfeasible for us to use.

Any alternatives?

 

Has anyone tried this? It's discord reverse engineered.

 

Inspired by this comment.

I'm curious.

 

Tldr we want a static website that will last a long time and also look pretty nice.

Right now, we have a wordpress website. It looks very nice. It also have 4 extensions that aren't configured to auto update. Also whenever I try to make changes to the website they don't apply because the website was configured via the extensions and I hate it.

I want a static site of some kind. It's simple to self host or host anywhere, and it's also simple to secure and keep maintained for a long time.

I am currently looking at static site generators, like quarto, or docusaurus

However, they are difficult to theme to the "niceness" that I want, and their nature results in these somewhat fixed output formats. Like, it is somewhat difficult and annoying to put images anywhere I want them and etc.

Is there like a fixed WYSIWYG html editor? Something between designing a website from scratch and a static site generator. Or is there a way to finagle static site generators to be more flexible than blogs or documentation sites?

 

I hate all three. I understand some of the decisions but other ones are frustrating.

Let me explain what I used to do. What I used to do, is take advantage of the fact that firefox profiles are completely separate instances of firefox, each with their own settings and extensions. I would run my personal profile with highly aggressive and experimental settings, because I was ok with it crashing if it meant I learned interesting things. On the other hand, the profiles related to schoolwork and other more important tasks would be defaults, so they would be much more stable. I no longer consider this a necessary feature, but it was fun to play with.

The other big reason why I relied on the old profiles, is because they have separate cookies and whatnot, which is useful for when I want to have an account for each profile. Although Google happily lets you sign into multiple accounts from the same browser, Microsoft, Discord, and many other apps do not, and force you to sign out before signing in again.

But this is painful. Things never open in the profile I want them to by default, which is annoying. In theory, and I am considering doing this, the way to fix it is by creating app menu shortcuts for each profile, and then having them be the apps I select whenever I want to open a website link or file (with no default profile/app set, so I just select every time).

In addition to that, each profile had to have it's own mozilla account for syncing, which was annoying.

Containers seemed like a nice in between. I could use a single mozilla account for sync, but have seperate microsoft or other accounts on the same browser instance.

Except nope, they actually suck and don't work like that. I can't decide a window is dedicated to a container, so all tabs from xyz site will open in that container and give me that account. It constantly prompts me and it's painful and the UX for what I'm trying to do is miserable.

Containers seem designed more for isolating cookies between two different sites, rather than hiding instances of sites from themselves. Like the original version was a "facebook container", which would hide the facebook cookies from other sites, but I don't want that. I want to be able to log into multiple facebook accounts (hypothetically, I don't actually have a single facebook account but you get the idea).

The new profiles, if you've heard of them, somehow manage to combine the worst of both worlds. Firstly they are an entirely separate system and can't be managed by the second profile system. But they exist within a single one of the old profiles, meaning I can't do tricks with desktop shortcuts to make apps open in one profile or the other. But at the same time, despite existing within one profile, they each require seperate Mozilla accounts for sync.

I am very frustrated, but als resetting up my system so I am considering what to do. I am probably going to continue with profiles, but add app menu shortcuts for them.

Any better ideas?

view more: next ›