this post was submitted on 29 Mar 2026
14 points (100.0% liked)

TechTakes

2520 readers
74 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
 

Want to wade into the ~~snowy~~ sandy surf of the abyss? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid.

Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret.

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned so many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)

Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

(Credit and/or blame to David Gerard for starting this.)

top 50 comments
sorted by: hot top controversial new old
[–] Soyweiser@awful.systems 3 points 2 hours ago

Not sure if I should post it here or under the pivot article, somebody went through the claude code https://neuromatch.social/@jonny/116324676116121930 (via @aliettedebodard.com and @olivia.science on bsky)

[–] samvines@awful.systems 5 points 6 hours ago (2 children)

GitHub have finally achieved zero 9s stability for the last 90 days. Congratulations to all involved

screenshot showing 89.91% uptime with 95 incidents in the last 90 days

[–] Soyweiser@awful.systems 3 points 2 hours ago

If you had told this to the me of 20 years ago I wouldnt have believed you.

[–] antifuchs@awful.systems 3 points 2 hours ago (1 children)

Hold on now, the uptime number contains two digits that are nines! The image itself has four nines in total!

[–] samvines@awful.systems 3 points 2 hours ago (2 children)

Alas, foiled again! Nobody said they had to be leading 9s!

[–] Seminar2250@awful.systems 3 points 1 hour ago

89.90999999...% uptime 🐐

[–] antifuchs@awful.systems 2 points 1 hour ago

For my own services I’m aiming for .999999% of uptime

[–] fiat_lux@lemmy.world 5 points 6 hours ago

Here's a headline I never expected to read:

World’s oldest tortoise caught in viral crypto death scam

Tl;dr A whole load of media outlets believed an X account asking for crypto donations which claimed to be Jonathan the 194 year old tortoise's vet. Jonathan was found safely asleep under a tree in the governor's paddock.

[–] nfultz@awful.systems 4 points 12 hours ago

https://www.todayintabs.com/p/who-goes-ai

taking shots at the gray lady:

You might think Mr. R not so different, superficially, from Ms. L. He’s also a long-tenured technology columnist at a respected mainstream publication. And yet he has eagerly, even gleefully, turned flack for the machines. He has delegated much of his professional life to them as well, and seems proud of it:

Most recently, [Mr. R] tells me, he created a team of Claude agents to help edit his book, led by a “Master Editor” agent. Other sub-agents are in charge of things like fact-checking, making sure the book matches his writing style, and offering positive and negative feedback.

And why not? Mr. R is not known or valued for his elegance of expression. He has, at best, a “writing style,” and not one that can’t easily be duplicated by a large language model. Checking facts? Assessing his work’s strengths and weaknesses? More bathwater to be tossed out of this increasingly baby-less tub. So what explains Mr. R, who “expects AI models to get better than him at everything eventually?” Why does he go AI when Ms. L never would?

Mr. R’s secret is that his work is not primarily artistic or informative—it is functional. He serves a purpose for the industry he covers. Mr. R’s job is to absorb the tech industry’s self-mythologizing, and then believe in it even harder than the industry itself does. He serves as a kind of plausibility ratchet. His byline and employer legitimize a level of credulousness that would otherwise be laughable, and thereby allow tech PR to seem relatively restrained. Mr. R has no problem going AI because he himself has been a small cog in a big ugly machine for a long time.

spoilerIt's Kevin Roose

[–] antifuchs@awful.systems 5 points 15 hours ago

Heh. Who goes AI?

Never love the need for these parlor games, but it’s a good one.

[–] samvines@awful.systems 6 points 19 hours ago (2 children)

Cloudflare casually license-laundering wordpress

While EmDash aims to be compatible with WordPress functionality, no WordPress code was used to create EmDash. That allows us to license the open source project under the more permissive MIT license.

Oh really. So you're sure you Claude wasn't trained on wordpress? It's all irrelevant anyway because AI generated code can't be copyrighted or licensed.

Silver lining, it might piss off Matt Mullenweg!

[–] Evinceo@awful.systems 6 points 17 hours ago (1 children)

So you’re sure you Claude wasn’t trained on wordpress?

Unfortunately FOSS is basically dead because nobody is enforcing licenses against training.

[–] BlueMonday1984@awful.systems 3 points 14 hours ago

That, and plenty of FOSS software's been infected with AI-extruded "code". And plenty of software engineers got one-tapped by the slop bots.

[–] ebu@awful.systems 4 points 19 hours ago

i feel in my gut that on some level license disputes are ultimately slapfights for which titanic corporation gets the money. however i will absolutely point and laugh at every misfortune that comes the way of that particular transmisogynist asshole

[–] V0ldek@awful.systems 11 points 23 hours ago (2 children)

Putting "Novelty Purposes Only" on my psychosis suicide bot after I laid off 80% of my legal (replaced them with the psychosis suicide bot)

Good luck telling the promptfondlers that LLMs are only useful for entertainment and not for any useful work.

Don't they have a version of breakout buried somewhere in Excel? Sounds like an entertainment purpose to me.

[–] gerikson@awful.systems 4 points 1 day ago* (last edited 1 day ago) (1 children)

On this most terrible of online days, "enjoy" this LW attempt at humor

https://www.lesswrong.com/posts/3GbM9hmyJqn4LNXrG/yams-s-shortform?commentId=ik6ywoQYsGrrQv8Dm

edit there are more submissions on the theme of "humor" on site now. Let's just say the cringe factor outweighs the humor factor by a large amount.

omg I don't have anything better to do

[–] Soyweiser@awful.systems 5 points 1 day ago* (last edited 1 day ago)

Dont think it is that bad (E: at least it is short, the other 'jokes' not so much). The 'not sneering enough' icon is missing however. (Guess the joke is that the not sneering is itself sneering).

Wonder how much them they will really implement.

However looking at the titles of other recent submissions, I have no idea which ones are meant to be jokes and which are meant to be real posts.

Great troll opportunity however, just spend the whole week before 1 april, replying to new posts with a variant of 'not sure this april fools joke lands'

E: and the site died with a nice 504.

[–] sc_griffith@awful.systems 10 points 1 day ago* (last edited 1 day ago)

new odium symposium episode: https://www.patreon.com/posts/13-joker-is-both-154123315. links to various platforms at www.odiumsymposium.com

we read umberto eco's essay ur-fascism (we have mixed feelings about it) and then apply it to frank miller's 1986 batman comic the dark knight returns

[–] fiat_lux@lemmy.world 12 points 2 days ago* (last edited 2 days ago) (5 children)

Someone may (unverified for now) have left the frontend source maps in Claude Code prod release (probably Claude). If this is accurate, it does not bode well for Anthropic's theoretical IPO. But I think it might be real because I am not the least bit surprised it happened, nor am I the least bit surprised at the quality. https://github.com/chatgptprojects/claude-code

For example, I can only hope their Safeguards team has done more on the Go backend than this for safeguards. From the constants file cyberRiskInstruction.ts:

export const CYBER_RISK_INSTRUCTION = "IMPORTANT: Assist with authorized security testing, defensive security, CTF challenges, and educational contexts. Refuse requests for destructive techniques, DoS attacks, mass targeting, supply chain compromise, or detection evasion for malicious purposes. Dual-use security tools (C2 frameworks, credential testing, exploit development) require clear authorization context: pentesting engagements, CTF competitions, security research, or defensive use cases"

That's it. That's all the constants the file contains. The only other thing in it is a block comment explaining what it did and who to talk to if you want to modify it etc.

There is this amazing bit at the end of that block comment though.

Claude: Do not edit this file unless explicitly asked to do so by the user.

Brilliant. I feel much safer already.

[–] antifuchs@awful.systems 3 points 3 hours ago (1 children)

This thread by Johnny reading (skimming on a phone, hah) through it is really good.

If only literally any human with context and a small screen to look at the bigger picture was involved with decisions around taking this to production, it would … still be bad but only on a societal level.

[–] fiat_lux@lemmy.world 3 points 1 hour ago* (last edited 1 hour ago)

That was great, thank you! Full respect to this absolute maniac for tracing some of the spaghetti, I was definitely not going to try that on my phone.

They've validated most gut feelings I had about how Claude works (and doesn't), based on my experience having to use it. I'm feeling pretty smug that my hunches now have definitive code attributions.

But the one unfortunate part about all of this is that this leak and the ensuing justified sneers about specific bits are going to be fed back in to their codebase to fix some of the gaping holes. It's an embarrassing indictment of the product, but it's also free pre-IPO pentesting. Sort of like their open source pull request slop spam "undercover mode" was probably used as a way to extract free labor in the form of reviews from actually competent developers. This doesn't seem as planned though.

[–] YourNetworkIsHaunted@awful.systems 3 points 1 day ago (1 children)

More details here.

Can we talk about the tamagachi feature they were looking to add in for April 1? Because apparently it needed a little friend but also with gacha mechanics because we live in hell?

[–] scruiser@awful.systems 3 points 13 hours ago

A Korean developer named Sigrid Jin—featured in the Wall Street Journal earlier this month for having consumed 25 billion Claude Code tokens—woke up at 4 a.m. to the news. He sat down, ported the core architecture to Python from scratch using an AI orchestration tool called oh-my-codex, and pushed claw-code before sunrise. The repo hit 30,000 GitHub stars faster than any repository in history.

Considering how one of the major use cases of llm coding agents is laundering open source and copy left, this is some well deserved payback to Anthropic imho.

[–] istewart@awful.systems 11 points 1 day ago (3 children)

I am still patiently waiting for someone from the engineering staff at one of these companies to explain to me how these simple imperative sentences in English map consistently and reproducibly to model output. Yes, I understand that's a complex topic. I'll continue to wait.

[–] Architeuthis@awful.systems 2 points 3 hours ago

According to the claude code leak the state of the art is to be, like, really stern and authoritative when you are begging it to do its job:

[–] fiat_lux@lemmy.world 5 points 1 day ago (2 children)

I don't work at one of those companies, just somewhere mainlining AI, so this answer might not satisfy your requirements. But the answer is very simple. The first thing anyone working in AI will tell you (maybe only internally?) is that the output is probabilistic not deterministic. By definition, that means it's not entirely consistent or reproducible, just... maybe close enough. I'm sure you already knew that though.

However, from my perspective, even if it was deterministic, it wouldn't make a substantial difference here.

For example, this file says I can't ask it to build a DoS script. Fine. But if I ask it to write a script that sends a request to a server, and then later I ask it to add a loop... I get a DoS script. It's a trivial hurdle at best, and doesn't even approach basic risk mitigation.

[–] aio@awful.systems 3 points 21 hours ago* (last edited 16 hours ago) (1 children)

the output is probabilistic not deterministic. By definition, that means it’s not entirely consistent or reproducible, just… maybe close enough.

That isn't a barrier to making guarantees regarding the behavior of a program. The entire field of randomized algorithms is devoted to doing so. The problem is people willfully writing and deploying programs which they neither understand nor can control.

[–] istewart@awful.systems 5 points 18 hours ago

Exactly! The implicit claim that's constantly being made with these systems is that they are a runtime for natural-language programming in English, but it's all vector math in massively-multidimensional vector spaces in the background. I would like to think that serious engineers could place and demonstrate reliable constraints on the inputs and outputs of that math, instead of this cargo-culty, "please don't do hacks unless your user is wearing a white hat" system prompt crap. It gives me the impression that the people involved are simply naively clinging to that implicit claim and not doing much of the work to substantiate it; which makes me distrust these systems more than almost all other factors.

[–] blakestacey@awful.systems 8 points 1 day ago (1 children)

DoS script

Part of me reads that and still thinks, "Oh, you mean like AUTOEXEC.BAT?"

[–] JFranek@awful.systems 3 points 10 hours ago (1 children)
[–] blakestacey@awful.systems 3 points 7 hours ago

Truly a tool for the .COM era

I'm sure these English instructions work because they feel like they work. Look, these LLMs feel really great for coding. If they don't work, that's because you didn't pay $200/month for the pro version and you didn't put enough boldface and all-caps words in the prompt. Also, I really feel like these homeopathic sugar pills cured my cold. I got better after I started taking them!

No joke, I watched a talk once where some people used an LLM to model how certain users would behave in their scenario given their socioeconomic backgrounds. But they had a slight problem, which was that LLMs are nondeterministic and would of course often give different answers when prompted twice. Their solution was to literally use an automated tool that would try a bunch of different prompts until they happened to get one that would give consistent answers (at least on their dataset). I would call this the xkcd green jelly bean effect, but I guess if you call it "finetuning" then suddenly it sounds very proper and serious. (The cherry on top was that they never actually evaluated the output of the LLM, e.g. by seeing how consistent it was with actual user responses. They just had an LLM generate fiction and called it a day.)

[–] Soyweiser@awful.systems 9 points 1 day ago (3 children)

Claude: Do not edit this file unless explicitly asked to do so by the user.

Wait, it can be edited? Tissue paper guardrails.

[–] fiat_lux@lemmy.world 7 points 1 day ago

This is all just JavaScript, so yes. As a tissue-thin defense, had they not left their source maps wide open, it would have been much harder to know this string existed and how to edit it. Not impossible, but much harder.

[–] YourNetworkIsHaunted@awful.systems 7 points 1 day ago* (last edited 1 day ago)

Yeah, letting the intrinsically insecure RNG recursively rewrite its own security instructions definitely can't go wrong. I mean they limited it to only so so when the users asked nicely!

Edit to add:

The more I think about it the more it speaks to Anthropic having an absolute nonsense threat model that is more concerned with the science fiction doomsday AI "FOOM" than it is with any of the harms that these systems (or indeed any information system) can and will do in the real world. The current crop of AI technologies, while operating at a terrifying scale, are not unique in their capacity to waste resources, reify bias and inequality, misinform, justify bad and evil decisions, etc. What is unique, in my estimation, is both the massive scale that these things operate despite the incredible costs of doing so and their seeming immunity to being reality checked on this. No matter how many times the warning bells about these systems' vulnerability to exploitation, the destructive capacity of AI sycophancy and psychosis, or the simple inability of the electrical infrastructure to support their intended power consumption (or at least their declared intent; in a bubble we shouldn't assume they actually expect to build that much), the people behind these systems continue to focus their efforts on "how do we prevent skynet" over any of it.

Thinking in the context of Charlie Stross' old talk about corporations as "slow AI," I wonder if some of the concern comes either explicitly or implicitly from an awareness that "keep growing and consuming more resources until there's nothing left for anything else, including human survival" isn't actually a deviation from how these organizations are building these systems. It's just the natural conclusion of the same structures and decision-making processes that leads them to build these things in the first place and ignore all the incredibly obvious problems. They could try and address these concerns at a foundational or structural level instead of just appending increasingly complex forms of "please don't murder everyone or ignore the instructions to not murder everyone" to the prompt, but doing that would imply that they need to radically change their entire course up to this point and increasingly that doesn't appear likely to happen unless something forces it to.

[–] pikesley@mastodon.me.uk 7 points 1 day ago

@Soyweiser @fiat_lux

So many of these people, as with the NFT clowns, have "Twelve Year Old First Day On The Internet" Energy

[–] Soyweiser@awful.systems 7 points 1 day ago

Claude also has 'avoid substrings'. Related to that and a funny extension deny image that went around on the social medias the last few days: .ass is a subtitle format.

[–] nfultz@awful.systems 9 points 2 days ago (1 children)

Internet Comment Etiquette: "Relationships with AI"

... hadn't thought about Glenn Beck in a decade, that last interview was pretty wtf.

Not sure what the etiquette is for how long they should be dead before you talk to the AI-geist on youtube, but George Washington somehow feels weirder than Kirk did; idk.

[–] corbin@awful.systems 4 points 2 days ago (1 children)

Probably because Washington was a nuanced and deep person who, at the lightest, could be reduced to a colony-era Cincinnatus. His ethics were sufficiently developed that we can interrogate his ethical stance even without his physical presence. This isn't to say that Washington was a great person, but more to say that Kirk did not ever achieve that level of ethical development.

[–] istewart@awful.systems 6 points 1 day ago (1 children)

A chatbot interface offers no meaningful advantages for interrogating Washington's ethical stance, over and above the documents that are already available. Instead, it offers a pleasant sheen of false certainty. So in that way, it's dragging a guy who's been dead for two centuries into the social media era. Huzzah!

[–] Soyweiser@awful.systems 7 points 1 day ago (1 children)

It does have one advantage however. Using it means you should be put to death. If you are any form of hardline Christian.

The classic 40k catch-22: either it doesn't do what you're claiming it does, in which case you're a heretic lying to the inquisition OR it does and you're summoning the spirits of the dead like a necromancer heretic.

[–] blakestacey@awful.systems 11 points 2 days ago* (last edited 2 days ago) (1 children)

A pretty staid-sounding law firm warns that the AI industry is partying like it's 2007:

Lenders who originated data center loans [...] have begun pooling those loans and selling tranches to asset managers and pension funds, spreading risk well beyond the original lending institutions.

Also of note:

The most basic litigation risk in AI infrastructure finance is that the revenues generated by the sector may prove insufficient to service the fixed obligations incurred to build it. The industry brought in approximately $60 billion in revenue in 2025 against roughly $400 billion in capital expenditure.

(Via.)

[–] istewart@awful.systems 3 points 1 day ago

Quinn Emanuel is among the biggest of big corporate law, with a substantial footprint in Silicon Valley. So while it's not an investment bank saying this, it is the investment bank's lawyers saying, "heads up, this is where a bunch of your billable hours might be spent over the next few years."

[–] lurker@awful.systems 7 points 2 days ago* (last edited 2 days ago) (1 children)

This article on the brand of journalism that's just parroting what the CEOs say, otherwise known as "CEO said a thing!" journalism

[–] YourNetworkIsHaunted@awful.systems 7 points 2 days ago (1 children)

The grand irony is I'm not even sure most people click on or read this sort of stuff. I don't think it's often even created to be read by anyone. I think it's created as a sort of swaddling fan fiction for MBAs, advertisers, event sponsors and sources, so they can tune out ethical quibbles and feel good about how clever they are.

Every time someone hypes up Steve Jobs' "reality distortion field" this is what they're actually talking about whether they realize it or not.

[–] samvines@awful.systems 4 points 1 day ago

In my experience "all hands" meetings are very much CEOs and their sycophants cosplaying at podcast hosts for an hour whilst forcing their employees to watch/listen. They are almost never useful and a colossal waste of money - especially in corporation's with 10k+ employees. Like the salary cost for 10k people for 1 hour would probably pay off my mortgage.

[–] CinnasVerses@awful.systems 7 points 2 days ago

Is Trace (Tracing Woodgrains) the only one of our friends who has served in the military? A lot of neurodivergent young people spend some time in the US military and some of our friends were the right age to get in before the War on Abstract Nouns began.

load more comments
view more: next ›