this post was submitted on 19 May 2025
302 points (96.6% liked)

A Boring Dystopia

12176 readers
488 users here now

Pictures, Videos, Articles showing just how boring it is to live in a dystopic society, or with signs of a dystopic society.

Rules (Subject to Change)

--Be a Decent Human Being

--Posting news articles: include the source name and exact title from article in your post title

--If a picture is just a screenshot of an article, link the article

--If a video's content isn't clear from title, write a short summary so people know what it's about.

--Posts must have something to do with the topic

--Zero tolerance for Racism/Sexism/Ableism/etc.

--No NSFW content

--Abide by the rules of lemmy.world

founded 2 years ago
MODERATORS
 

cross-posted from: https://hexbear.net/post/4958707

I find this bleak in ways it’s hard to even convey

all 46 comments
sorted by: hot top controversial new old
[–] SoftestSapphic@lemmy.world 8 points 2 hours ago (1 children)

Nothing will meaningfully improve until the rich fear for their lives

[–] DeathsEmbrace@lemm.ee 0 points 1 hour ago

In a way that the relief is to give us our demands subliminally. This way the only rich person who is safe is our subject.

[–] markovs_gun@lemmy.world 8 points 2 hours ago (1 children)

I can't wait until ChatGPT starts inserting ads into its responses. "Wow that sounds really tough. You should learn to love yourself and not be so hard on yourself when you mess up. It's a really good thing to treat yourself occasionally, such as with an ice cold Coca-Cola or maybe a large order of McDonald's French fries!"

[–] thermal_shock@lemmy.world 4 points 2 hours ago* (last edited 2 hours ago) (1 children)
[–] Retrograde@lemmy.world 1 points 1 hour ago* (last edited 1 hour ago)

That episode was so disturbing 😅

[–] Cyberflunk@lemmy.world 3 points 2 hours ago

I've tried this ai therapist thing, and it's awful. It's ok to help you work out what you're thinking, but absymal at analyzing you. I got some structured timelines back fro. It that I USED in therapy, but AI is a dangerous alternative to human therapy.

My $.02 anyway.

[–] idunnololz@lemmy.world 12 points 4 hours ago* (last edited 4 hours ago)

This is terrible. I'm going to ignore the issues concerning privacy since that's already been brought up here and highlight another major issue: it's going to get people hurt.

I did a deep dive with gen AI for a month a few weeks ago.

It taught me that gen AI is actually brilliant at certain things. One thing that gen AI does is it learns what you want and makes you believe it’s giving you exactly what you want. In a sense it's actually incredibly manipulative and one of the things gen AI is brilliant at. As you interact with gen AI within the same context window, it quickly picks up on who you are, then subtly tailors its responses to you.

I also noticed that as gen AI's context grew, it became less "objective". This makes sense since gen AI is likely tailoring the responses for me specifically. However, when this happens, the responses also end up being wrong more often. This also tracks, since correct answers are usually objective.

If people started to use gen AI for therapy, it's very likely they will converse within one context window. In addition, they will also likely ask gen AI for advice (or gen AI may even offer advice unprompted because it loves doing that). However, this is where things can go really wrong.

Gen AI cannot "think" of a solution, evaluate the downsides of the solution, and then offer it to you because gen AI can't "think" period. What gen AI will do is it will offer you what sounds like solutions and reasons. And because gen AI is so good at understanding who you are and what you want, it will frame the solutions and reasons in a way that appeals to you. On top of all of this, due to the long-running context window, it's very likely the advice gen AI gives will be bad advice. For someone who is in a vulnerable and emotional state, the advice may seem reasonable, good even.

If people then act on this advice, the consequences can be disastrous. I've read enough horror stories about this.

Anyway, I think therapy might be one of the worst uses for gen AI.

[–] match@pawb.social 2 points 3 hours ago (1 children)

unlike humans, the ai listens to and remembers me to me [for the number of characters allotted]. this will help me feel seen i guess

[–] Viking_Hippie@lemmy.dbzer0.com 2 points 3 hours ago

You know a reply's gonna be good when it starts with "unlike humans" 😁

[–] november@lemmy.vg 11 points 8 hours ago
[–] SouthEndSunset@lemm.ee 12 points 9 hours ago

Cheaper than paying people better, I suppose.

[–] drmoose@lemmy.world 16 points 10 hours ago (1 children)

People's lack of awareness of how important accessibility is really shows in this thread.

Privacy leaking is much lesser issue than not having anyone to talk to for many people, especially in poorer countries.

[–] samus12345@lemm.ee 1 points 3 hours ago

Is Australia a poorer country?

[–] ZILtoid1991@lemmy.world 30 points 12 hours ago (1 children)

Am I old fashioned for wanting to talk to real humans instead?

[–] GreenMartian@lemmy.dbzer0.com 39 points 12 hours ago* (last edited 12 hours ago) (2 children)

No. But when the options are either:

  • Shitty friends who have better things to do than hearing you vent,
  • Paying $400/hr to talk to a psychologist, or
  • A free AI that not only pays attention to you, but actually remembers what you told them last week,

it's quite understandable that some people choose the one that is a privacy nightmare but keeps them sane and away from some dark thoughts.

[–] Natanael@infosec.pub 27 points 12 hours ago (1 children)
[–] anus@lemmy.world 2 points 4 hours ago

Ahh yes the random rolling stone article that refutes the point

Let's revisit the list, shall we?

[–] ZILtoid1991@lemmy.world 11 points 12 hours ago (3 children)

But I want to hear other people's vents...😥

[–] Viking_Hippie@lemmy.dbzer0.com 20 points 12 hours ago

Maybe a career in HVAC repair is just the thing for you!

[–] lilmo037@infosec.pub 10 points 10 hours ago

Please continue to be you, we need more folks like you.

[–] GreenMartian@lemmy.dbzer0.com 10 points 12 hours ago

You're a good friend. I wish everyone has someone like this. I have a very small group of mates where I can be vulnerable without being judged. But not everyone are as privileged, unfortunately...

[–] ryedaft@sh.itjust.works 38 points 13 hours ago (1 children)
[–] drmoose@lemmy.world 4 points 10 hours ago (1 children)

Yeah we have spiritual delusions at home already!

Seriously, no new spiritual delusions could ever be more harmful than what we have right now.

[–] DeceasedPassenger@lemmy.world 8 points 10 hours ago* (last edited 9 hours ago) (1 children)

Totally fair point but I really don't know if that's true. Most mainstream delusions have the side effect of creating community and bringing people together, other negative aspects notwithstanding. The delusions referenced in the article are more akin to acute psychosis, as the individual becomes isolated, nobody to share delusions with but the chatbot.

With traditional mainstream delusions, there also exists a relatively clear path out, with corresponding communities. ExJW, ExChristian, etc. People are able to help others escape that particular in-group when they're familiar with how it works. But how do you deprogram someone when they've been programmed with gibberish? It's like reverse engineering a black box. This is scaring me as I write it.

[–] theneverfox@pawb.social 3 points 8 hours ago (1 children)

This isn't a new thing, people have gone off alone into this kind of nonsensical journey for a while now

The time cube guy comes to mind

There's also temple OS written in holy C, he was close to some of the stuff in the article

And these are just two people functional and loud enough to be heard. This is a thing that happens, maybe LLMs exacerbate a pre existing condition, but people have been going off the deep end like this long before LLMs came into the picture

[–] DeceasedPassenger@lemmy.world 3 points 6 hours ago* (last edited 6 hours ago)

Your point is not only valid but also significant, and I feel stands in addition, not contradiction, to my point. These people now have something to continuously bounce ideas off; a conversational partner that never says no. A perpetual yes-man. The models are heavily biased towards the positive simply by nature of what they are, predicting what comes next. You (may or may not) know how in improv acting there's a saying called "yes, and" which serves to keep things always moving forward. These models effectively exist in this state, in perpetuity.

Previously, people who have ideas such as these will experience near-universal rejection from those around them (if they don't have charisma in which case they start a cult) which results in a (relatively, imo) small number of extreme cases. I fear the presence of such a perpetual yes-man will only accelerate all kinds of damage that can emerge from nonsensical thinking.

[–] SpicyLizards@reddthat.com 9 points 10 hours ago

Enter the Desolatrix

[–] JustJack23@slrpnk.net 25 points 14 hours ago (2 children)

If the title is a question, the answer is no

[–] sawdustprophet@midwest.social 9 points 11 hours ago (1 children)

If the title is a question, the answer is no

A student of Betteridge, I see.

[–] JustJack23@slrpnk.net 6 points 11 hours ago

Actually I read it in a forum somewhere, but I am glad I know the source now!

[–] Viking_Hippie@lemmy.dbzer0.com 3 points 12 hours ago* (last edited 12 hours ago)

What is a sarcastic rhetorical question?

[–] ininewcrow@lemmy.ca 93 points 17 hours ago (4 children)

A human therapist might not or is less likely to share any personal details about your conversations with anyone.

An AI therapist will collate, collect, catalog, store and share every single personal detail about you with the company that owns the AI and share and sell all your data to the highest bidder.

the AI therapist probably can't force you into a psych ward though, a human psychologist is obligated to (under the right conditions).

[–] DaddleDew@lemmy.world 51 points 17 hours ago* (last edited 17 hours ago) (1 children)

Neither would a human therapist be inclined to find the perfect way to use all this information to manipulate people while they are being at their weakest. Let alone do it to thousands, if not millions of them all at the same time.

They are also pushing for the idea of an AI "social circle" for increasingly socially isolated people through which world view and opinions can be bent to whatever whoever controls the AI desires.

To that we add the fact that we now know they've been experimenting with tweaking Grok to make it push all sorts of political opinions and conspiracy theories. And before that, they manipulated Twitter's algorithm to promote their political views.

Knowing all this, it becomes apparent that we are currently witnessing is a push for a whole new level of human mind manipulation and control experiment that will make the Cambridge Analytica scandal look like a fun joke.

Forget Neuralink. Musk already has a direct connection into the brains of many people.

[–] fullsquare@awful.systems 13 points 17 hours ago

PSA that Nadella, Musk, saltman (and handful of other techfash) own dials that can bias their chatbots in any way they please. If you use chatbots for writing anything, they control how racist your output will be

[–] bitjunkie@lemmy.world 2 points 10 hours ago

The data isn't useful if the person no longer exists.

[–] Crewman@sopuli.xyz 7 points 17 hours ago (2 children)

You're not wrong, but isnt that also how Better Help works?

[–] essell@lemmy.world 10 points 16 hours ago

Better help is the Amazon of the Therapy world.

[–] Kyrgizion@lemmy.world 29 points 16 hours ago

I suppose this can be mitigated by installing a local LLM that doesn't phone home. But there's still a risk of getting downright bad advice since so many LLM's just tell their users they're always right or twist the facts to fit that view.

I've been guilty of this as well, I've used ChatGPT as a "therapist" before. It actually gives decently helpful advice, compared to what's out there available after a google search. But I'm fully aware of the risks "down the road", so to speak.

[–] Zagorath@aussie.zone 22 points 16 hours ago
[–] adarza@lemmy.ca 18 points 17 hours ago (1 children)

how long will it take an 'ai' chatbot to spiral downward to bad advice, lies, insults, and/or promotion of violence and self-harm?

[–] Whats_your_reasoning@lemmy.world 8 points 11 hours ago

We're already there. Though that violence didn't happen due to insults, but due to a yes-bot affirming the ideas of a mentally-ill teenager.

[–] stinky@redlemmy.com -4 points 10 hours ago

Other people: my outlook is so bad I turned to religion to feel better

You: EW GROSS DON'T DO THAT

Other people: ok then I'll turn to technology instead

You: EW GROSS DON'T DO THAT