this post was submitted on 24 Jun 2024
4 points (100.0% liked)

Today I Learned

26518 readers
498 users here now

What did you learn today? Share it with us!

We learn something new every day. This is a community dedicated to informing each other and helping to spread knowledge.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules (interactive)


Rule 1- All posts must begin with TIL. Linking to a source of info is optional, but highly recommended as it helps to spark discussion.

** Posts must be about an actual fact that you have learned, but it doesn't matter if you learned it today. See Rule 6 for all exceptions.**



Rule 2- Your post subject cannot be illegal or NSFW material.

Your post subject cannot be illegal or NSFW material. You will be warned first, banned second.



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Posts and comments which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding non-TIL posts.

Provided it is about the community itself, you may post non-TIL posts using the [META] tag on your post title.



Rule 7- You can't harass or disturb other members.

If you vocally harass or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.

For further explanation, clarification and feedback about this rule, you may follow this link.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here.

Unless included in our Whitelist for Bots, your bot will not be allowed to participate in this community. To have your bot whitelisted, please contact the moderators for a short review.



Partnered Communities

You can view our partnered communities list by following this link. To partner with our community and be included, you are free to message the moderators or comment on a pinned post.

Community Moderation

For inquiry on becoming a moderator of this community, you may comment on the pinned post of the time, or simply shoot a message to the current moderators.

founded 2 years ago
MODERATORS
 

Roko's basilisk is a thought experiment which states that an otherwise benevolent artificial superintelligence (AI) in the future would be incentivized to create a virtual reality simulation to torture anyone who knew of its potential existence but did not directly contribute to its advancement or development, in order to incentivize said advancement.It originated in a 2010 post at discussion board LessWrong, a technical forum focused on analytical rational enquiry. The thought experiment's name derives from the poster of the article (Roko) and the basilisk, a mythical creature capable of destroying enemies with its stare.

While the theory was initially dismissed as nothing but conjecture or speculation by many LessWrong users, LessWrong co-founder Eliezer Yudkowsky reported users who panicked upon reading the theory, due to its stipulation that knowing about the theory and its basilisk made one vulnerable to the basilisk itself. This led to discussion of the basilisk on the site being banned for five years. However, these reports were later dismissed as being exaggerations or inconsequential, and the theory itself was dismissed as nonsense, including by Yudkowsky himself. Even after the post's discreditation, it is still used as an example of principles such as Bayesian probability and implicit religion. It is also regarded as a simplified, derivative version of Pascal's wager.

Found out about this after stumbling upon this Kyle Hill video on the subject. It reminds me a little bit of "The Game".

top 15 comments
sorted by: hot top controversial new old
[–] moosetwin@lemmy.dbzer0.com 2 points 2 years ago (2 children)

roko's basilisk is a type of infohazard known as 'really dumb if you think about it'

also I have lost the game (which is a type of infohazard known as 'really funny')

[–] decivex@yiffit.net 1 points 2 years ago (1 children)

Thanks! I just won the game!

[–] PlexSheep@infosec.pub 0 points 2 years ago (1 children)

Winning wasn't in the set of rules I received, can you explain?

[–] decivex@yiffit.net 1 points 2 years ago

Make your own rules.

[–] AnarchistArtificer@slrpnk.net 1 points 2 years ago (1 children)

Oh damn, I just lost the game too, and now I'm thinking about the game as if it were a virus - like, I reckon we really managed to flatten the curve for a few years there, but it continues to circulate so we haven't been able to eradicate it

[–] PlexSheep@infosec.pub 1 points 2 years ago

I lost too. I agree, it's been going around at least in the threadiverse. I've seen it at least 3 times in a couple months.

[–] samus12345@lemmy.world 0 points 2 years ago (1 children)

Pascal's Wager always seemed really flawed to me even through a purely Christian perspective. You're saying that god is so oblivious (even though he's supposed to be omniscient) that he'll be fooled by you claiming to believe just because you're hedging your bets? The actual reason it's dumb is that it's not a binary choice since there are thousands of ways people claim you can be saved in various religions.

[–] UnderpantsWeevil@lemmy.world -1 points 2 years ago

You’re saying that god is so oblivious (even though he’s supposed to be omniscient) that he’ll be fooled by you claiming to believe just because you’re hedging your bets?

More that repetition reinforces an idea. By commiting to the bit and accepting a God at face value, you reduce your psychological defenses when the priest or prophet comes around with the next ask.

So you admit you believe in God? Then you won't mind putting a few coins in the collection plate to prove it.

Oh, you've already donated? Surely you'd be comfortable making a confession.

My son, you've got so many sins! Surely you'd like to join our prayer group to get yourself right with the God we all agree exists.

Can't have prayer without works! Time to do some penance.

[–] nicknonya@lemmy.blahaj.zone 0 points 2 years ago (1 children)

it has been said before and i'll say it again: Pascal's wager for tech bros

[–] Cosmicomical@lemmy.world 0 points 2 years ago (1 children)

but not as easily dismissable

[–] Saledovil@sh.itjust.works 1 points 2 years ago

Roko's Basilisk hinges on the concept of acausal trade. Future events can cause past events if both actors can sufficiently predict each other. The obvious problem with acausal trade is that if you're the actor B in the future, then you can't change what the actor A in the past did. It's A's prediction of B's action that causes A's action, not B's action. Meaning the AI in the future gains literally nothing by exacting petty vengeance on people who didn't support their creation.

Another thing Roko's Basilisk hinges on is that a copy of you is also you. If you don't believe that, then torturing a simulated copy of you doesn't need to bother you any more than if the AI tortured a random innocent person. On a related note, the AI may not be able to create a perfect copy of you. If you die before the AI is created, and nobody scans your brain (Brain scanners currently don't exist), then the AI will only have the surviving historical records of you to reconstruct you. It may be able to create an imitation so convincing that any historian, and even people who knew you personally will say it's you, but it won't be you. Some pieces of you will be forever lost.

Then a singularity type superintelligence might not be possible. The idea behind the singularity is that once we build an AI, the AI will then improve itself, and then they will be able to improve itself faster, thus leading to an exponential growth in intelligence. The problem is that it basically assumes that the marginal effort of getting more intelligent grows slower than linearly. If the marginal difficulty grows as fast as the intelligence of the AI, then the AI will become more and more intelligent, but we won't see an exponential increase in intelligence. My guess would be that we'd see a logistical growth of intelligence. As in, the AI will first become more and more intelligent, and then the growth will slow and eventually stagnate.

[–] UnPassive@lemmy.world 0 points 2 years ago (1 children)

I was raised Mormon (LDS) and there are parallels; basically they believe Mormonism is the one true and complete denomination of Christianity and once you learn this, you need to spread that truth (mandatory 2 year missions for men, and a STRONG culture of missionary work through life), also, no one goes to hell in Mormonism except those who learned this truth and then later denied it/left it (called a son of perdition).

So my parents believe I'll go to hell without the likes of Hitler because he never was taught "the truth" lol

[–] UnderpantsWeevil@lemmy.world -1 points 2 years ago

Hell without Hitler doesn't really sound so bad

[–] Mango@lemmy.world 0 points 2 years ago (1 children)
[–] UnderpantsWeevil@lemmy.world -1 points 2 years ago

More like a Chain Letter.

There was a trend when I was a little kid of people sending you mail that said something to the effect of "You have been cursed by reading this letter. If you don't mail a copy to ten other people, you will die in thirty days."

Roko's Basilisk is a modern manifestation of human paranoia and superstition. It exists to exploit and extort the gullible.