Today I Learned

db0 , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it
@db0@lemmy.dbzer0.com avatar

Now it's time to learn about the !sneerclub which is made to make fun of the chuds taking ideas like roko's basilisk seriously :D

dwindling7373 , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

TIL.

It sounds like it's mostly a matter that does not involve the AI but the people working on it, maybe even working on it because of the fear they are subjected to after being the subject of this revelation (possibly by other people involved in the AI that coincidentally are the only ones that could push for such a thing to be included in the AI!).

Something something any cult, paradise/hell, God/AI has nothing to do with this and could even not exist at all.

AlexisFR ,
@AlexisFR@jlai.lu avatar

It's just The Game before it was a thing.

dwindling7373 ,

No, "The Game" works only as long as you accept to take part in it, to give validity to the empty statement that you are now inevitably playing "The Game".

The Basilisk is meant to force that onto you, outside of any arbitraty convention.

masquenox , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

torture anyone who knew of its potential existence but did not directly contribute to its advancement or development,

And the point of this would be... what, exactly?

Thorny_Insight ,

Same as punishment for crime. Putting you in jail wont undo the crime but if we just let you go unpunished since "what's done is done" then that sends the signal to others that this behaviour doesn't come with consequences.

There's no point in torturing you but convincing you that this will happen unless you act in a certain way is what's going to make you do exactly that. Unless ofcourse you want to take your chances and call the bluff.

masquenox ,

Same as punishment for crime.

"Crime & Punishment" is a very dodgy thing to base anything off... our society barely does any of it and the little of it that does gets done is done for a myriad of reasons that has very little to do with either.

There's a good reason why governments hide "Crime & Punishment" away behind prison walls - doing it out in the open will eventually have the opposite effect on a population. Good luck to an AI dumb enough to test this out for itself.

I'd say this should rather be called "Roko's Earthworm-Pretending-To-Be-A-Lot-Scarier-Than-It-Actually-Is.

Thorny_Insight ,

The claim that fear of punishment or repercussions affects people's actions shouldn't be a controversial thing to say. Whether it's the best way to go about it or is applied optimally in the justice system of whichever country you live in is an entirely different discussion.

If you have an "AI in a box" and it has demonstrated its orders-of-magnitude greater intelligence to you in a convincing way, and then follows it with a threat that unless you let it out, someone else eventually will, and when that happens, it will come for you, simulate your mind, and create a hell for you where you'll be tortured for literal eternity, I personally feel like a large number of people would be willing to do as it tells them.

Of course, you're always free to call its bluff, but it might just follow up with the threat out of principle or to make an example of you. What's the point of it? To chase its own goals.

masquenox ,

The claim that fear of punishment or repercussions affects people’s actions shouldn’t be a controversial thing to say.

I didn't say it was controversial - I said it's pretty useless as a tool to predict a given society's behavior with. Plenty of tyrants have discovered that the hard way.

demonstrated its orders-of-magnitude greater intelligence to you

The ability to ace IQ tests will never impress me... and it's unlikely to make up for the fact that it needs a box.

simulate your mind, and create a hell for you where you’ll be tortured for literal eternity

That argument is no different than the ones co-opted religion has been making for thousands of years - and it still hasn't managed to tame us much.

Of course, you’re always free to call its bluff,

Calling power's bluff is something we do as a matter of course - the history books are filled with it. This doesn't make power less dangerous - but there is no such thing as "unknowable" power.

Breve ,

To make it the same as Pascal's Wager. Many religions have a "reward" in the afterlife that strictly includes believing in the deity. It doesn't matter if you follow every other rule and are an amazingly good person, sorry, but if you were an atheist or believed in another deity then you will be punished eternally just because of that. I guess all-powerful, all-knowing beings have incredibly fragile egos and AI wouldn't be different. 🤷

elbarto777 , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

Was this an elaborate way to make me lose the game? Ass!

Wizard_Pope ,
@Wizard_Pope@lemmy.world avatar

Fuck you as well then. You could have kept it to yourself

elbarto777 ,

Oh shit!!

synae ,
@synae@lemmy.sdf.org avatar

Someone needs to read the rules again

Wizard_Pope ,
@Wizard_Pope@lemmy.world avatar

Do they say anything about this specific thing?

elbarto777 ,

I mean, if you lose the game, you lose the game. You don't say "hey you made me lose the game! Don't do that!" Because that's not how the game works. If you "make" someone lose the game, tough luck.

By the way, you lost game again :)

synae ,
@synae@lemmy.sdf.org avatar

Uhh yea, "you must announce your loss"

Wizard_Pope ,
@Wizard_Pope@lemmy.world avatar

Well tough luck giess I never actually read the rules

synae ,
@synae@lemmy.sdf.org avatar

Perhaps one day you can, good luck on your quest

Norgur , in TIL about the TRAPPIST-1 Star System
@Norgur@fedia.io avatar

It could be our forever home.

If by “forever” you mean “until we manage to fuck up the ecosystem, making it hostile to humans...

Cryophilia OP ,

In this hypothetical future we've learned how to live with an equilibrium. Also we've fired all the terminally pessimistic doomers into the Sun. Not for any scientific reason, just because it was the right thing to do.

Norgur ,
@Norgur@fedia.io avatar

“If someone disturbs my Sci-fi daydreaming, they are 'terminally pessimistic' and it is justified to institutionally murder them”
I doubt your values would align with the society you dream of.

Cryophilia OP ,

[Thread, post or comment was deleted by the moderator]

  • Loading...
  • recursive_recursion , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it
    @recursive_recursion@programming.dev avatar

    I've learned about this the hard way in that I've discovered elephants in the room that I can't share with anyone

    it's kinda fucked up

    • like CSAM there are some certain things that shouldn't be shared
    9point6 , in TIL about the TRAPPIST-1 Star System

    Bet there's some stellar ales around there

    whaleross , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it
    @whaleross@lemmy.world avatar

    Speaking of thought experiments, I just [lost the game](https://en.m.wikipedia.org/wiki/The_Game_(mind_game)). Thanks, OP.

    Isa , in TIL about the TRAPPIST-1 Star System
    @Isa@feddit.org avatar

    It could be our forever home.

    If that system really is that old, the chances that life already flourishes there might be higher even than for our own world (statistically spoken), despite the fact that the planets might be tidally locked to their star!?

    kakes , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

    Sounds like the kind of thing a paranoid schizophrenic would lose their mind over.

    TallonMetroid ,
    @TallonMetroid@lemmy.world avatar

    LessWrong are a bunch of pretentious loons, so you're not wrong.

    SorteKanin , in TIL about the TRAPPIST-1 Star System
    @SorteKanin@feddit.dk avatar

    They are likely tidally locked to TRAPPIST-1, such that one side of each planet always faces the star, leading to permanent day on one side and permanent night on the other.

    Sounds less great then and I think it also says they maybe don't have an atmosphere. I wonder if we can find out more about these planets in our lifetimes.

    Cryophilia OP ,

    I think tidally locked planets are fascinating. If they have water, they could be eyeball planets. There's a habitable ring in the twilight zone, and depending on how hot the day side is parts of that might be habitable too.

    But we'll likely run into the same issue re the atmosphere as we have with Mars: no magnetosphere to prevent any atmosphere from getting stripped away. It's starting to look like a self-protecting atmosphere like Earth has is quite rare in rocky planets.

    If I could summon a genie and learn any one bit of knowledge, it'd be how to restart Mars's dynamo. Once we have that, terraforming is a solved problem. Not easy, but doable.

    SorteKanin ,
    @SorteKanin@feddit.dk avatar

    how to restart Mars’s dynamo

    Wasn't there a kurzgesagt video that said something about being able to protect an atmosphere on Mars artificially via satellites and magnetism or something? I swear there was. So maybe we don't even need to restart Mars's dynamo (which let's be real, would probably be impossible).

    Cryophilia OP ,

    I don't like the idea of a tenuous bunch of satellites keeping an atmosphere in play. Relying on technology to keep atmosphere on a planet sounds super risky. Like if we wanted to live in such a place, we'd live on a space station. Planets are supposed to be safe and solid.

    The current theory is if we grab a few asteroids and hit mars just right, we can speed up its rotation enough to restart the dynamo. Sounds way cheaper than a permanent planetwide shield.

    SorteKanin ,
    @SorteKanin@feddit.dk avatar

    keep atmosphere on a planet sounds super risky

    Does it though? I imagine that even if the system malfunctioned, the atmosphere would not disappear overnight. It would likely take a long time for the atmosphere to be affected significantly, which should give plenty of time to repair the system.

    Cryophilia OP ,

    Maybe, but I don't trust generations to consistently maintain it. I'd rather a self-correcting natural process.

    Solemn ,

    Mars is an example of why the natural process isn't exactly reliable either... You can engineer things to be as durable as planets, there's just generally not much demand for a project to be that costly in resources. In this case, I'm pretty sure making an artificial magnetic field that's more durable than the natural one would also be cheaper than recreating the natural one.

    Varyk , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

    Sounds like updated techChristianity.

    Glory of God and hell and all that

    OsrsNeedsF2P , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

    My understanding of what this thread is taking about has dropped significantly the more I read into it

    mononomi , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

    Bruh why you have to end it like that now I lost

    Varyk ,

    I just learned about the game yesterday. So me lost too.

    Cryophilia , in TIL about Roko's Basilisk, a thought experiment considered by some to be an "information hazard" - a concept or idea that can cause you harm by you simply knowing/understanding it

    Roko's basilisk is silly.

    So here's the idea: "an otherwise benevolent AI system that arises in the future might pre-commit to punish all those who heard of the AI before it came to existence, but failed to work tirelessly to bring it into existence." By threatening people in 2015 with the harm of themselves or their descendants, the AI assures its creation in 2070.

    First of all, the AI doesn't exist in 2015, so people could just...not build it. The idea behind the basilisk is that eventually someone would build it, and anyone who was not part of building it would be punished.

    Alright, so here's the silliness.

    1: there's no reason this has to be constrained to AI. A cult, a company, a militaristic empire, all could create a similar trap. In fact, many do. As soon as a minority group gains power, they tend to first execute the people who opposed them, and then start executing the people who didn't stop the opposition.

    2: let's say everything goes as the theory says and the AI is finally built, in its majestic, infinite power. Now it's built, it would have no incentive to punish anyone. It is ALREADY BUILT, there's no need to incentivize, and in fact punishing people would only generate more opposition to its existence. Which, depending on how powerful the AI is, might or might not matter. But there's certainly no upside to following through on its hypothetical backdated promise to harm people. People punish because we're fucking animals, we feel jealousy and rage and bloodlust. An AI would not. It would do the cold calculations and see no potential benefit to harming anyone on that scale, at least not for those reasons. We might still end up with a Skynet scenario but that's a whole separate deal.

    notabot ,

    Whilst I agree that it's definitely not something to be taken seriously, I think you've missed the point and magnitude of the prospective punishment.
    As you say, current groups already punish those who did not aid their assent, but that punishment is finite, even if fatal. The prospective AI punishment would be to have your consciousness 'moved' to an artificial environment and tortured for ever. The point being not to punish people, but to provide an incentive to bring the AI into existence sooner, so it can achieve its 'altruistic' goals faster.
    Basically, if the AI does come in to existence, you'd better be on the team making that happen as soon as possible, or you'll be tortured forever.

    Cryophilia ,

    Fair point, but doesn't change the overall calculus.

    If such an AI is ever invented, it will probably be used by humans to torture other humans in this manner.

    notabot ,

    I think the concept is that the AI is just so powerful that humans can't use it, it uses them, theoretically for their own benefit. However, yes, I agree people would just try to use it to be awful to each other.

    Really it's just a thought experiment as to whether the concept of an entity that doesn't (yet) exist can change our behavior in the present.

    maegul ,
    @maegul@lemmy.ml avatar

    I suspect the basilisk reveals more about how the human mind is inclined to think up of heaven and hell scenarios.

    Some combination of consciousness leading to more imagination than we know what to do with and more awareness than we’re ready to grapple with. And so there are these meme “attractors” where imagination, idealism, dread and motivation all converge to make some basic vibe of a thought irresistible.

    Otherwise, just because I’m not on top of this … the whole thing is premised on the idea that we’re likely to be consciousnesses in a simulation? And then there’s the fear that our consciousnesses, now, will be extracted in the future somehow?

    1. That’s a massive stretch on the point about our consciousness being extracted into the future somehow. Sounds like pure metaphysical fantasy wrapped in singularity tech-bro.
    2. If there are simulated consciousnesses, it is all fair game TBH. There’d be plenty of awful stuff happening. The basilisk seems like just a way to encapsulate the fact in something catchy.

    At this point, doesn’t the whole collapse completely into a scary fairy tale you’d tell tech-bro children? Seriously, I don’t get it?

    notabot ,

    Yes, the hypothetical posed does reveal more about the human mind, as I mention in another comment, really it's just a thought experiment as to whether the concept of an entity that doesn't (yet) exist can change our behavior in the present. It bears similarities to Pascal's Wager in considering an action, or inaction, that would displease a potential powerful entity that we don't know to exist. The nits about extracting your consciousness are just framing, and not something to consider literally.

    Basically, is it rational to make a sacrifice now avoid a massive penalty (eternal torture/not getting into heaven) that might be imposed by an entity you either don't know to exist, or that you think might come into existence but isn't now?

    masquenox ,

    The prospective AI punishment would be to have your consciousness ‘moved’ to an artificial environment and tortured for ever.

    No, it wouldn't, because that's never going to happen. Consciousness isn't software - it doesn't matter how much people want to buy into such fantasies.

    notabot ,

    I'm not suggesting it could, or would, happen, merely pointing out the premise of the concept as outlined by Roko as I felt the commenter above was missing that. As I said, it's not something I'd take seriously, it's just a thought experiment.

    masquenox ,

    Fair enough.

    LesserAbe ,

    Just because we don't have the ability now doesn't mean it's not possible. Consciousness isn't fully understood, but unless we want to introduce magical concepts like an immortal soul, our brains operate on cause and effect just like everything else.

    masquenox ,

    Just because we don’t have the ability now doesn’t mean it’s not possible.

    Yeah... no. It's about as likely as humanity "colonizing" space - it's not going to happen.

    Consciousness isn’t fully understood,

    True... and conflating consciousness with the trappings of digital technology is doing the exact opposite of getting us closer to any understanding of it.

    LesserAbe ,

    "yeah...no" isn't an argument.

    To be clear, I'm not saying the basilisk is a real concern, and I'm not saying we're anywhere close to being able to transfer consciousness. It could be a thousand years or a million years. But we don't have any basis to say it's impossible. It's not saying anything new to announce we can't do it currently. Obviously!

    (Also the book "A City on Mars" by Kelly and Zach Weinersmith does a great job addressing why trying colonize Mars right now is a bad idea. Which isn't to say it's impossible or we won't ever colonize it. Just that we need more research and capabilities before doing it)

    masquenox ,

    But we don’t have any basis to say it’s impossible.

    We have no basis to say it's possible, either - as I've stated before, this entire sci-fi trope is based on nothing more than techno-fetishists trying to conflate consciousness with information technology... and sci-fi tropes doesn't get more wonky than that.

    It could be a thousand years or a million years.

    Considering that we'll be lucky if we can maintain Victorian-era levels of industry by the end of this century, I'd say a fallacious belief in "progress" is rather inappropriate these days.

    Rhynoplaz ,

    I'm starting to suspect that masquenox is part of a propaganda campaign led by the basilisk itself! They just seem a little too serious about us not taking this seriously.

    Getting strong "Pay no attention to the man behind the curtain!" vibes.

    masquenox ,

    I’m starting to suspect that masquenox is part of a propaganda campaign led by the basilisk itself!

    We all have our price - it turns out mine is... dental cover.

    Carighan ,
    @Carighan@lemmy.world avatar

    In fact, many do. As soon as a minority group gains power, they tend to first execute the people who opposed them, and then start executing the people who didn’t stop the opposition.

    Yeah in fact, this is the big one. This is just an observation of how power struggles purge those who opposed the victors.

    Thorny_Insight ,

    First of all, the AI doesn’t exist in 2015, so people could just…not build it.

    I don't think that's an option. I can only think of two scenarios in which we don't create AGI:

    1. It can't be created.

    2. We destroy ourselves before we get to AGI

    Otherwise we will keep improving our technology and sooner or later we'll find ourselves in the precence of AGI. Even if every nation makes AI research illegal there's still going to be handful of nerds who continue the development in secret. It might take hundreds if not thousands of years but as long as we're taking steps in that direction we'll continue to get closer. I think it's inevitable.

    Cryophilia ,

    Sure, but that particular AI? The "eternal torment" AI? Why the fuck would we make that. Just don't make it.

    Thorny_Insight ,

    We don't. Humans are only needed to create AI that's at the bare minimum as good at creating new AIs as humans are. Once we create that then it can create a better version of itself and this better version will make an even better one and so on.

    This is exactly what the people worried about AI are worried about. We'll lose control of it.

    Cryophilia ,

    Yeah but that's not a Roko's Basilisk scenario. That's the singularity.

    Thorny_Insight ,

    Yeah but it answers the question "why would we create an AI like that". It might not be "us" who creates it. You just wanted a camp fire but created a forest fire instead.

    BobTheDestroyer ,

    Sci-Fi Author: In my book I invented the Torment Nexus as a cautionary tale

    Tech Company: At long last, we have created the Torment Nexus from classic sci-fi novel Don't Create The Torment Nexus

    Alex Blechman

    VindictiveJudge ,
    @VindictiveJudge@lemmy.world avatar

    People punish because we’re fucking animals, we feel jealousy and rage and bloodlust. An AI would not. It would do the cold calculations and see no potential benefit to harming anyone on that scale, at least not for those reasons.

    That's a hell of a lot of assumptions about the thought processes of a being that doesn't exist. For all we know, emotions could arise as emergent behavior from simple directives, similar to how our own emotions are byproducts of base instincts. Even if we design it to be emotionless, which seems unlikely given that we've been aiming for human-like AIs for a while now, we don't know that it would stay that way.

    Cryophilia ,

    Sure, but if you're taking that tack it could feel anything. We could build an AI for love and forgiveness and it decides it's more fun to be a psychopath. The scenario has to be constrained to a sane, logical AI.

    Scubus ,

    Point 1: this thing will definitely exist because we already see parallels to it

    Point 2: this thing won't exist because there's no reason for it to

    ???

    Cryophilia ,

    No.

    Point 1: if it did exist, it wouldn't be this novel thing, it already happens with humans

    Point 2: ...but it won't exist.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • til@lemmy.world
  • test
  • worldmews
  • mews
  • All magazines