r/SneerClub archives
newest
bestest
longest
51

Long time lurker who finally sat down to try and understand the dreaded acausal robot god. Not going to lie I did get anxious at first (probably due to my OCD), but after a while I realized the whole notion of acausal trade is absolutely ridiculous.

Am I missing something or is acausal trade just imagining some make believe creature in your head and then making a “deal” with the imaginary creature because somewhere in the (alleged) multiverse it has to exist?

edit: Apologies in advance if formatting is wrong or this is the wrong sub to post this in.

[deleted]

It is isn't it. I guess the fear of the thought experiment comes from people who don't necessarily understand it - not that it's their fault since the whole thing revolves around convoluted jargon imo.
The jargon is what makes the whole thing a thing. LW people love to write about nothing with big sounding words. It's like 90% of the scene.

Yes it is pretty silly, but it can make for some interesting science fiction. As cstross wrote in one of his books:

“I am the Eschaton. I am not your God. I am descended from you, and exist in your future. Thou shalt not violate causality within my historic light cone. Or else.”

E: The or else here involves things like pieces of rock traveling at near light speed and peoples homes.

In Rationalia, science fiction writes *you.*
This is from 2005 so, in proper Rationalia fashion, the good bits are not new. E: and I liked that book series, sucks you managed to write yourself in a corner Stross.
> E: and I liked that book series, sucks you managed to write yourself in a corner Stross. I still have no idea what exactly was the problem that caused Stross to abandon the series.
Same. Hope he tells us.
He wrote about it here: http://www.antipope.org/charlie/blog-static/2010/09/books-i-will-not-write-4-escha.html Basically the Unborn God is an AI on the level of the Eschaton which he didn't really consider.
Right thanks!
I mean, the difference is in those books Eschaton has actual time travel to whack you from the future if you do stuff it doesn't like. That's a whole lot less convoluted than the acausal trade thing.
Yep totally true. It is acausal, but not acausal trade, my bad.

In rationalism or neoreaction, the answer to “what the fuck” is always anime.

I didn’t understand this bit of the Basilisk until I read “Death Note”. Acausal trade is Yudkowsky describing the “I know that you know that I know” mental battle between Light and L.

See, donating to MIRI is spending years developing an immunity to acausal torture.
Yud: "i take the dust speck... AND BLOW IT!!"
Yudkowsky actually told me once that HPMOR was inspired in part by *Shinji and Warhammer 40K.* This was in a Twitter interaction long since lost; he randomly appeared and replied to a joke I made about writing *Shinji Ikari and the Methods of Rationality.* Of course, he replied to (and seemingly only read) the first tweet in a thread where I spelled out how being "rational" in the world of *Neon Genesis Evangelion* would just be another route to loneliness, despair and lots of screaming.
Yud really was watching/reading Death Note and thought to himself, "this....but with robots". But tbf, Death Note with robots does sound entertaining haha.
I liked the anime up to that part, that was just so convoluted and 'this is what a dumb person thinks a smart person is' it broke my enjoyment.
I think it helps if the viewer thinks of the creators of the anime as being in on the joke: Take the tennis scene. All of the 12th dimensional hypothetical double bluff head games is just *so* hyperbolically OTT for stakes so transparently low. Really, Sherlock Holmes' superpower isn't deduction. It's that the writer is on his side. Death Note is the story of two emo goth pretty boys who each have the writer on their side... until they don't, and its ridiculous and glorious.
> "I know that you know that I know" mental battle Reminds me of that puzzle Terry Tao wrote about in 2008 https://terrytao.wordpress.com/2008/02/05/the-blue-eyed-islanders-puzzle/
Acausal trade is Otherkin saying that they are reincarnations of cartoon characters. (Or at least Roko could have come up with an as convincing argument for that using the same ideas.)

[deleted]

>this is people just coming up with an idea and writing an essay about it who don’t even know the logical grounds of where they’re coming from. that applies to pretty much anything written at LW Edit: also, the idea of a SUPERINTELLIGENT AI using LW's theories is weird, to say the least.
You’re absolutely right, I’m just even more irritable than usual thanks to my current living and health situation, and getting sick of the fact that one of my habits is reading these venture capital programming twats pontificating about shit they don’t even have the barest understanding of and smelling their own farts every five minutes
I'm sorry things suck for you right now. I hope they get better.
Thanks for the kind reply
> Nobody who actually works in game-theory or decision theory (or both) would really entertain such a nonsense idea... > Both game-theory and decision-theory (...) are constructs which attempt to address real world situations. Too strong a statement. When, for example, Gale and Stewart analyzed infinite two-player games of perfect information in the 50s, they weren't addressing real world situations. (Who in the real world has time to play infinitely many innings of a game before seeing who wins?) Certainly a lot of stuff in game theory and decision theory attempt to address real world situations—I'm maybe being too generous to economists here ;)—but to say they are just about the real world is to reduce them to just the most boring bits. In any case, there's plenty of obscene or nonsensical game theory stuff that nevertheless got taken seriously, e.g. von Neumann's contributions to the Cold War.
That’s a fair point, I was more addressing the stuff you put in at the bottom there
There's also some very obviously practical situations for decision theory which actually can pop up but miraculously seem to get resolved correctly without much analysis. Lets suppose we are making an autopilot. For redundancy, we have 3 computers running identical software, the outputs go into a very reliable majority vote circuit, the result controls the airplane. This is curiously problematic for "causal decision theory" as originally formulated. If you consider possible actions by one computer in that autopilot, they have no influence on the flight (it gets outvoted). Of course, in practice, anyone who writes software for that autopilot treats this whole 3 computer situation as an irrelevant implementation detail, and everything works perfectly fine. Alternatively, the outputs of computers are extremely highly correlated, so evidential decision theory works just fine (and it is not clear what the "smoking lesion" would look like for an autopilot). edit: TBH my take from this is that the evidential decision theory is probably the correct one; as far as "smoking lesion" goes, it is very unconvincing; if there was sufficient evidence to establish the lesion theory then the decision to smoke would no longer correlate with cancer (with the lesion forcing the hand, bypassing the decision). Of course, in real life, tobacco companies just made it up, without even faking enough evidence to support the idea. Various folk philosopher attempts to "fix" causality (see LessWrong) fall prey to the case of 3 redundant autopilots made by different developers, and ignorant of each other inner workings with the exception of a knowledge of a correlation. The autopilot makes a good example because there is no reasonable disagreement as to what the correct behavior is; we all want reliable airplanes and we want to get to the destination. Clever tricks like "each computer should act as if there's a failure and its actions matter" don't work either if in case of failure the flight must divert to the nearest airport. Real airplanes have all sorts of complex redundancy schemes with multiple methods of calculation and sanity checks (e.g. Airbus fly-by-wire software is pretty crazy), all of it engineered very cleverly to prevent causal chains from a bit flip to a plane crash.
> Concievability issues: within modal logic (which is another aspect of the whole thing) the standard view of possible worlds is that there are near and far possible worlds, and that we should be concerned with the nearer ones rather than the further. The acausal trade stuff doesn’t take account of this at all. https://philpapers.org/archive/sinpg
Classic Neil and I are actually friends on twitter btw

fuck you, it’s the best idea that has ever come up, you miserable little piece of shit.

o fuq! It's all fun and games until the real acausalrobotgod shows up. Please don't simulate me, I'm too stupid to help you anyways
Well, I think you need the most [acausal attention](https://www.youtube.com/watch?v=3lRbwPKbveM).
Pounded in the butt by the butt-pounding tulpa I created specifically so I wouldn't be pounded in the butt.

It’s gnosticism for tech bros. The evil demiurge (err I mean AI) has trapped us in this material universe.

Their “proofs” are about as convincing as any other religious cosmology.

The fun thing is that even if it wasn’t dumb enough by itself, if we’re dumb enough not to get it it becomes dumb because the bargaining becomes impossible.

LessWrong has invented a hypothetical threat that can be defeated by saying, “That’s stupid.”

It invented a hypothetical threat that can be defeated by sending money to MIRI.
I would simply not be tortured.
Or, just become a painslut, and like being tortured. Bdsm is praxis.
lament.cfg
[Such](https://www.youtube.com/watch?v=lKfupO4ZzPs) sights. Much wow. E: Hellraiser, as envisioned by Barker is actually a lot more interesting than the hellraiser 3 and later movies, in which they started to talk about the otherworld as an actual hell. In the first stories the otherworld was actually a world created by a [large geometric shape](https://cenobite.fandom.com/wiki/Leviathan) who only wanted to create a realm of perfect order and wanted to prevent chaos. And to do that it created the cenobites (who are summoned not by just solving the puzzlebox, but solving the puzzlebox while having a manic desire to solve it, the desire part is important somehow) who murdered people (as life is chaos) and created more cenobites to torture forever, as well, at least eternal pain isn't chaotic. It was way more interesting vision than 'christian hell take 404'. If you can dig up the old hellraiser comics they can be an interesting read.
Their founder is held hostage by terrors he himself invented, it's pretty apt.

(I go into some detail about acausal decision theory here, maybe I should start with a WARNING, even on /r/sneerclub)

According to the official history, it all started with an attempt to beat the Prisoner’s Dilemma, a classic scenario of game theory. Ordinary self-interest says you should “defect”, but if only both players could “cooperate”, then they could both have a higher payoff. Douglas Hofstadter dubbed this “irrational” decision to cooperate superrationality.

Then someone called Gary Drescher published a book justifying superrationality, in the case where the two players are computer programs running the same source code. If you’re in a Prisoner’s Dilemma with an exact copy of yourself, then you might reason that you can rationally choose to cooperate, since your other self will do the same thing. But you don’t actually control your other self, so what is the exact justification for this confidence? Drescher apparently introduced the word “acausal” to the discussion.

Drescher, incidentally, is an enigmatic figure to me. There’s very little information about him. His book “Good and Real” came out in 2006, the year in which the group blog “Overcoming Bias” was also launched (“Less Wrong” emerged from “Overcoming Bias” three years later). I have not read the book, but it contains acausal decision theory and a defense of the many-worlds interpretation.

Another “application” of this thinking is to resolve Newcomb’s paradox, which is a little like being in a Prisoner’s Dilemma, not with a copy of yourself, but with a superintelligence which will defect if you defect, and cooperate if you cooperate. I won’t go into the details, but you have a choice between being greedy and being restrained, and the superintelligence has promised you a big reward if you are restrained and a small reward if you are greedy. And the paradox is that the superintelligence already predicted your choice and determined the size of the reward. Ordinary causal thinking then says, you may as well be greedy and grab everything, because the reward is already set; but if you do that, you will be retro-causing yourself to have a small reward. How do you rationally justify being restrained?

The answer is an extended version of Drescher superrationality. The superintelligence is not a copy of you, but its decision is a copy of your decision. You should be restrained now, because that implies the superintelligence will have modeled you as restrained, and left the big reward. This violates the usual dictum that the future cannot influence the past, so, “acausal”. Perhaps it would have been better to speak of logical causality or atemporal causality, but this is the dominant temrinology now.

In any case, there has been a progressive generalization of the concept, to agents that are only vaguely similar, agents located in different universes, populations of agents in different universes reasoning their way to a collective equilibrium, and even quasi-theologies like, all possible godlike superintelligences that rule their respective universes, arriving at an acausal decision equilibrium among themselves.

Recounting this history leads me to think that, from the perspective of history of ideas, Roko’s basilisk should be regarded as an episode in the history of decision theory, that can usefully be placed alongside Newcomb’s paradox and Hofstadter’s approach to the Prisoner’s Dilemma. Its notoriety as an object of fear or derision obscures the fact that it is also a thought experiment for decision theorists.

People get tripped up on the fact that it’s a scenario in which there is a causal link as well as an acausal link - the possible AI is in our future, that’s the causal link - and the acausal part is overlooked in favor of the grand guignol of “punishment by the robot god”. But viewed abstractly, it’s just a variation on Newcomb’s paradox, but with the superintelligence that models your decision in the future, rather than in the past.

Returning to the post above: is “acausal trade” really just a kind of daydream, an exercise in “trading” with imaginary friends and enemies? Certainly the multiverse version seems problematic on multiple levels. We don’t even know that other universes exist, so how can we know that we’re trading with them? And even if they do exist, I would question whether any relationship possessing the mutuality implied by the designation of trade, is actually possible among them. If Chuang-tzu does something for the sake of the butterfly, and the butterfly does something for the sake of Chuang-tzu, is that a “trade”, or just a fortuitously consistent folie a deux?

At a mundane level, a sneer may be enough to chase away the basilisk doldrums. But I suppose the idea needs a more formal way of being countered too. So let me propose a hypothesis of omniversal autarky: That the set of superintelligences which attempt acausal trade has measure zero, because it quickly becomes clear to a hyper-rational being that you should only care about things you can influence causally. I can’t prove it, but neither can the philosophes of acausality disprove it. Let them try to do so, and until they do so, feel free to focus on this universe alone.

Thank you for the detailed write-up. I def appreciated the detail you went into, helped me to understand this whole scenario better. Side note: when I was first reading about RB I came across your post on wordpress (A compilation of comments) about why Acausal Blackmail was BS. Reading through that alleviated a ton of the anxiety I had surrounding this thing. So thank you for that as well :)
Yudkowsky liked Drescher's book (which is I think his Ph.D thesis cleaned up and published), and [called it](https://www.lesswrong.com/posts/wfJebLTPGYaK3Gr8W/recommended-reading-for-new-rationalists?commentId=2B5NKt6j76vcf72Rv) "practically Less Wrong in book form." Drescher also [commented a bit](https://www.lesswrong.com/users/gary_drescher) on LW, but didn't stick around.
Don't you think that whether CDT fails in newcomb's paradox is still not so sure kinda debate. I mean , even if a computer simulated me, somewhere in his timeline, newcomb's paradox is a random event. Anytime even I can change my own decision. Leave my copy, if I do this game twice in different timeline, let's say on consecutive days. And one day I choose two box and on other day I choose one box. How will my simulation correctly set the order that on what day I choose what. I mean isn't the choice always random? And so how CDT fails?

isn’t part of the idea that if the robot god can simulate you, you’re not sure if you’re currently the simulation created by them?

If that's the case, then if the AI wants to send me a message or tell me to do something, it could appear before me as an avatar wreath'd in flame and speak with a voice that shakes mountains. That would get my attention a lot better than simulating a tenuously logical theory on the internet.
but then it would confirm you're in the simulation, and it gets nothing out of compelling you to do stuff in the simulation the more i think about it, the less sense this makes
Or it would confirm to me that God is real and so is the world. Also even if I were sure I was in a simulation, I would still feel physical pain. So if the AI wanted to torture me, me knowing that I'm in a simulation wouldn't stop me from suffering.
You could also try to break out, or break the simulation, a thing which happens when people try to simulate evolution (or unexpected glitches happen). That this doesnt seem to happen irl is a big strike against the simulation theory. (And yes you can counter this counter by going 'but what if the simulation is perfect', but that is just inventing god again)
Honestly, throwing me some shiny 1s & 0s at the screen at startup would be enough.
True, but even if our world is being simulated by RB that doesn't mean we are being acausally blackmailed. It's just searching for other "trading" partners - which we are not since we can't acausally trade with a super AI. At least that's how I understand it.
yeah ill be honest i have no fucking idea
That makes the two of us lmao
I think being simulated and trying to be acausually blackmailed are two mutually exclusive different things. If we are being ablackmailed it would imply not being simulated and vice versa.

It’s basically Anselm’s Ontological Argument but for robots.

I mean, it makes sense if you accept the assumptions (namely, possibility of easily simulating consciousness, AI that behaves in a certain way)

Correct me if I'm wrong but even after taking the assumption that you can simulate consciousness and that an AI will be god one day, isn't it still impossible for humans to actually model any entity in their head well enough (especially a super AI) to trade with it - making acausal trade impossible and just imaginary?
I know it, you know it, Yudkowsky doesn't know it because he's too good for such boring, mundane things as complexity theory. The guy is positively obsessed with the idea of simulating people, so it _has_ to be possible in his world because a lot of things are indistinguishable from magic if you don't go to college.
I liked the way someone on r/badmathematics explained it: "once true ai arrives, we can assume it will enhance itself exponentially. So we can assume it can run algorithms requiring arbitrary resources. So we can assume it can run algorithms requiring infinite resources. So we can assume future ai can reverse entropy and simulate the entire universe. QED, all perfectly logical, no jumps of reasoning anywhere."
That's actually Yud's own refutation of the idea. According to him, he got mad at Roko because he presented the idea of humans making deals with powerful AIs from the future, although the exact way Roko proposed doesn't work for humans because it's impossible to simulate any AI god in your human mind. His concern is about someone managing to come up with a way to make human-future AI deals that actually work, now that the idea is out there.
Which is weird btw, as his (yud) own theory also says that godAI's can figure out a lot of things for very limited data (see his theory about 2 frames of an apple falling is all that is needed for a boxed AGI to make a model of our physics). Which means just that somebody thought of this from our human data means that an AGI can do it as well, from the same data. [Aka, the Basilisk is inevitable.](https://www.youtube.com/watch?v=7HypZPRXWE0)
My man was seriously upset about the theoretical possibility of humans making mind deals with god like AIs from the future. This is some HP Lovecraft type shit, read the forbidden tome that teaches you what words to think to make deals with cosmic entities
“Do not call up that which you cannot put down.” H P Lovecraft, The Case of Charles Dexter Ward. Turns out this doesn't just apply to demon summoning [E:Sorry resurrecting the dead, don't want to be accused of misrepresenting Charles, unrelated but look at [this weird tales cover from wikipedia](https://upload.wikimedia.org/wikipedia/commons/c/c3/Weird_Tales_July_1941.jpg)], but also to just thinking very hard.
That's why you have the Safety Cut Revenant Axe Man. That cover owns in multiple ways.
Why is the robot saying he killed Hitler but then grinning like he did something bad.
>see his theory about 2 frames of an apple falling is all that is needed for a boxed AGI to make a model of our physics That has to be one of the stupidest ideas I've ever heard of. Given the self-evident fact that there are a theoretically infinite variety of possible sets of physical laws which would lead to an apple falling at that speed, the falling apple cannot be used as evidence to favor one model over another. Maybe we're in a universe in which Newton's laws are absolutely correct, maybe we're bound by relativity, maybe the earth exerts a pressure that causes apples to accelerate *away* from it at a rate of 9.8 m/s^(2) and the camera was upside down. Without further information, it's *literally impossible* to know the correct answer. It's like saying a smart enough AI could crack the encryption of a one-time pad, or correctly guess my date of birth given no information besides the fact that I am currently alive.
One frame is enough if you just look at how the grass bends. E: [not even a joke, that is what he actually said](https://www.lesswrong.com/posts/5wMcKNAwB6X4mp9og/that-alien-message)
I clicked through, and, like every time I accidentally visit that site, immediately thought "lifes too short to read all that".
Idk about the deal part, the way I understand roko's basilisk is that a hypothetical future AI would torture simulated copies of you to motivate you to work towards building the AI. The reason Yud suppressed info about the basilisk and called it an infohazard is because the logic only works if you know about it, the AI wouldn't torture a random person since that wouldn't motivate them in the present, it would only torture those who were worried about being tortured for not working hard enough to build the AI. You're probably thinking about "the AI in the box boxes you" which has you precommit to not releasing an AI onto the internet so the AI can't threaten you with torture of simulated copies, since the AI would know (through simulating you) that you already committed to not releasing it so torture would just be a waste of resources. It doesn't require you to model an AI in your head as far as I know though.
I’ve genuinely never understood why I should care about a simulated version of me being tortured. It takes some serious and not entirely credible leaps of faith to believe that that is both possible, and ethically equivalent to your current and present self.
Also, the bastards totally deserve it.
Past human me barely seems to give a shit about future human me, presumably I might be completely hostile to hypothetical future simulated me.
Past human me fucking sucks and future human me can definitely deal with the stuff I'm procrastinating on, I have faith in you, future human me!
Reminds me of one of [my favorite Calvin & Hobbes stories](https://web.mit.edu/manoli/mood/www/calvin-full.html). Actually, come to think of it, [the Calvin duplicator strip](https://www.gocomics.com/calvinandhobbes/1990/01/11) is probably a good summation of how a normal person would react if being told that they needed to work on developing AI or else some alternate version of themselves would be tortured.
Yeah there's also some philosophical assumptions about personal identity in there as well
In addition to this, a super-powerful A.I. should know that torturing simulations isn't time-travel. Torturing a simulation of you int he future is nothing but a literal waste of resources. You can't even make it change peoples behavior, and anything it could accomplish (threaten people in the future now) could also be accomplished by just claiming you're infini-torturing all those people that didn't donate to MIRI.
Yeah, you can easily apply the "illogic of hell" argument to RB too. Punishing after death accomplishes NOTHING, and it's only retributive. We, in 2021, have no way of verifying if it will carry out the punishment or not. In other words, people who take RB seriously had their behavior altered by the *potential* to be punished (since we have no way of verifying the punishment with a time machine). The threat drove behavior, not the torture. Theres literally no incentive for it to carry out any punishment. Any superAI that doesn't understand this isn't that super.
So I think there are two models here. In the 'altruistic' one, the idea is that a simulation of you is basically your sibling, and certainly a person, and this is basically a hostage situation. The other thing is more like a Pascal's Roulette. A figure approaches you, and it says "Lo, Abraham, the God of Torture and Ruination has sent 600 angels and one normal man to you and your 600 twin brothers. We all carry the one commandment, that you should sacrifice Isaac. If anyone of you should refuse the command of one of the 600 angels, the God of T&R shall level your entire city, and commit everyone therein to agonizing and endless torture. If the one of you approached by a mere man is **not** rebuked, then the God of T&R shall level your entire city, and yada yada yada. So, how do you like your odds that I'm the one normal man, as opposed to one of the 600 angels?"
The trick is that once you say "I don't care about simulated versions of me." the acausal robot god reveals that you're a simulation. At which point instead of shouting "Death to the Demiurge!" or something similar you're supposed to be cowed. Yud has a fucked up understanding of what you're supposed to do if it turns out there's a malevolent Creator imo.

not only is it mind bogglingly stupid, but some days i think it barely makes it into the the top five stupid things the lesswrong crew has put together

If you want the silly and yet much more thought out version of this, read Neil Sinhababu’s Possible Girls. It will either convince you that stuff across universes is silly, or that lesswrong just has it , well, wrong, and they are actually more wrong.

oh, i fucking love that paper
what did i just read haha
An actual philosophy paper about acausal interactions between worlds :D

Uhm, guys? Isn’t acausal trade just living in a society explained so obtusely as to be illegible? Because when I think about acausal cooperation, I think about me and all the other people who pick up someone else’s trash in the park, who will never meet each other but are nevertheless helping each other across time.

Part of acausal trade is accepting as something that’s true that you can easily be recreated. This is really the keystone of Yud’s thought because of his thanatophobia and conviction that he’s worthy of being recreated by the acausal robot god (ARG) at the end of time. If you don’t accept that the you that the ARG creates (nb that I’m just accepting as a given ARG’s existence because there’s no real reason to stretch this out even longer not because I actually think ARG is real) is the same as the you who is reading this then the bargain doesn’t work.

If I understand it correctly, the general idea of acausal trade (at least in a limited form) makes sense but isn’t useful for much of anything; on the other hand, the application of it in Roko’s Basilisk just doesn’t make sense.

The basic idea of acausal trade can be explained by the following thought experiment: Suppose you’re playing Prisoner’s Dilemma against a copy of yourself, or against someone you know will make the same decision as you. (Obviously this is unrealistic.) Because of this, you know that the only possibilities are both agents cooperating or both defecting, and your opponent will cooperate iff you cooperate. Since both cooperating has a higher payoff than both defecting, you should choose to cooperate, which logically necessitates that your opponent cooperate. The same applies in any Prisoner’s-Dilemma-like situation where the two agents can predict each other’s actions precisely and can decide to cooperate iff the other agent does (and are aware of this option).

On the other hand, the requirement that each agent be able to nearly perfectly predict the other’s action makes this fairly useless outside of contrived thought experiments. The only case I know of of any remotely plausible attempt to apply it is George Ainslie’s theory of willpower as intertemporal bargaining, which uses a similar concept as part of its theory that willpower is partly just people knowing that they will probably act the same way now as in the future and therefore that to avoid indefinite procrastination in the future they must act in the present as well, though that was developed independently of Yudkowsky & al. The application of it to future superintelligent AIs doesn’t make sense AFAICT because it would almost certainly be impossible either for an AI, however smart, to learn enough about a person who had died any significant amount of time ago to accurately simulate them, or for a modern person to precisely predict such an AI’s actions, so the basic preconditions are not met. (The other problem with Roko’s Basilisk is that the acausal trade requires that the AI predict that you will cooperate with it iff it doesn’t torture you, so it can be defeated simply by firmly deciding to ignore it and not cooperate with it no matter what.)