
Title text: I'm working to bring about a superintelligent AI that will eternally torment everyone who failed to make fun of the Roko's Basilisk people.
nb: Roko's basilisk. It's really stupid.
added link.
Moderators: Moderators General, Prelates, Magistrates


eSOANEM wrote:If Fonzie's on the order of 100 zeptokelvin, I think he has bigger problems than difracting through doors.
xkcdfan wrote:nb: Roko's basilisk. It's really stupid.
ilduri wrote:Sooo... Roko's Basilisk is basically Pascal's Wager repackaged for Singularitarians?
"You'd better fund the creation of a super-intelligent AI, because if you don't, it'll create a simulation of hell and put a simulation of you in it."
It seems to me that if a super-intelligent AI might send people to hell, the only moral thing to do is to not fund its creation.
EliezerYudkowsky wrote:I can't seem to post a reply because it repeatedly gets flagged as spam. I can't post a link to discussion elsewhere because that also gets flagged as spam. Does somebody know how to correct this?
RowanE wrote:On the one hand, I'm really glad to see my in-group mentioned somewhere as popular as xkcd. On the other hand, I don't like that Roko's Basilisk is still the first thing anyone ever hears about the LessWrong community, FFS that was one time! The only reason it wasn't forgotten instantly was that the original post got deleted and there was a kerfuffle over whether something so implausible needed to be censored. Apparently some people took it just seriously enough that they were starting to worry about AI-hellfire (but not enough to actually start trying to bring about the apocalypse), and that's why it was deleted - and also why it's called a "basilisk", i.e. it's a harmless bunch of words that can traumatise you (if you're so disposed) - literally no-one has ever been recorded actually having been seriously convinced to do what the basilisk says. There are no "Roko's Basilisk people"... unless you're talking about the LessWrong community more generally, which would be a very hurtful thing.
EliezerYudkowsky wrote:Tl;dr a band of internet trolls that runs or took over RationalWiki made up around 90% of the Roko's Basilisk thing; the RationalWiki lies were repeated by bad Slate reporters who were interested in smearing particular political targets; and you should've been more skeptical when a group of non-mathy Internet trolls claimed that someone else known to be into math believed something that seemed so blatantly wrong to you, and invited you to join in on having a good sneer at them. (Randall Monroe, I am casting a slightly disapproving eye in your direction but I understand you might not have had other info sources. I'd post the link or the text of the link, but I can't seem to do so.)
Flammifer wrote:Fellow Lesswronger here, I was about to post that joke
(It's also not clear for me whether I fall under "the Roko's Basilisk people", if so, yay! I've been mentioned on XKCD! (and all you people have to make fun of me or face ETERNAL TORMENT!)

Arancaytar wrote:For any decision X you could make in fear of an AI punishing you, you should also fear an AI punishing you for the opposite decision. Once you accept that this is pointless, you become immune to any AI's blackmail. If some version of you is going to end up in some version of hell regardless of what you do, you may as well follow your own conscience in all decisions.
EliezerYudkowsky wrote:Diadem, follow the link. http://www.reddit.com/r/Futurology/comm ... sk/cjjbqv1
Telling me that I ought to be happy about how this has been portrayed... seems a bit... (searches for words) unrealistic.
Diadem wrote:It's a wiki. Anyone can edit. Where exactly are you basing the claim that this is a systematic attack on the LessWrong community on? Looking at the article's history I see no pattern of corrections being systematically reverted, or anything like that.
Code: Select all
$ wget "rationalwiki.org/w/index.php?title=LessWrong&offset=&limit=2500&action=history" -q -O- |grep userlink |sed -e "s@.*userlink\"[^>]*>@@" -e "s@<.*@@" |sort |uniq -c |sort -n |tail -11
6 Tetronian
7 80.221.17.204
12 XiXiDu
13 Waitingforgodel
14 Stabby the Misanthrope
17 AD
23 Bo
28 Armondikov
30 Human
49 Baloney Detection
301 David Gerard
Diadem wrote:I'm curious why you call the RationalWiki article 'made up by trolls'. At first glance it seems fairly correct, and it casts you personally in a pretty favorable light. So where's the problem?
Code: Select all
rationalwiki.org/w/index.php?title=LessWrong&oldid=1059400As such, "You should try reading the sequences" is LessWrong for "fuck you."
The site has also been criticized for practically being a personality cult of Eliezer Yudkowsky. This is almost certainly not intentional on his part, just ask Brian of Nazareth.
Ironically, Less Wrong users rarely recognize biases that arise from the site's demographics[19], which can be summarized as the same problem in academic psychology of samples being WEIRD: mostly male, white, white-collar, 20-30-year-old United States residents coming from families with a Christian or Jewish background. When pointed out the sources and instances of collective bias, they typically ignore them or say that "this is just how things are here."
Indeed, if anyone even hints at trying to claim to be a "rationalist" but doesn't write exactly what is expected, they're likely to be treated with contempt, as criticism of
You'll be unsurprised to know that many in the LessWrong community self-diagnose themselves as being on the Asperger's/autism spectrum.[43] They do all this because they are bad at human interaction
Diadem wrote:EliezerYudkowsky wrote:Diadem, follow the link. http://www.reddit.com/r/Futurology/comm ... sk/cjjbqv1
I have a lot of respect for you as an author and researcher, but that reddit article has to be one of your poorer ones. Mudslinging is always a dangerous activity. Even if you are entirely justified, even if you are entirely in the right, what people will see is still you covered in mud. I think you'd do much better just telling your side of the story, with as much proof as you can. There's no need to even mention the other side exists.
Arancaytar wrote:For any decision X you could make in fear of an AI punishing you, you should also fear an AI punishing you for the opposite decision. Once you accept that this is pointless, you become immune to any AI's blackmail. If some version of you is going to end up in some version of hell regardless of what you do, you may as well follow your own conscience in all decisions
RowanE wrote:And the real solution is simpler anyway - "we do not negotiate with terrorists". If you stubbornly refuse to be blackmailed, anyone who knows this will know it's not worth the effort to try. Doesn't work for when you're likely to chicken out and acquiesce in actual blackmail situations, but this problem isn't really there in acausal trade.
RowanE wrote:And the real solution is simpler anyway - "we do not negotiate with terrorists". If you stubbornly refuse to be blackmailed, anyone who knows this will know it's not worth the effort to try. Doesn't work for when you're likely to chicken out and acquiesce in actual blackmail situations, but this problem isn't really there in acausal trade.
FeepingCreature wrote:RowanE wrote:And the real solution is simpler anyway - "we do not negotiate with terrorists". If you stubbornly refuse to be blackmailed, anyone who knows this will know it's not worth the effort to try. Doesn't work for when you're likely to chicken out and acquiesce in actual blackmail situations, but this problem isn't really there in acausal trade.
Of course, you still have to worry about other people folding and the AI coming to exist anyways, and it then decides that while you weren't, in this case, motivated by the threat of blackmail you plausibly could have been...
RowanE wrote:And the real solution is simpler anyway - "we do not negotiate with terrorists". If you stubbornly refuse to be blackmailed, anyone who knows this will know it's not worth the effort to try. Doesn't work for when you're likely to chicken out and acquiesce in actual blackmail situations, but this problem isn't really there in acausal trade.
FeepingCreature wrote:I'll let this histogram of contributors to the RationalWiki LessWrong page speak for itself.
Code: Select all
$ (\
USERS=$(wget "rationalwiki.org/w/index.php?title=LessWrong&offset=&limit=2500&action=history" -q -O- |grep userlink |sed -e "s@.*userlink\"[^>]*>@@" -e "s@<.*@@");\
echo "<html><body style=\"width: 50%; margin-left: 50px; \">";\
echo "$USERS" |while read LINE;\
do echo -n '<div style="float: left; background-color: #'; echo -n $(echo "$LINE" |md5sum |cut -c1-6); echo -n '; "> </div>';\
done;\
echo " <br><p style=\"clear:left;\"></p>";\
echo "$USERS" |sort |uniq -c |sort -nr |while read LINE;\
do echo -n '<div style="display: inline-block; background-color: #'; echo -n $(echo "$LINE" |sed -e "s/^[0-9]* //" |md5sum |cut -c1-6); echo -n '; "> </div>'; echo "$LINE<br>";\
done;\
echo "</body></html>"\
) > contrib.html

"Well, yes," Harry said. He was surprised that he wasn't feeling angrier at Captain Weasley, but his concern for Hermione seemed to be overriding that, for now. "The more you try to justify yourself to people like that, the more it acknowledges that they have the right to question you. It shows you think they get to be your inquisitor, and once you grant someone that sort of power over you, they just push more and more." This was one of Draco Malfoy's lessons which Harry had thought was actually pretty smart: people who tried to defend themselves got questioned over every little point and could never satisfy their interrogators; but if you made it clear from the start that you were a celebrity and above social conventions, people's minds wouldn't bother tracking most violations. "That's why when Ron came over to me as I was sitting down at the Ravenclaw table, and told me to stay away from you, I held my hand out over the floor and said, 'You see how high I'm holding my hand? Your intelligence has to be at least this high to talk to me.' Then he accused me of, quote, sucking you into the darkness, unquote, so I pursed my lips and went schluuuuurp, and after that his mouth was still making those talking noises so I put up a Quieting Charm. I don't think he'll be trying his lectures on me again."
Polemarchus said to me: I perceive, Socrates, that you and your companion are already on your way to the city.
You are not far wrong, I said.
But do you see, he rejoined, how many we are?
Of course.
And are you stronger than all these? for if not, you will have to remain where you are.
May there not be the alternative, I said, that we may persuade you to let us go?
But can you persuade us, if we refuse to listen to you? he said.
Certainly not, replied Glaucon.
Then we are not going to listen; of that you may be assured.
--Plato, The Republic
IlyaShpitser wrote:(a) The original thing was an overreaction,
(b) It is a sensible social norm to remove triggering stimuli, and Roko's basilisk was an anxiety trigger for some people,
(c) In fact, there is an entire area of decision theory involving counterfactual copies, blackmail, etc. behind the thought experiment, just as there is quantum mechanics behind Schrodinger's cat. Once you are done sniggering about those weirdos with a half-alive half-dead cat, you might want to look into serious work done there.
"I thought you said you could just read his brain electronically,'' protested Ford.
"Oh yes,'' said Frankie, "but we'd have to get it out first. It's got to be prepared.''
"Treated,'' said Benji.
"Diced.''
"Thank you,'' shouted Arthur, tipping up his chair and backing away from the table in horror.
"It could always be replaced,'' said Benji reasonably, "if you think it's important.''
"Yes, an electronic brain,'' said Frankie, "a simple one would suffice.''
"A simple one!'' wailed Arthur.
"Yeah,'' said Zaphod with a sudden evil grin, "you'd just have to program it to say What? and I don't understand and Where's the tea? --- who'd know the difference?''
"What?'' cried Arthur, backing away still further.
"See what I mean?'' said Zaphod and howled with pain because of something that Trillian did at that moment.
"I'd notice the difference,'' said Arthur.
"No you wouldn't,'' said Frankie mouse, "you'd be programmed not to.''
EliezerYudkowsky wrote:I can't post a link to discussion elsewhere because that gets flagged as spam. Does somebody know how to correct this? Tl;dr a band of internet trolls that runs or took over RationalWiki made up around 90% of the Roko's Basilisk thing; the RationalWiki lies were repeated by bad Slate reporters who were interested in smearing particular political targets; and you should've been more skeptical when a group of non-mathy Internet trolls claimed that someone else known to be into math believed something that seemed so blatantly wrong to you, and invited you to join in on having a good sneer at them. (Randall Monroe, I am casting a slightly disapproving eye in your direction but I understand you might not have had other info sources. I'd post the link or the text of the link, but I can't seem to do so.)
So far as I know, literally nobody has ever said, "You should build this AI because it'll torture you if you don't." Like, literally nobody. There are people who want you to believe somebody else says that, but there's literally nobody who does. Even the original "Roko" was claiming that Friendly AI was a terrible idea because it would torture people who didn't contribute to building it, and was using that to argue that nobody should ever try to build Friendly AI.
I can't say the thing is made up out of entirely thin air because there is, in fact, a corresponding question in Newcomblike decision problems (whose corresponding answer appears to me to be "no"), and one branch of Newcomblike decision theory was worked on by collaborators who shared posts at LessWrong.com (which is why the 10% actual fiasco happened there). Needless to say, nobody in the "Ha ha let's sneer at these nerds" section has ever, ever succeeded in understanding any of the technical work that was twisted to make up this thing, like http://arxiv.org/abs/1401.5577 which is our work proving cooperation in the Prisoner's Dilemma between agents that have common knowledge of each others' source code. Eventually, I expect, we'll prove a no-blackmail equilibrium between updateless agents with common knowledge of each others' source code... and nothing will change on the Internet, because bad Slate reporters are incapable of understanding that and wouldn't care if they did.
Return to “Individual XKCD Comic Threads”
Users browsing this forum: kalira and 16 guests