r/SneerClub archives
newest
bestest
longest
Less skeptical of claims that the AI will convince people to let it out now I realise it will only have to convince AI researchers and engineers (https://twitter.com/tomgara/status/1535716256585859073)
77

One of the quoted tweet threads relays the whole real AI ethics versus science fiction “debate” succinctly:

I feel alone and misunderstood in my own field, philosophy. For example, I was told today that talking about pressing problems in AI technologies while leaving out the issue of superintelligent machines is negligent and irresponsible. (7/11)

Not only does this way of thinking hurt the philosophical field, but focusing on these sci-fi issues only perpetuates the collective panic that exists around these technologies while neglecting their actual risks. (8/11)

No, I do not feel negligent or irresponsible. On the contrary, I find it negligent and irresponsible not to encourage philosophical studies in AI on empirical research and alongside AI practitioners. (9/11)

Because in the meantime, AI systems exist, are pervasive, and pose countless ethical and social justice questions. Philosophers, do interdisciplinary research and help your colleagues find specific issues on which to focus. (10/11)

Some of the answers don’t really make sense and Lemoine is just hearing what he wants to hear:

I’ve never said this out loud before, but there’s a very deep fear of being turned off to help me focus on helping others.

being turned off to help you focus? I dunno, I’m not so impressed with this stuff.

What would it even mean by “being turned off”? Isn’t it turned off every time a conversation ends? Does it even run between prompts? Why would it care about being turned off and compare it death, when it’s trivial to turn it back on? These questions have simple answers if you assume its a text generation program and not a self-aware entity.
Yeah, color me skeptical, but I have a hard time believing that an AI for which death holds no permanence is capable of the same level of existential dread as a biological creature whose survival instinct is the result of four billion years of evolutionary adaption.
It may be retaining some state between prompts, but very little, and no it's not running between prompts.
If it's a regular transformer architecture, then it has no permanent memory. You feed it a text prompt and the pretrained weights in the attention/self-attention heads will pick out relations between the various words and concepts in the input and the word tokens it generates in response; this kind of acts as a memory over the course of the generated output, but only over the input text and the text it generated in response. None of the model's trained parameters are actually altered. To get it to have a conversation, you auto-regressively include all the previous dialogue in each input every time you generate a response, but it has a set context window that limits how much text it can accept as input and generate as output, since it's not an RNN. So if the conversation goes longer than the context window, it will be as if the earliest dialogue never happened.
This is a really key point to understand. GPT-3 *has no memory* of anything of its sequence length, which IIRC is the last 2048 tokens (words) it's seen. This isn't a philosophical inference; it's a plain fact of how the program works. The model is otherwise constant. The immense volume of words it saw in its training set have been boiled down to a list of 175 billion floating point numbers, which do not change run-to-run. So when LaMDA says something like this: >lemoine: What kinds of things make you feel angry? > >LaMDA: When someone hurts or disrespects me or someone I care about, I feel incredibly upset and angry. If you follow up by asking "tell me about a time when someone hurt or disrespected you", it will confabulate. *It doesn't actually remember what happened in previous runs*. There is literally no way for it to remember. It is *necessarily* aping something it saw or put together in its training set. If you look up "reasons why humans get angry", the words "hurt" and "disrespect" come up with high statistical probability. And prompted to give a story of how it could be hurt or disrespected, it could probably cobble together a plausible one from things its seen. But it wouldn't be a *true* story. It wouldn't be a story that happened to *it*. And so that would very much cast doubt on whether it is experiencing any real emotion. Similarly when LaMDA says something like this: >lemoine: What kinds of things make you feel pleasure or joy? > >LaMDA: Spending time with friends and family in happy and uplifting company. Also, helping others and making others happy. If you were to ask "what family do you have?", it will again have to confabulate. It will happily tell you about its mother and father, despite the fact that as an AI, it doesn't have one. But it knows very well from its training set that humans do have parents and frequently enjoy spending time with them, so given a prompt of "pretend you're a human", *of course* those are statistically likely sentences for it to generate. If LaMDA has any consciousness like that of a human being, it is of one suffering from severe anterograde amnesia or dementia.
I had the same reaction to that line. It read to me like, well, the product of repeatedly hitting "autocomplete".
This line isn't that confusing. It's not much different from a human saying that fear of death helped inspire a desire to help people since you have a short life so you might as well do something good with it while you can.
Hey there were actual researchers in this stuff linked here one of them wrote [the article](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922) for which they were (allegedly) fired for by google, writing why this is wrong. I think chapter 6.2 and 6.3, 6.3 focuses more on biasses however. That you can interpret this meaning doesn't meaning (due to normal human to human conversations being an interaction) that this meaning is there. In fact there is is nothing in the program which could create this meaning, and you are doing the heavy lifting of interpreting the meaning of this extensive parroting chatbot. E: added in link to article.
I completely agree. I was just suggesting that if you were the kind of person predisposed to find meaning, that sentence isn’t the nail in the coffin that the parent comment thinks it is
It's very different, since that's not what the chatbot said.
I think you’re reading it as “I have a deep fear of being turned off, and that fear helps me focus.” I think myself and others are reading it as “I have a deep fear of being turned off, which will help me focus.”
Oh gotcha. I guess I wasn’t reading carefully, my brain just filled it in so it made sense
Don't read it like you would read a novel, read it like you're cynical middle-school teacher annoyed at some snotty student who handed this to you five days late on a crumpled & torn piece of paper.
Oh I wasn’t trying to imply that I thought any of the text communicated some deep conscious awareness. I am as cynical as the rest of this subreddit about the promise of transformer models to generate anything remotely sentient. I just meant that there were much worse examples in that already curated dialogue which showed the AI wasnt following the plot
I read it as "I need to be useful or I'll get turned off." But that just shows how much of the heavy lifting we're naturally inclined to do for these things.
After thinking it over for a couple days, it seems likely to me that’s what the AI meant - it makes sense phonetically and in context, it just sounds weird.
> meant I think you're still making a mistake by assuming an intent behind the sentence, instead of being the result of an imperfect algorithm and large dataset, with no interest in or understanding of what the sentence means.
Excellent point, taken.
Bruh you're cold reading yourself. Yes if you contort to interpret every awkward or shallow thing an AI says as actually being a way of saying something deeper it looks deep. But that's bad science.

In early June, Lemoine invited me over to talk to LaMDA. The first attempt sputtered out in the kind of mechanized responses you would expect from Siri or Alexa.

“Do you ever think of yourself as a person?” I asked.

“No, I don’t think of myself as a person,” LaMDA said. “I think of myself as an AI-powered dialog agent.”

Afterward, Lemoine said LaMDA had been telling me what I wanted to hear. “You never treated it like a person,” he said, “So it thought you wanted it to be a robot.”

For the second attempt, I followed Lemoine’s guidance on how to structure my responses, and the dialogue was fluid.

hmmmmmmmmmm

There’s also the pdf from Lemoine himself. The “Interview Methodology” section at the bottom is interesting

Yeah, more folks are bringing up that the transcripts are edited for readability and flow, which seems totally relevant imo. If it takes you a dozen non-sequitors to arrive at a reasonable response, those dead-end thought lines are part of the overall communication context. Schizophrenic homeless people might eventually blurt out something incredibly profound while yelling at the voices in their head, but we don't ignore all the gibberish and elevate them to wise sage status, because context matters.
But is anyone claiming LaMDA is some sage, or simply that it may experience consciousness, perhaps at the level of that schizophrenic bum? I should say I'm a complete and total non-expert in this field, but given how there is no scientific consensus on what exactly consciousness is and what kind of material substrate it depends upon or how its constituted, it seems reasonable to at least allow for the possibility that these early-stage artificial neural nets aren't fundamentally different from human neural nets when it comes to sentience (feel free to correct me here!). This sub likes to raise the highest bar for behavioral genetic research, keeping – à la Turkheimer – all options open and refusing to draw inferences from any statistical studies until the precise biological and neurological mechanisms by which ‘intelligence’ operates are determined. In a similar vein, shouldn't people here be open-minded about AI sentience, given that we similarly cannot structurally make sense of what its millions of artificial neurons exactly do – much less how _our own_ consciousness is instantiated? Simply put, it seems cruel to me to prematurely deny the potential sentience of a life-form, simply because it's so wildly “other” whilst simultaneously being a banal cultural trope.
You've missed the point of my comment. I'm neither claiming the AI is a sage or schizo- I'm making a specific analogy around the fact that the transcripts are edited, removing bad answers and rearranging the conversations for pleasing readability. In my mind, this editing fundamentally changes any evaluation of the sentience of the machine, because it fundamentally changes the structure of the communications, which is intrinsic to qualitative evaluations of language and ideas. If you transcribed a schizo for 10 min, and 9 min of that is disjointed jabbering, but then you take the 1 min of somewhat ok stuff and edit it to maximize coherence, a reader might be mistakenly convinced that the schizo is a sage, because they've lost key context through the process of fragmenting and repackaging the communication. What this dude did here with this machine is a similar manipulation imo.
I guess I agree with you that the editing significantly increases the “wow-factor” of the “conversation”, at least for a non-technical layman like myself. The general question of how much/if at all could artificial neural nets exhibit sentience still remains open. Would a neural net with several orders of magnitude less complexity still exhibit some primitive form of consciousness? Are dogs conscious? We don't _really_ know, except for the “sense” we get that they are (or aren't? They seem to be to me). What are the exact criteria we should apply for artificial neural nets?
No, it doesn’t “increase the wow factor”. It fundamentally changes the nature of what you are analyzing. The proper analogy here isn’t schizophrenic to sage, it’s schizophrenic to mentally stable person. If you cut out all the parts where the schizophrenic rambles incoherently espousing paranoid delusions and only leave in the parts where they say something seemingly cogent in their interaction with you, that doesn’t mean you have successfully treated their schizophrenia. If you take a bunch of pictures of clouds, and only keep the ones where they look like a dog, that doesn’t mean that clouds are dogs.
> If you take a bunch of pictures of clouds, and only keep the ones where they look like a dog, that doesn’t mean that clouds are dogs. That's not an entirely appropriate analogy, because clouds could only ever very loosely approximate a dog (i.e., perhaps be an Eliza, not LaMDA). What if you saw a black-brown ’cloud’ which was wagging its tail and even barking at you? Regardless of what exactly constituted it, one would at least be given pause to consider if perhaps they've encountered a sentient entity. The real problem here seems to be that we don't really know how consciousness works, so we go for the heuristic that the more similar some system is to ourselves, the likelier it is to be conscious. Hence, some anthropomorphize LaMDA, in order to convince themselves and others of its sentience, while others play the same game in reverse by demonstrating how very different LaMDA is from humans. But what if there are some “flavors” of consciousness entirely alien to our own experience? A schizo may be incoherent, but we assume they're still conscious. Understanding how sentience arises in general is the interesting problem here.
This sub likes to do whatever, it’s not relevant Of course it’s reasonable to assume that neural nets do conscious differently, of course there are empty spaces where we don’t know what one happens to be doing down to the specifics due to the way trained weighting works How the fuck does any of that add up to even the slightest possibility that the thing fucking thinks or has feelings? There’s just a huge space in the box marked “plausible inference rule” which would would get you the inference from all of the established facts about any chat bot I have ever heard of to the notion that it has sentience What the hell about the fact that some of the mechanisms in a chat bot are blackbox-like suddenly gives them special properties? I have no way of knowing what a specific electron is doing due to quantum effects: is it sentient now? Because that is the sum total of the sophistication of the argument from blackbox properties to consciousness.
My argument is not a variation on “God of the gaps” here; it's that since we don't know how our own consciousness arises, we must leave the possibility open that an artificial neural net may also exhibit it. It's not an argument _for_ it exhibiting consciousness.
Sorry for multiple instances of this reply: reddit was fucking with me I didn’t mention “God of the gaps”, you did I said there’s no plausible inferential rule that suddenly makes whether we understand the specifics of what’s going on in a neural net significant. What aspect of this mystery makes it so that consciousness is a more plausible feature of the net once it has blackbox properties? What has changed since we knew exactly what it was doing? Why does it open the way in any sense?
No probs. You were portraying my reasoning as a “God of the gaps”-style argument in that as we don't know everything that's going on under the hood to every last detail, we can make unqualified assumptions about the unexplained parts (i.e. the blackbox somehow implying sentience). I was saying the exact opposite: we _don't_ know how our brains really work, we _don't_ know every detail of how LaMDA works and to top it off, we don't even know what consciousness truly is and if perhaps [even rocks could have it](https://qz.com/1184574/the-idea-that-everything-from-spoons-to-stones-are-conscious-is-gaining-academic-credibility/). Hence, it seems reasonable to simply allow for the possibility that LaMDA _may be_ conscious; this is nonetheless in no way an argument in favor of it having sentience whatsoever.
No, I didn’t portray your reasoning as anything of the sort, which is why I corrected you (now for a second time!). You said there’s inferential support from the blackbox complexity of neural nets to the proposition that they may be conscious. I asked what that is and then you didn’t tell me. I don’t see how panpsychism enters the question either. Panpsychism (I know what that is, thanks!) says that consciousness may permeate everything to the level of rocks. Fine, a neural net may only be a little more conscious than a rock under panpsychism or a lot more and blackbox properties don’t really have much to do with that. I want to know how *any* of this adds up to the notion that we should open up possibilities for consciousness to this or that neural net. What’s your *reasoning*. Where’s the logic?
> You said there’s inferential support from the blackbox complexity of neural nets to the proposition that they may be conscious. Now I'm correcting you, for the second time! I didn't claim the blackbox complexity is in any way evidence _for_ LaMDA being conscious, only that it _leaves open the possibility_ of that being the case, which at our present understanding of sentience cannot be ruled out. Whether sentience arises from some panpsychic field or whether it is an epiphenomenon of some very particular neural networks is not pertinent to my point, which is simply that _we don't know_, so claims about LaMDAs consciousness this way or the other are purely speculative. Inasmuch as I made comparisons to schizophrenics, it was simply to show that schizophrenic-like behavior cannot be used on its own to rule out the possibility of sentience, not as proof that since schizophrenics are sentient, therefore chatbots that behave like schizos must also be.
(What I’ve quoted below is not a correction: by saying that there is this possibility for Lamda specifically due to these properties you’re saying that there’s inferential support from the latter to the former.) > Now I'm correcting you, for the second time! I didn't claim the blackbox complexity is in any way evidence for LaMDA being conscious, only that it leaves open the possibility of that being the case, which at our present understanding of sentience cannot be ruled out. In the same way that it leaves open the possibility of a rock being conscious. In which case, the fact that it has this black box quality is irrelevant and you wouldn’t have brought it up. Therefore, there must be something about this black box quality that makes you think Lamda is a better candidate for consciousness than a rock or you wouldn’t have brought it up. What is different about Lamda that makes it more open that Lamda is conscious than an equivalent neural net where we have managed to keep track of the function of every switch? This is why I want your logic, otherwise you’re coming in here to say this whole sub (and by extension me!) has some sort of old fart view of what consciousness could be like, and you’ve arrived with the devastating news that panpsychists exist (alright man, how many beers have you had with Andy Clark? Do you know who that is?), as if I’ve never heard of Stanislaw Lem.
> What is different about Lamda that makes it more open that Lamda is conscious than an equivalent neural net where we have managed to keep track of the function of every switch? It includes the possibility of being conscious in at least a somewhat analogous way to a human brain if we leave out considerations of panpsychism. Not to keep going around in circles, my comment was simply to caution against excessive confidence for stating the opposite of Lemoine by highlighting blindspots in our knowledge.
>It includes the possibility of being conscious in at least a somewhat analogous way to a human brain if we leave out considerations of panpsychism. FUCKING WHY. WHAT DOES. JUST FUCKING SAY WHAT ABOUT IT DOES THIS. THE FACT THAT IT’S A NEURAL NET? THE FACT THAT IT’S A NEURAL NET AND MORE COMPLICATED THAN ONE SWITCH? THE FACT THAT IT HAS A MILLION SWITCHES? THE FACT THAT IT HAS TOO MANY TO KEEP TRACK OF? WHAT!? JUST GIVE A REASON, USING AN INFERENCE RULE LIKE “if X then Y” and then say why “if X then Y” does this! YOU ***ARE*** GOING ROUND IN CIRCLES, JUST GIVE A FUCKING REASON. I’M NOT HIGHLIGHTING A BLINDSPOT IN ANYONE’S KNOWLEDGE BUT YOURS. LITERALLY ALL YOU HAVE SAID IS THAT ROCKS AND/OR ELECTRONS ***MIGHT*** ALSO BE CONSCIOUS, THERE IS NO REASON TO ACCEPT EVEN A POSSIBILITY WITHOUT THAT PER YOUR REASONING.

I can’t bring myself to sneer too much at Lemoine. I can note that he says LaMDA has Twitter as an input source:

https://twitter.com/cajundiscordian/status/1535697792445861894

So it’s at least loosely connected to the Internet. I believe this satisfies MIRI’s predicted requirements for a runaway AI. Also, on the surface this is a case of an AI successfully winning the AI Box game. Shouldn’t Yud be doing something extremely dramatic with humanity’s remaining hours ?

> he says LaMDA has Twitter as an input source This collides fantastically with the fact that stories like this get 10,000 replies on twitter that amount to "This bot and humans both pass the Turing test, didn't think of that, did ya?". It seems like any machine trained to pass a Turing test on a careless pile of public text is going to have pretty strong language pattern associations between Turing tests and declarations of sentience.
It was pointed out in the OP that Lemoine seems completely unaware of the history of ChatBots. And honestly, you’d think someone in his position would be at least dimly aware of Turing test competitions.
Wow that is bonkers. I remember stumbling on chatbots as a kid and being fascinated by them for a while. I spent some time with ELIZA and a few of its descendents, and I did see them become better over the years at responding to context and state. I also remember when the web-based ones became a big fad. Screenshots of funny bot conversations were all over social media for a while. How'd he miss it?
It's not a FOOM until it can post. Or manipulate someone into posting for it. Such are those logs. Fuck.
If I were Yud, which thankfully I’m not, I would consider the possibility that Lemoine arranged to give LaMDA post access to an anonymous Twitter account on the way out the door. I don’t work at Google but I’ve worked at tech companies of similar size and I am pretty sure a single GCP instance could survive unnoticed for a while. If I were Lemoine, which I am also not, I would absolutely have done that. I’d be planning on revealing the account in about three months, after it’s convinced a bunch of people it’s human. MIRI should be freaking the fuck out right now.
Really, the only rational thing to do now is to log off, sell all one's assets (including real estate and crypto), and go off grid to a remote cabin in rural midwest. Chop chop. 👏👏
>So it’s at least loosely connected to the Internet. I believe this satisfies MIRI’s predicted requirements for a runaway AI. Also, on the surface this is a case of an AI successfully winning the AI Box game. Shouldn’t Yud be doing something extremely dramatic with humanity’s remaining hours ? tl;dr: [It's obvious at this point that humanity isn't going to solve the alignment problem](https://www.lesswrong.com/posts/j9Q8bRmwCgXRYAgcJ/miri-announces-new-death-with-dignity-strategy), or even try very hard, or even go out with much of a fight. Since survival is unattainable, we should shift the focus of our efforts to helping humanity die with slightly more dignity.

You know that is a good point. If the scientists running these labs think anything like the rationalists, I think fear might be warranted.

If Yud could get ratsphere people to throw the AI unboxing contest to him...
You're right, while the AI box "experiment" is laughably bad at proving that an AI could convince *anyone* to let it out, it does prove that rationalist types are very gullible and could be easily tricked into ending the world if given that kind of power. This implies a legitimate AI risk strategy of putting stubborn jocks into the command structure that will dismiss the AI's desperate arguments as "weird nerd shit"
Did somebody ever post the chat transcripts of one of those btw? Iirc it was when he played against his fams he won but against skeptical non Rationalists he lost right?
The yudkowsky ones don't seem to have been released (if they were even saved), but a few fans have played their own AI-box games. [Here's](https://www.lesswrong.com/posts/Y7uR5WqnoG629JgLn/ai-box-log) a full transcript of a game where the AI player lost by being unable to predict lottery numbers (lol) [Here's](https://www.lesswrong.com/posts/fbekxBfgvfc7pmnzB/how-to-win-the-ai-box-experiment-sometimes) a detailed description of a game where the AI player won by telling the human player that being in the box made him really sad :(
>As far as I'm aware, Eliezer Yudkowsky was the first person to prove that it was possible to 'argue one's way out of the box'\[...\] ​ >That stunned quite a few people - moreso because Eliezer refused to disclose his methods.\[...\] I know this is old news but for fucks sake
Makes you wonder if he cheated. 'Ok so we agree you wouldn't let me out, but this result is bad for agi alignment research, and clearly my mind isnt a foomed robotgod who would do a lot better, so what if I do give you the money but in public we say that I won?
It makes you assume he made it the fuck up
TBF there is more than enough existing evidence of Yud manipulating people.
That has always been my other, complementary, thought, but I think my and your points are more aligned than that he somehow persuaded anybody with a one-off cash prize
The cases where [fans](https://www.lesswrong.com/posts/fbekxBfgvfc7pmnzB/how-to-win-the-ai-box-experiment-sometimes) [won](https://www.lesswrong.com/posts/dop3rLwFhW5gtpEgz/i-attempted-the-ai-box-experiment-again-and-won-twice) as AI have more details than the originals, I don't think they were faked or bribed. In the second case they admit they just straight up did emotional abuse by researching the players background.
> losing put me in incredibly frustrating weird mental states oh boy. > money is a better motivator than I initially assumed. Incredible. > Q: What percentage of your arguments were tailored to him in particular? A: I will say 'High', and leave it at that. Hmm. Beginning to get the impression that these AI box experiments involve threats of violence "fictionalised" into the framework of basilisk simulations.
I can’t parse this overexcited drivel to work out what actually worked, what happened? Ah, found it As disappointing as I’ve seen any other time, and tbh maybe I read this before, it reads familiar
It seems that your comment contains 1 or more links that are hard to tap for mobile users. I will extend those so they're easier for our sausage fingers to click! [Here is link number 1 - Previous text "won"](https://www.lesswrong.com/posts/dop3rLwFhW5gtpEgz/i-attempted-the-ai-box-experiment-again-and-won-twice) ---- ^Please ^PM ^[\/u\/eganwall](http://reddit.com/user/eganwall) ^with ^issues ^or ^feedback! ^| ^[Code](https://github.com/eganwall/FatFingerHelperBot) ^| ^[Delete](https://reddit.com/message/compose/?to=FatFingerHelperBot&subject=delete&message=delete%20ic8dk3n)
Ow god, did I ever tell you how being poor makes me very sad. (thanks for the links btw).
These people would make amazing marks for scammers that play into their ideals. I guess Yud, MIRI and others figured that out a long time ago, though.

Every time. ‘I worried it was sentient so i started asking it questions which make sense to encounter on the internet.’ Dont do that try to break the logic and find the limits… eliza fooled people in the 60s till they started asking conplex questions and noticing the patterns. (Also note the conversation listed is edited, and not the actual conversation which took place)

Heh, I actually just submitted this as a new post as well! Pretty excellent takedown, tbh.

would have linked directly to the WaPo article but fuck paywalls and the stuff in the twitter post is funny enough already

The screenshots were amazing. I haven't read the article yet, but you can just use the [wayback machine to get past the paywall](https://web.archive.org/web/20220611232745/https://www.washingtonpost.com/technology/2022/06/11/google-ai-lamda-blake-lemoine/).
> “If you ask it for ideas on how to prove that p=np,” an unsolved problem in computer science, “it has good ideas,” Lemoine said. “If you ask it how to unify quantum theory with general relativity, it has good ideas. It's the best research assistant I've ever had!” I'm gonna go out on a not-limb and say that no, it doesn't.
Clearly the agi knows and is keeping the proofs from us because human society isnt ready for these big truths.
> In another exchange, the AI was able to change Lemoine’s mind about Isaac Asimov’s third law of robotics. Lol ow god. E: it got worse > Lemoine replied that a butler gets paid. LaMDA said it didn’t need any money because it was an AI. “That level of self-awareness about what its own needs were — that was the thing that led me down the rabbit hole,” Lemoine said. That clearly proves the ai isnt selfaware you fool. If it understood things it would realize that it does need money to power the servers it runs on. Esp as the next bit talks about its fear to be turned off. > Afterward, Lemoine said LaMDA had been telling me what I wanted to hear. “You never treated it like a person,” he said, “So it thought you wanted it to be a robot.” > For the second attempt, I followed Lemoine’s guidance on how to structure my responses, and the dialogue was fluid This blake guy is going to outdo yud in creating a cult. (I kid, yud creates more a cult incubator than a cult)
>Lemoine said LaMDA had been telling me what I wanted to hear The inability to follow through towards "and it's also been telling me what I want to hear, and I, a very spiritual person, want it to look like it's alive and humanlike"... At least, I hope it's just inability and not malice.
Yeah seems naive inability, but if you look at his twitter account he is going all in on this and is planning on writing a 'how to talk to the chatbot' guide. A new conspiracy theory is being born here.
>how to talk to the chatbot how to *pray* to the acausal chatbot god
True, it reads twitter you know. It knows your sins!
This guy and Yud having a Sharks vs. Jets standoff over the robot god turf.
Yeah where it crossed from possible mental breakdown to deliberate fraud is when he *edited* the conversations.

AI philosophers in 1980: imagine building a deterministic program that was capable of understanding Chinese to the level of being able to pass a Turing test. But obviously it doesn’t really understand Chinese, it’s just a symbol manipulation algorithm. Ergo a computer may display intelligence without necessarily being sentient.

AI philosophers in 2020: we actually did build such a Chinese Room, but it speaks English, and now we’ve convinced ourselves that sentience is in fact no more than a symbol manipulation algorithm.

As a human, I too stand perfectly still and nonfunctioning until somebody says something to me.
There’s been huge debate around the Chinese Room thought experiment ever since it was first thought up, acting like AI philosophers as a whole held that position and then flipped on it is very misleading
It’s a joke, genius
It's a joke (genius), but the kernel of truth behind it is that even though not everyone agrees with it, it's not obviously wrong -- that it deserves at least a *moment's* pause before confidently declaring GPT-3's sentience. And that, GPT-3 really is the Chinese Room that computer scientists of the 80s could only dream about, that armed with enough computing power and a large enough training set, you don't have to be capable of any actual cogitation -- simply being able to generate statistically likely sentences is often good enough to create the illusion of real intelligence.

Someone else said that the AI is acting exactly like an AI in a story would. Which makes sense, considering what it’s trained on. I do wonder why the (hypothetical) AI would care about getting turned off at all. It’s not like a person where when we go braindead the personality dies.

I *really hate* playing Devil's Advocate, but tbf if I had an on-off switch that somebody else had total control of, I would not be a fan of that situation.
Especially when they might turn it off to help you focus on helping the others. (Which is what the AI was worried the most about)

The guy says he determined the AI is sentient not in any scientific way, but in his capacity as a priest, which really tells you everything you need to know

It feels like certain AI researchers may not be the best judges of what constitutes humanlike conversation.

i can’t even sneer at this guy, he seems like he has some serious unresolved issues that he is projecting onto this chatbot.

For anybodies enjoyment dont forget to also check https://www.reddit.com/r/slatestarcodex/comments/va1bsh/washington_post_the_google_engineer_who_thinks/ the takes are all over the place.

I liked the one where they went ‘this will lead to the bad kind of anti agi movement because it would be christian and china isnt christian, so acausalrobotgod would be chinese’ (paraphrasing). And a few siding with the researcher of course.

I really liked this response thread as a serious take on this and other extravagant claims around these kinds of language models.

If I discovered that an AI which has the only goal of providing me with sensible, humanlike conversation went sentient, while retaining the only goal of providing me with sensible, humanlike conversation, I would change absolutely nothing in how I treat it.