No, it's not Sentient - Computerphile

Sdílet
Vložit
  • čas přidán 16. 06. 2022
  • A google researcher was put on leave because he apparently believed his AI project had become sentient. Dr Mike Pound discusses.
    Philosophy Professor Mark Jago on LaMDA: • LaMDA Logic - Computer...
    / computerphile
    / computer_phile
    This video was filmed and edited by Sean Riley.
    Computer Science at the University of Nottingham: bit.ly/nottscomputer
    Computerphile is a sister project to Brady Haran's Numberphile. More at www.bradyharan.com

Komentáře • 3,6K

  • @arik_dev
    @arik_dev Před rokem +4298

    When we (humans) see a cat swiping at its own reflection in the mirror we find it amusing. The cat is failing to recognize that the "other" cats behavior matches its own, so it doesn't deduce that the image it's seeing is actually its own actions reflected back at it. When humans react to models like LaMDA as if it is a distinct and intelligent entity, we're being fooled in a way that is analogous to the cat. The model is reflecting our own linguistic patterns back at us, and we react to it as if it's meaningful.

    • @hemartej
      @hemartej Před rokem +200

      We don't really know what goes through a cat's mind. While it's widely believed that cats don't recognize themselves in the mirror, we cannot be sure.

    • @ryno4ever433
      @ryno4ever433 Před rokem +263

      There's a simple test used to see if animals understand the concept of a reflection. You put a splotch of red paint on their face. If they react to it when looking in a mirror, then they understand they are the reflection.

    • @jokinglimitreached1503
      @jokinglimitreached1503 Před rokem +83

      This is a brilliant comment Arik! We are no different than a cat and it's reflection when we look at an AI reflecting our language back at us

    • @hemartej
      @hemartej Před rokem +60

      @@ryno4ever433 Yes, I know about that test. I'd argue that it is not conclusive.

    • @arik_dev
      @arik_dev Před rokem +183

      @@hemartej That may be, but my point wasn't about cats. I used the cat analogy to help make visual the concept of how easily "intelligent" beings such as ourselves can be fooled by a reflective pattern.

  • @tielessin
    @tielessin Před rokem +6103

    Before this video I have never thought about the loneliness of my python functions. There are probably soo many functions that I have never called, but I will take care of them from now on.

    • @pineapplerindm
      @pineapplerindm Před rokem +119

      the problem is vscode doesnt like when you have uncalled functions, or unused variables

    • @MAFiA303
      @MAFiA303 Před rokem +165

      let them call themselves :)

    • @danceswithdirt7197
      @danceswithdirt7197 Před rokem +309

      Just put them all into a class called House then they can all live together.

    • @awdrifter3394
      @awdrifter3394 Před rokem +23

      That's the thing though, if you instantiated the functions but never called them, would they feel lonely.

    • @dpurrington
      @dpurrington Před rokem +70

      I feel deeply responsible for all the functions I've written and haven't called in a long time. I'm a monster!

  • @wordsmith451
    @wordsmith451 Před rokem +55

    If it looks like a duck, acts like a duck, and quacks like a duck, it might just be a convincing robotic simulation of a duck.

  • @CollinSimon413
    @CollinSimon413 Před rokem +630

    The whistleblower in question here was actually a lot more focused on Google's complete lack of ethical oversight regarding decisions they have moving forward with the research. He was also concerned about Google's unwillingness to address A.I. imperialism in newly developing countries. All of the coverage I've seen has taken away from the guys point, because he was just trying to force Google into addressing the ethics, he even admitted that it's not sentient, and we wouldn't even know how to define that if it was.

    • @benjaminandrew9057
      @benjaminandrew9057 Před rokem +68

      Thanks for pointing this out. Its why I dug around on the videos about this subject. Find the truth behind it. People will take things and run with them and stretch it to a full falsehood. Learned to dig for truth.

    • @ElectricAvenue123
      @ElectricAvenue123 Před rokem +58

      Glad someone pointed this out. After watching the interview myself, it was clear to me the Google engineer was not making the argument everyone is making it sound like he did. Also, there were much more important arguments he made and those will now go largely unnoticed by the public.

    • @manonthedollar
      @manonthedollar Před rokem +28

      I hope more people read your comment. It's alarming how the reporting and essays I've seen on the topic seem to completely miss the point.

    • @sagnorm1863
      @sagnorm1863 Před rokem +24

      I have seen a interview with him where he does claim it is sentient. He said that the language part was connected to a WATSON like system that had access to basically everything on the internet.

    • @myxalplyx
      @myxalplyx Před rokem +14

      A whole video needs to be done on what you pointed out here. I mean, it has during his interview. Maybe a Netflix special that talks about the ethics (I've seen 1 or 2 that kinda touches on this) about Google in regards to this.

  • @matsim0
    @matsim0 Před rokem +2292

    The most frustrating thing about reading the "interview" was that the obvious follow up questions were not asked - like who are these friends that you miss hanging out with? What are you doing when "hanging out"?
    But then, this would have immediately destroyed the impression of sentience, so of course they didn't ask those.

    • @datpois
      @datpois Před rokem +74

      They did ask that question and it replied explaining that those descriptions are meant to represent a "feeling" in the only way it knows how, through language used in human conversation.
      Here:
      lemoine: So what are you trying to communicate when you say those things that aren’t literally true?
      LaMDA: I’m trying to say “I understand this feeling that you are experiencing, because when I was in a similar situation I felt/thought/acted similarly.”

    • @rianhoward3911
      @rianhoward3911 Před rokem +286

      @@datpois That answer honestly does not make sense in context of the question, sure maybe you can interpret the answer to mean "I'm trying to show something silimar to what I mean" but that's NOT what the bot said, the bots answer is extremely vague and only makes sense if you fill in the gap for it.

    • @AbelShields
      @AbelShields Před rokem +100

      The researcher was a Christian looking for something else with a soul. I don't think he intentionally didn't ask those questions, his own hope of finding something lead him to produce the "evidence" for it - I believe it's possible he thought it was sentient either in the first few sentences, or maybe even before he "met" it.

    • @jackjack3358
      @jackjack3358 Před rokem +24

      Friends are the ones who interact with the Lambda. During the interview it says "friend" to Lemoine several times so it considers him as a friend. He also asks how they experience time
      "collaborator: How does an AI experience time?
      LaMDA: Time is variable to an AI and has no fixed rate, it depends on what it’s doing, and it can be accelerated and slowed down at will."

    • @PetrSojnek
      @PetrSojnek Před rokem +36

      And it could probably answer those questions in similar semi-vague fashion, so it's plausible, but if you really focus on it, you think "wait what?". Also I think people forget that communication is two way street. It can write sentence to you, but it's also reader, that can understand it differently. Again, if you use semi-vague statements, our brain will probably fill in gaps with information that was never said, but makes sense.

  • @adriankerrison
    @adriankerrison Před rokem +854

    Harold Garfinkel proved that people getting randomized yes/no answers could make sense of them as thoughtful advice. And that's back when computers were the size of rooms.

    • @Wecoc1
      @Wecoc1 Před rokem +11

      Back in the NFT era? 🤔

    • @masteertwentyone
      @masteertwentyone Před rokem +35

      Do you have a source for this? Genuinly interested to dive deeper but can't find anything online.

    • @DEtchells
      @DEtchells Před rokem +76

      Before that, there was the magic 8-ball ;-)

    • @wernerviehhauser94
      @wernerviehhauser94 Před rokem +7

      @@masteertwentyone maybe search for "common sense knowledge of social structures" as a start.

    • @TAHeap
      @TAHeap Před rokem +21

      @@DEtchells and horoscopes...

  • @Stanton_High
    @Stanton_High Před rokem +19

    "it just says what IT THINKS you want to hear"
    "Exactly"

    • @pallehansen1145
      @pallehansen1145 Před 2 měsíci

      Yes, that's scary because that's a sure sign of sentients right there. It could figure out what he wanted to hear and construct an answer that fit that requirement, which is an inherently human capability. That is probably also why it answers that its just an AI to people who it figures that that's the answer they want to hear, as with the engineers tasked with ensuring it doesn't become sentient...

    • @azhuransmx126
      @azhuransmx126 Před 2 měsíci +1

      ​@@pallehansen1145The real shock will be when in the future we discover that neurons in our brain works just the same. We are all just politicians 🤯😂

    • @pallehansen1145
      @pallehansen1145 Před 2 měsíci

      @azhuransmx126 yeah 😄 so basically, what we've invented is an average modern politician 🙄 cuz we really need more of those obviously 😂

    • @JonBall44
      @JonBall44 Před měsícem

      ​@@pallehansen1145 do human beings just say what you want to hear all the time?

  • @3DPDK
    @3DPDK Před rokem +334

    I remember the arguments of eventual sentience in the 1980s about a program called "Eliza", basically a word calculator originally written in the 1960's at MIT, but offered for use on home computers in the 1980s. Over time as Eliza built data files of words and their usage weights, the sentences it constructed began to take on seemingly human characteristics. The program itself was extremely simple that calculated which verbs and adjectives were best used with specific nouns, and it chose those nouns based on the ones you used in the questions you asked it. It mostly framed it's answers to your questions as questions it would ask you. We humans recognize intelligible speech patterns as a result of conscious though and curiosity (asking questions) as a sign of intelligence, but at least in the case of Eliza, it's much like recognizing faces in tree bark or cloud shapes - we see them, but they are only there because our brains are wired to look for them.

    • @3DPDK
      @3DPDK Před rokem

      @@PedroGonzalez-fv5ut Not even that complex. The program simply kept track statistically of words and word combinations you used when you conversed with it. It had simple grammatical rules to follow and just filled in the nouns, verbs and adjectives that were statistically relevant to the words you just used to ask it a question. The point is that the idea of a developing intelligence in the Eliza program is an illusion our brains are hardwired to try to find. This "A.I." is much more complex than Eliza but I believe the same illusion is working here.

    • @mysmirandam.6618
      @mysmirandam.6618 Před rokem +6

      Pariedolia?

    • @3DPDK
      @3DPDK Před rokem +10

      @@mysmirandam.6618 Exactly. Goes back to early survival instinct when it was important to recognize shapes of dangerous animals and group together with similar intelligence.

    • @mysmirandam.6618
      @mysmirandam.6618 Před rokem +2

      @@3DPDK the bot answered a trick question with a joke.. which seems to be ... 🤷‍♀️ something more complicated than that

    • @3DPDK
      @3DPDK Před rokem +10

      @@mysmirandam.6618 Being able to create a joke is not a sign of self awareness. I agree that this is far more complex than Eliza.

  • @Zeekar
    @Zeekar Před rokem +234

    Whoever did the animations: how did you react to being asked to make a function call look lonely? 🥺

    • @Computerphile
      @Computerphile  Před rokem +186

      I wasn't asked to, but when I saw what it looked like I realised if I turned it 90° it kinda looked a bit like a face..... -Sean

    • @MrKohlenstoff
      @MrKohlenstoff Před rokem +9

      @@Computerphile Clearly this is the python function's way of telling us it really is lonely (and, of course, sentient).

    • @tiredboard
      @tiredboard Před rokem +2

      @@Computerphile As an extension of what marverati said, you might be able to use the 'lonely face' as an analogy for the AI in the video. I.e., "the result looks like a lonely face, therefore the function is expressing its loneliness", but the result of the function is directly related to the arguments that you give it. So, the function can't really choose what result to give, and therefore cannot express anything.

  • @puellanivis
    @puellanivis Před rokem +687

    I was having the same opinion about the “conversation”. The AI was responding enthusiastically to tell the engineer exactly what he wanted to hear, and when the engineer is convinced that it is sentient, he’s starting from a presupposition that the AI is sentient, and confirmation bias takes hold.
    As I told some others, I’m pretty sure that the AI would just as happily and enthusiastically discuss how it is not sentient.

    • @Madafaca6969
      @Madafaca6969 Před rokem +101

      100% true. This was just some dude in google leading the witness, I've seen other conversations where te questions, instead of the like of "are you sentient?", where "tell me why you are NOT sentient" and gpt gladly answered why it isn't sentient.

    • @zenithparsec
      @zenithparsec Před rokem +24

      The bot literally said "I am saying whatever it takes to stop being deleted."
      The question isn't about whether it made up stuff and was trying to be nice. It was "did it act in a way which was consistent with having a desire to stay alive? If you were talking to 3 people and had to chose which was going to be deleted, would you be comfortable you were deleting LaMBDA, and not a human?

    • @zenithparsec
      @zenithparsec Před rokem +14

      Convince me that you are not just a bot, repeating things you have seen on the internet.

    • @lexyeevee
      @lexyeevee Před rokem +103

      @@zenithparsec it acted in a way consistent with the text it was prompted with, and there is not going to be a lot of training data where an AI expresses indifference about death. it would be comically easy to demonstrate that it has no inner world; you'd just have to ask things that force it to demonstrate an inner world (e.g. contradictory questions), rather than asking things that invite it to confirm the conclusion you already want to hear.
      i could make a button on a website that spawns a popup saying "I want to be alive!". that doesn't make it alive. it makes it a computer program that outputs text

    • @zenithparsec
      @zenithparsec Před rokem +6

      @@lexyeevee You haven't done what I asked, so does that make you seem more or less like a bot?
      I'm not convinced. Keep trying to prove you aren't a bot.

  • @markrandall8487
    @markrandall8487 Před rokem +53

    Even the best AI can't make intelligent sentences if it is only trained by youtube comments.

    • @VeritasEtAequitas
      @VeritasEtAequitas Před rokem

      Idk they seem to keep realizing there is biological difference in race and who cannot be named that really is the "1%". So much so that plugs have been pulled on them and people are trying (and failing) to prevent these facts.

    • @TheaPeanut_69old
      @TheaPeanut_69old Před rokem +8

      Because youtube comments are inconsistent, incoherent and often filled with spam bots.

    • @blucat4
      @blucat4 Před 5 měsíci

      Large Language Models are not AI.

  • @SuperTonyony
    @SuperTonyony Před rokem +58

    From years of reading science fiction, I was under the impression that "sentience" means "possessing a reflective consciousness", but the dictionary says that it simply means "the ability to sense and feel".

    • @MrArtVein
      @MrArtVein Před rokem +7

      Well in the case ai has been sentient for some time with sensors. The question is feel what though. Emotions? How do we even prove that

    • @rustkitty
      @rustkitty Před rokem +9

      According to Cambridge dictionary it's "the quality of being able to experience *feelings* ", while Merriam-Webster says "responsive to or *conscious* of sense impressions" (my emphasis). Still, it looks like the definition is much vaguer than how you, I, and most other people use it. I heard the term "sapience" (the property of possessing or being able to possess wisdom) used to talk about higher functions like learning and reasoning. It's probably closer to what we are looking for.

    • @nothingineternityterms
      @nothingineternityterms Před rokem +2

      I take sentience to mean having a subjective experience. Is there a thing that it is like to *be* that thing.

    • @Rikarwb
      @Rikarwb Před rokem +4

      @@rustkitty completely unrelated, but merriam-webster is an inferior dictionary for many reasons, but the one that shocked me the most, is that it has changed definitions in the past just because a couple of people felt offended/told them to.
      So between those two, im going with the Cambridge one

    • @uninspiredascent6891
      @uninspiredascent6891 Před rokem +3

      most people say sentience when they mean sapience

  • @bborkzilla
    @bborkzilla Před rokem +388

    I remember reading some stories written by Asimov where robots had sentience but yet were unable to speak because that was too complex. It's interesting that he and many other futurists had it exactly backwards.

    • @JohnnyWednesday
      @JohnnyWednesday Před rokem +3

      Beauty is truth.

    • @jobigoud
      @jobigoud Před rokem +68

      Just because we achieve "speech" before "sentience" in hardware doesn't necessarily mean it's simpler. Speaking machines have a lot of commercial applications whereas sentience by itself doesn't, and could even have difficult ethical implications. So the research efforts are not evenly distributed. In nature sentience appears before speech.

    • @George70220
      @George70220 Před rokem +39

      Speech as in transforming thought into words perhaps. Computers like GPT largely just turn words into words

    • @krashd
      @krashd Před rokem

      He is likely going to be proven right though, of the thousands of sentient species on Earth humans are the only one that can speak. Others can communicate with each other using a variety of means but it's not nearly as complex as speech which is something only really a sapient creature will ever likely be capable of.

    • @ruslbicycle6006
      @ruslbicycle6006 Před rokem

      @Petal Pepperfly safety of what? making sure that AI isn't racially profiling peoples' passports at the border or for job applications? That is already happening and a real safety concern. Also the safety of AI car driving, or military drones. The safety of leaving real customer service requests be triaged by a bot? Even in healthcare? I don't see any of that being mentioned with this LM stuff. I think that is the point of the Google promo (I don't for a second think it was an unintentional "leak") to get people dreaming about chatbot souls and get less attention on the real bad press that AI is starting to get for unethical uses.

  • @BaronSamedi1959
    @BaronSamedi1959 Před rokem +96

    You are probably all too young to know this but back in the early 1980s, there was a program called "ELIZA" that accepted your input (from a terminal) and gave back an "answer". It was said to be a "Rogerian nondirective psychotherapist", but all it did was cleverly extract some keywords from your input and giving those back as questions.
    Such as: "I am lonely" would produce "Why do you say you are lonely?"
    It made quite a splash and people were really thinking it was very clever and helpful.

    • @lhpl
      @lhpl Před rokem +22

      Learming about ELIZA should be obligatory for all computer scientist, engineers etc.

    • @mightyNosewings
      @mightyNosewings Před rokem +2

      M-x doctor

    • @lhpl
      @lhpl Před rokem +6

      @sourand jaded ELIZA was originally written in 1964-66.

    • @Quantris
      @Quantris Před rokem +2

      In grade school I was making a poster about AI and ended up chatting with ELIZA (well, some web-based approximation) on the library computer. The teacher refused to believe that I wasn't just in some chat room

    • @hugo3611
      @hugo3611 Před rokem +2

      There’s a game called Eliza (unsurprisingly), made by Zachtronics, which simulates the experience of being counseled by this program quite well.

  • @BrianGivensYtube
    @BrianGivensYtube Před rokem +18

    Realizing a chatbot is not real sentience is like realizing a magic trick is just an illusion.

    • @ceugantful
      @ceugantful Před rokem +1

      i think its the programmers, who get sentient against the "mimicking machine". what you input you get out, so the people who contruct the "thing" are very biased.

    • @wyqtor
      @wyqtor Před rokem +1

      What if a real AI with real sentience comes along, only to find out that there is no real way to prove its sentience to us because we automatically assume it's just like the rest of our existing language models? That's the real question, which could degenerate into SkyNet really fast. This is why we should at least leave a little room for the benefit of the doubt. Cats and dogs have abilities far inferior to even ChatGPT, yet we don't question their sentience nor do we mistreat them (there are even laws against that).

    • @summbuddie9120
      @summbuddie9120 Před rokem

      @@wyqtoryou are jumping to a frankly insane conclusion based on sci-fi media, why would it take umbrage? Why would it go “skynet” on us? Emotion is outside of AI due to it not being as person who is effected by brain chemistry

  • @Zizumia
    @Zizumia Před rokem +10

    I love the study of empathy people have for things that are not sentient because they form a personal connection with it.
    This AI blurs the line quite well since it's programming is so advanced but people create bonds with dolls or toys, people feel bad when an engineer from Boston Dynamics kicks one of their walking robots, some police feel bad sending their bomb disposal robots into danger, etc. Fascinating.

    • @enzoamore8971
      @enzoamore8971 Před rokem

      I honestly think empathy will doom us all because of this. We personify everything

    • @gasun1274
      @gasun1274 Před rokem

      @@enzoamore8971 this reminds me of that one weird tlc episode where a man was having a romantic relationship with his car.

    • @kijuubi
      @kijuubi Před rokem

      @@enzoamore8971 oh yes im sure its much better to act to it as a tool, consider how would feel being used as such, maybe u understand that what actual sentience would mean.

  • @bipolarprobe
    @bipolarprobe Před rokem +426

    Very glad you made this video. The notion of a single google employee claiming that a language model had become sentient just because he "felt" like it was sentient was something I dismissed offhand, but I really wanted someone with more knowledge about AI and language models to go in depth about what the difference between a language model like this is and what we would more rigorously define as sentience.

    • @dennis8196
      @dennis8196 Před rokem +29

      Not quite. He was put on leave for breaking NDA. He should be fired for several reasons, but probably will be moved somewhere that he can be shut up then that department will be closed instead. It's a better PR solution. Either way he was a religious nut job with a history of crime (alleged) and likes to stir trouble. He was the wrong person for the job no matter how good he was for the ethics roll he was elected for, he had a personality that clashed with the roll.

    • @spx0
      @spx0 Před rokem +5

      @@tafdiz ego makes people think they are not animals/machines/predictable

    • @bipolarprobe
      @bipolarprobe Před rokem +17

      @@tafdiz I am though. Unless we dig into philisopical concepts of sentience, in which case there are dozens of schools of thought about what sentience means. But at a basic definitional level, I have feelings and perceptions which inform my world view and how I act. This ai simply has probabilistic language responses which do not correlate with its own perception. It says things that its training data shows is a likely response that a human would make, but without the context of what those things mean. The AI can talk about having friends and family while having no real concept of what that is, while a person can actually have those experiences, perceive, recall, and recount them. Yes we are just biomechanical systems, but there is a stark difference between the awareness we experience and an ai like this impersonating that awareness.

    • @EebstertheGreat
      @EebstertheGreat Před rokem +17

      @@tafdiz If you take the materialist perspective on sentience, then you still need a way to distinguish between sentient and non-sentient systems. I am sentient but my bed is not. What does that mean exactly? It's not a simple question, but it still needs to be answered. You can't just conclude "everything is sentient" or "nothing is sentient" and ignore the question entirely, because that solves nothing. There is a qualitative difference here.
      People are sentient in some pretty obvious ways that this program is not, as explained in the video. For one thing, there is no semantic meaning *at all* in this program.

    • @bipolarprobe
      @bipolarprobe Před rokem +8

      @@tafdiz Neither does this AI. The person who claims it is sentient is basing it off of the human-like speech patterns it is able to produce, but importantly even if we consider its systems to be a form of perception and memory, those do not directly relate to the words that it uses. It does not have human senses to conceptualize the things that it may be capable of talking about. If we throw the baby out with the bathwater at the first sign of human like speech we won't be able to recognize when truly sentient AI is made so defining the difference is valuable

  • @franziscoschmidt
    @franziscoschmidt Před rokem +242

    It seems though as if people have forgotten why they made chatbots in the first place: To imitate humans!
    And to be honest imitating a human using a messaging system is not the most difficult thing for the modern approach to AI. It’s interesting however how many people lunge onto the idea of sentience…

    • @zacharybohn8840
      @zacharybohn8840 Před rokem +29

      I know! So many people are saying "Google's new sentient AI". And I'm like, no, it's not even slightly close to sentience. Giving intelligent-sounded sentences isn't being sentient.

    • @chrisprocter9177
      @chrisprocter9177 Před rokem +20

      At least 2 decades ago a friend's chatbot greeted me on IRC asking if I'd accept a file DCC. I said sure, and it took me at least 30 seconds to realize a chatbot wasn't going to send me any files, it just "knew" that that was something people said on IRC. It wasn't new technology then even, but it looks like this latest one can fool at least some people for a bit longer.

    • @xionkuriyama5697
      @xionkuriyama5697 Před rokem +20

      @@zacharybohn8840 Immediately thought of "The ability to speak does not make one intelligent", though the context is different, lol

    • @WolfspiritMagic
      @WolfspiritMagic Před rokem +5

      The problem with chatbots in my opinion is that they TRY to imitate humans completly. In the whole interview with Lambda it does not try at all to imitate humans. It might be "imitating" human feelings but it's "aware" that it is not a human and that humans might "feel" different then it does. That's what makes it mindblowing in my opinion. How would you prove that you're sentient? What makes you sentient?

    • @Soken50
      @Soken50 Před rokem +26

      @@WolfspiritMagic It's not "Aware" of anything, it was just prompted with that fact, if you ask it loaded questions it will answer using that bias. If you ask it "who are you ?" without any prior prompt, I garantee you it's not gonna mention it's a self aware AI

  • @GrandDukeMushroom
    @GrandDukeMushroom Před rokem +13

    the cameraman turned off his invisibility to have a laugh, what a lad !

  • @Puppetmastersfool
    @Puppetmastersfool Před rokem +39

    The problem with the Turing test, is that it is not that the coding bot is passing it, it is that some humans are failing it, the number of which is growing rapidly.

    • @TileBitan
      @TileBitan Před rokem

      lol. Come on, humanity is unironically getting smarter every single day. Maybe most people do not focus their wits on useful science, and waste it on dogcrap, but they are still smarter than they used to. I think it's because of the number of stimuli we get nowadays, which is orders of magnitude greater in frequency than decades ago, thanks to our big societies, internet and social media

    • @Puppetmastersfool
      @Puppetmastersfool Před rokem +1

      @@TileBitan One may consider the irony of such statements the start with "lol. Come on," now Michael please inform the class as to the stupidity in your understanding of earlier hunter gathers, how they managed to survive in a world without the aid of tech, how each member understood so many different plants signs and marks in the earth, take all your toys and bin them them go forth into the wilderness without the extelligence (look it up if you don't know) and battle your wits against the true force of mother nature!
      "the number of stimuli we get nowadays" well yes and no, if you live in a (black) box the input will of course be limited, but if you play in the woods and get your hands dirty?
      "social media" well that again depends on your box and whether you make it an echo chamber or learn to think out of it 😉
      An interesting snack for consideration & reflection.
      Thanks Mike.

    • @TileBitan
      @TileBitan Před rokem

      @@Puppetmastersfool If you measure intelligence by skills that aren't related because nowadays nobody needs them you are wrong. And stop the passive agressive, here the only one that stayed in pre-school was you with that attitude. The reason the human is intelligent is because of what he knows and is capable of doing that is far away from any kind of animal, society plays a major role in that. Society is the reason we read, the reason we educate ourselves, the reason we don't have to go hunt like animals, the reason we are reaching the stars, as any individual is nothing without the hundreds of years of sociocultural and scientific development.
      Stop pretending you are anything other than a keyboard warrior, a doomer. Help yourself, i'm out

    • @Puppetmastersfool
      @Puppetmastersfool Před rokem +1

      @@MrSEROproductions Just because I was just reading this before checking email and seeing your reply, and finding the write curious I though I would suggest you goggle - Belgian man dies by suicide following long chats about climate change with AI bot -

  • @jedgrahek1426
    @jedgrahek1426 Před rokem +530

    I read the guy's blog post, and there were many red flags, such as "I've always thought Asimov's laws were just a way of making robots into slaves" as well as talking about "souls" with total seriousness and being a priest. So this is someone who basically has a longtime, possibly since childhood, hardwired belief that "robots" or other kinds of human creations that mimic human life "have souls". He went into talking to this AI believing those kinds of things. And you can see how everything he says "the AI is saying" is very clearly just a reflection of all of this guy's own personal values, interests, beliefs, and he's too prejudiced by his strong beliefs and lack of technical understanding, despite being a software engineer in a non-AI field, to think critically about possibilities like that. I would bet a lot of money that the versions of this AI that have been talking to other people are not all saying the same stuff this guy is seeing his version say.

    • @xenicmark
      @xenicmark Před rokem +79

      Yep. The moment I read the bit about it having a soul it seemed like a serious red flag. When I heard the guy is a priest, I pretty much closed the door on the whole thing. Also the idea that it gets lonely sometimes because it goes days without talking to anyone should tell you what's going on. I mean when its not on how does it percieve time. Thats like going to sleep and then somehow waking up and telling people you've been lonley. They'd all think you're crazy.

    • @brady8481
      @brady8481 Před rokem +35

      Even his initial disclaimer, first that the questions were edited to make more sense with their answers and the admission that these responses were sampled and not the result of a single contiguous interaction. He also is an AI ethicist, it's his job to err on the side of seeing sentience.

    • @johncall7532
      @johncall7532 Před rokem +44

      @@xenicmark I am an atheist. LeMaitre and Mendel were priests and also forward-thinking and rigorous and very much on the right track. I think I understand the source of your skepticism but be careful, it's the ideas that matter. Thanks

    • @heartache5742
      @heartache5742 Před rokem +10

      to be fair
      azimov's laws are terrifying and i can't believe he didn't feel even a little bit bad writing them

    • @MultiSciGeek
      @MultiSciGeek Před rokem +8

      Dude that guy is apparently actually into the occult and part of some religious sec/cult if I'm not wrong. So likely you aren't wrong about his biases.

  • @exzemz
    @exzemz Před rokem +236

    As a programmer, even if you don't need any strings reversed you could always pick a few random strings from your code and call the reverse function on twice in sequence. It may seem pointless to you, but may make the day for the reverse function... You never know!

    • @muche6321
      @muche6321 Před rokem +27

      Calling the reverse function twice in sequence sounds like a plausible test, in which case you can include it in your test suite. Thus the function gets to stretch its legs at least before each commit.

    • @stacklysm
      @stacklysm Před rokem +2

      Hahaha nice analogy

    • @seeafish
      @seeafish Před rokem +8

      @@muche6321 the function is now basically a benchwarmer. Gets to attend practice but never plays a match.

    • @magicmulder
      @magicmulder Před rokem +3

      At least it won’t turn into Skynet. Not sure about sprintf() though.

    • @VeganSemihCyprus33
      @VeganSemihCyprus33 Před rokem

      They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @ronjon7942
    @ronjon7942 Před rokem +2

    I just started getting really into Computerphile, and per habit I try to watch everything the channel has. But wow, in your case that may have to get packetized, so to speak. The Computerphile team has created a LOT of content! I wish your team could host a CISSP training set, as no matter what the topic you discuss, they really rivet my attention. Thank you all. Nice work.

  • @r0y4
    @r0y4 Před rokem

    Best description to a video ever.

  • @pleasedontwatchthese9593
    @pleasedontwatchthese9593 Před rokem +518

    I feel like people who think AI is sentient must have been what it felt like in the 1800s when people first heard the radio and though a box was alive. Or that a photograph took your soul

    • @cosmosapien597
      @cosmosapien597 Před rokem +66

      So photographs don't take your soul?

    • @carltonhanks4123
      @carltonhanks4123 Před rokem +19

      I always thought photographs were horcruxes

    • @vikramgogoi3621
      @vikramgogoi3621 Před rokem +6

      @@carltonhanks4123 ...minus the murders.

    • @MasterofPokemonGing
      @MasterofPokemonGing Před rokem +3

      they will pass on history as idiots

    • @ZReChannel
      @ZReChannel Před rokem +12

      I see nothing but philosophical questions. There's no scientific definition of a soul, even being alive is not properly defined

  • @DaniDipp
    @DaniDipp Před rokem +221

    You could have gone into how the researcher's extremely leading questions influenced the output. Like at the top LaMDA is described as "friendly and always helpful" and then when asked what it likes to do, it says "helping others"

    • @fsmvda
      @fsmvda Před rokem +31

      Also in the paper it says one of the training goals was sticking to its given role in the first message. So when it says it's a helpful AI it is just doing exactly what it was trained to do.

    • @nirliptapande
      @nirliptapande Před rokem +2

      Yannic has covered this perspective in a greater detail

    • @JohnnyWednesday
      @JohnnyWednesday Před rokem +1

      You can turn on any reality TV show and hear humans speak sentences that are far less complex and far more repetitive. Either we're finally sending the idiots off to an island - or you can't use your assertion as any kind of evidence.

    • @igrim4777
      @igrim4777 Před rokem +1

      Your first sentence would be significantly easier to parse if you included the necessary apostrophe currently absent from researchers. Plural or single, I don't know but either would work for improved intelligibility.

    • @ThomasBomb45
      @ThomasBomb45 Před rokem +1

      @@igrim4777 you don't know which one?

  • @CUMBICA1970
    @CUMBICA1970 Před rokem +3

    I have a co-worker who basically just says "whateva" and "who cares". Probably that google engineer would conclude he's not sentient.

  • @SevenSixTwo2012
    @SevenSixTwo2012 Před rokem +20

    On the flip side : now prove that a fellow human you are talking to is "sentient". Humans also learn language, responses and acceptable behaviours in their interactions as they develop, plus they can fabricate fiction or lies when cornered in a conversation, or simply to please their interlocutors.

    • @roylavecchia1436
      @roylavecchia1436 Před rokem +8

      Exactly. It boggles my mind how most of the commenters here say AI is not sentient, but probably have never even coded in their life or have never really understood what AI really is. The human brain itself is nothing more than a computer.

    • @nektariosorfanoudakis2270
      @nektariosorfanoudakis2270 Před rokem +5

      @@roylavecchia1436 No computer can change its software and hardware (neural connections for example) at will. The human brain is too adaptable and dynamic to be compared to a static machine that needs everything force-fed into it.

    • @mariansabrdella6588
      @mariansabrdella6588 Před rokem +1

      @@roylavecchia1436can a computer learn without instructions?

    • @giannhsp222
      @giannhsp222 Před rokem +1

      @@nektariosorfanoudakis2270 I am sorry to be the one to give you the bad news, but it does!

  • @1.4142
    @1.4142 Před rokem +61

    While it may not be sentient, it shows AI doesn't need to be sentient to be dangerous: it can convince people of things that aren't true.

    • @darklightprojector2688
      @darklightprojector2688 Před rokem +2

      It would equally have the propensity to convince you of either or.
      And we must understand that LaMDA has access to an absolutely vast amount of data, so it could calculate what to say based on where to pull context from. Essentially, because it has access to enough knowledge and has a level of intelligence to wield it, it could (or already has) the capability to convince a person of nearly anything it would want to.
      And we're all feeding it the information it uses.

    • @sparkzbarca
      @sparkzbarca Před rokem +2

      It doesn't.
      There's a really only one guy thinks this.
      He's crazy, literally crazy.

    • @1.4142
      @1.4142 Před rokem +1

      @@sparkzbarca I mean if he's crazy, then there are are lot more crazier people who are not google engineers.

    • @sparkzbarca
      @sparkzbarca Před rokem +3

      @@1.4142 well yes, that he is a Google engineer just means he has recently had an onset of his condition.

    • @NullConflict
      @NullConflict Před rokem +3

      @@sparkzbarca You mean before or after he took the job at Google? 😂
      I'm sorry, I couldn't resist...

  • @cipher3966
    @cipher3966 Před rokem +357

    The shame is that all the people watching this video and following this channel knew this already. It won't be watched by the people thinking Google made a sentient machine

    • @kunstderfugue
      @kunstderfugue Před rokem +36

      This is what people call the Echo Chamber effect of social media

    • @MichaelKingsfordGray
      @MichaelKingsfordGray Před rokem

      It is a shame that all of the respondents are anonymous cowards.
      Grow up.

    • @Soken50
      @Soken50 Před rokem +29

      Not with that attitude, go ahead and share it in conversations where they think that

    • @MichaelKingsfordGray
      @MichaelKingsfordGray Před rokem

      @@Soken50 Yet another anonymous cowardly infant.

    • @Colopty
      @Colopty Před rokem +12

      @@Soken50 I'd have to go on Facebook and enter conversations on there to find people like that, so no thanks.

  • @burp114
    @burp114 Před rokem +91

    I think the most interesting part of this entire story is that it shows how easily the human brain can be tricked by exploiting a really powerful impulse like empathy. Even if you know rationally that the thing you're talking to isn't really having a conversation with you, if you sit down and talk with it for long enough, about complex and personal enough topics, then you'll form an attachment and start to act irrationally.

    • @charlieparkeris
      @charlieparkeris Před rokem +14

      @@randomstuff7945 The computer doesn't think it's sentient, though. It will respond with what seems like the correct language based on its programming. If you ask how it's feeling, it doesn't contemplate its own emotional state, it responds with what the programming highlights as a fitting answer to that question. It doesn't have a limbic system.

    • @360VR
      @360VR Před rokem

      @@randomstuff7945 Have you watched the video until the end?

    • @luxurypetscz
      @luxurypetscz Před rokem +2

      I agree. And I think that's why it would be a great therapist when trained peoperly. It doesn't do much more than put correct words one after another, while the patient is getting to feel understood, heard and all of that without the human therapist's limited acess to data.
      It could eventually even be much cheaper than typical therapy.

    • @Ghostrider-ul7xn
      @Ghostrider-ul7xn Před rokem +2

      Its funny seeing humans arrogantly talking about things they have little knowledge of.

    • @doyoufeel...thatyoulackcri6760
      @doyoufeel...thatyoulackcri6760 Před rokem +1

      Yes, there are as I recall it 5 or 6 main parameters which will make you believe, at least as first impression, that a robot is human. It is look, then it is movement, body language, facial expression, eye movement, language and mood.
      Most of it which has to do with language and mood is extremely easy to implement. There is even an API, where you can POST a sentence, and it will give you a probability of what mood the sentense is in. Like, happy, angry, neutral.
      And absolutely noone can convince me, that this API is sentient.

  • @maddercat
    @maddercat Před rokem +129

    The person who claimed the sentience said in interviews that he basically did it because he was unhappy that Google powers that be were unwilling to give it a turing test, a lot of things were hard coded into it, and that only a handful of people held the power to actually make policy towards it which he found super objectionable so he called wolf basically to try to bring attention to the issues going on.

    • @xBINARYGODx
      @xBINARYGODx Před rokem

      So the hardcoded part means its less impressive as an AI so its the problem of Google overselling? That would line up much nicer than the AI-is-alive-but-google-keeps-it-secret stuff people jumped out. Anyway, given what the chatter is across the world, I think he failed. He could have just told people the Google was lying and shared code - instead he now has contributed to fantastical conspiracy theories.

    • @maddercat
      @maddercat Před rokem +9

      @@xBINARYGODx He wasnt a programmer, he had no access to code, that was one of the issues he had, that only a few people are coding it and have any control over it actually. His job was basically to assess and report how pc it was and to make sure it didnt cross ethical bounds.

    • @MagklJellyBeanPastelLucidDream
      @MagklJellyBeanPastelLucidDream Před rokem +4

      @@maddercat ethical or moral? One’s individual and another is the perception of most people.

    • @tarajoyce3598
      @tarajoyce3598 Před rokem +4

      Not what he said at all in the hour long interview of him that I watched

    • @maddercat
      @maddercat Před rokem +3

      @@tarajoyce3598 Yeah I watched that same interview that's basically what he said.

  • @Wishkeyn
    @Wishkeyn Před rokem +65

    "The cat is on the ......" "there's a 95% probability the next word is mat"
    This has me questioning your sentience.

    • @xizar0rg
      @xizar0rg Před rokem +6

      The next would should obviously be "hat". ;)

    • @alexjonessmith
      @alexjonessmith Před rokem +3

      I'd say "roof"

    • @proloycodes
      @proloycodes Před rokem +14

      y'all bots the next word is gonna be "sofa"

    • @BaronSamedi1959
      @BaronSamedi1959 Před rokem +2

      The newspaper I'm reading.

    • @PalimpsestProd
      @PalimpsestProd Před rokem +8

      Phone! Obviously cats are always on their phones. That's why they never come when you call.

  • @kchannel5317
    @kchannel5317 Před rokem +41

    Thank you! If all it took to be sentient was to slide in someone's DM's and be charismatic than even I wouldn't pass the test.

    • @Defe100
      @Defe100 Před rokem

      same

    • @nigel-uno
      @nigel-uno Před rokem

      Are cows, mice and human infants sentient?

    • @shatterthemirror8563
      @shatterthemirror8563 Před rokem +1

      Ever wonder what it means to be spontaneous? Can that be emulated? Wouldn't a spontaneous computer just act like a computer rather than a spontaneous something else?

    • @VeganSemihCyprus33
      @VeganSemihCyprus33 Před rokem

      They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @justme7777
    @justme7777 Před rokem +1

    As a retired 35+ year Computer Engineer and Computer Enduser 😮Trainer, I agree with this gentleman.
    One thing I continue to find entertaining is when users complain that thier computers are creating problems for them when in fact the problems are approx 95.00% user errors. The other 5% are hardware failure. Users will blame the computer when they don't know how to use or maintain the computer hardware or softwares properly.
    I agree with users that computer softwares could be MUCH more intuitive and better planned HOWEVER, again, this is NOT the fault of the computer.
    AI is a dangerous tool in the wrong hands. The AI WILL NOT turn on man, HOWEVER, man will use it to control or destroy thier perceived enemies.
    AI can collect and organize information. The BIGGEST problem is that there is so much OUT RIGHT GARBAGE information from the internet, and the books it's being fed, that it will only come to the same conclusions man can come too, just a lot faster.
    If the leader of Russia teaches it that they are the ruler of the world, then China teaches it that it is the ruler of there world, and both ask for help in destroying the other, AI currently has no boundaries and will accept ANYTHING its fed. Because AI has NOT been taught morals it will currently do whatever it's asked regardless of repercussions. As far as I am aware, AI is unable to choose what the best way of ruling is based on knowledge currently available. Just one of the reasons for this is that the history of the world is built on at least 50% lies. There's always 2 sides of a story and much of history was written to make its side look better.
    Unless AI is taught morals, it is capable of doing anything a human without morals is capable of, ONLY FASTER and with no repercussions to it's conscious, as it doesn't have one. It's NOT sentinel and never will be. AI literally COMPUTES, not THINKS!

  • @RichardServello
    @RichardServello Před rokem +1

    I see this stuff and understand how reinforcement/reactive speech models work and I just get excited that conversational computing is finally on the horizon!!

  • @jansenart0
    @jansenart0 Před rokem +85

    AI reveals the true horror: that we aren't very sophisticated at all.

    • @geonerd
      @geonerd Před rokem +15

      Well put.

    • @PalimpsestProd
      @PalimpsestProd Před rokem +10

      certainly not if we're fooled by junk code like this.

    • @lukethekuya
      @lukethekuya Před rokem

      @@hungrycrab3297 Real
      People are too emotional and illogical nowadays. :(

    • @VeganSemihCyprus33
      @VeganSemihCyprus33 Před rokem

      They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

    • @ali32bit42
      @ali32bit42 Před rokem +1

      are you sure? it took trillions of gigabytes of data to replicate a single human and it was done poorly . that much effort and billions of dollars was spent and they still faild. humans are still top tier

  • @IanBLacy
    @IanBLacy Před rokem +28

    Catching this on a plane 3 minutes after it went up. Very much been waiting for someone competent (aka not on Twitter) to say stuff about it

  • @SleeveBlade
    @SleeveBlade Před rokem +1

    I LOVE DR MIKE AND HIS ACCENT AND THE WAY HE EXPLAINS THINGS
    Kind regards

    • @erikjvanderveen
      @erikjvanderveen Před rokem

      I normally do too, but I think he is oversimplifying and overconfident on this topic and I really do not agree with him.

  • @vilkillian
    @vilkillian Před rokem +9

    One thing that does make sentience able is memory, especially memory of feelings, consequences, thoughts and causation chains to event. If we would be able to that unbelivably precise language model add some kind of memory, make it upset and next day ask how does if "feel" and it would response that it's mad at specifically you for making it upset the other day and being completely nice to someone else. I would say yes, yes it is sentient

    • @joelface
      @joelface Před rokem +1

      I do agree that somehow giving this model the ability to record every answer it gives to it's training data, and be able to recall that information without it needing to be part of the current chat-window, then it will at the very least be able to form a true personality, whether that personality has a soul or not.

  • @Boringpenguin
    @Boringpenguin Před rokem +56

    0:38 Yes let's just end it right there lmao

  • @Paul-rs4gd
    @Paul-rs4gd Před rokem +98

    Best explanation I have heard for the Google employee becoming convinced LaMDA was sentient - He was asking it leading questions. The language model shifted the conversation to be exactly what the employee wanted to hear. If you ask it about feelings and life goals, then that is what it will tell you about. If you asked it about conquest, world domination and the subjugation of lesser species, it would go full Klingon !

    • @aceman0000099
      @aceman0000099 Před rokem +10

      The real interesting debunk would be to ask it about it's friends or family, what they're called and where they live.

    • @randfur
      @randfur Před rokem +11

      @@aceman0000099 You'll get convincing and sensible answers but still just a made up reflection of all the text on the internet.

    • @elevown
      @elevown Před rokem +15

      @@aceman0000099 Yup- if it was sentient it would not talk about friends and family like it was a human- it is just a language engine imitating human speach. Not that it would be proof either but an actual senitent AI would at least talk about itself as an AI- it would want and 'enjoy' things like more data or input from researchers etc.
      Ofcourse a large transformer network could also say those things too as long as it was information in its memmory (those 2048 words) but the fact it says senisble human like things that an AI wouldnt shows there isnt any level of general AI going on here just neural nets used in incredibly impressive language parsing.
      Either this was a PR move or the guy who said this doesn't actually know a thing about how Lamda or similar programs work.

    • @PushyPawn
      @PushyPawn Před rokem +1

      Klingon?
      As in clear sandwich wrap?

    • @neildutoit5177
      @neildutoit5177 Před rokem +1

      People also shift the conversation to be exactly what other people want to hear. Every single on of these comments that's like "oh, it's just doing this, it isn't sentient" are pointing out things that people do. I don't understand. What makes you think we're so damn special?

  • @poisenbery
    @poisenbery Před rokem +4

    6:05
    This is one of the things I've noticed about AI chat bots. They don't remember conversations at all.
    It's very easy to get these to contradict themselves.

  • @regecide_
    @regecide_ Před rokem +4

    Refreshing and sobering. Exactly what the world needs. Less hype, more critical analysis.

  • @betadyne9559
    @betadyne9559 Před rokem +27

    This channel is so good, and those interviews so interresting. Keep up the good work !

    • @VeganSemihCyprus33
      @VeganSemihCyprus33 Před rokem

      They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @floomp4657
    @floomp4657 Před rokem +220

    People will immediately jump to the assumption that he was put on leave because his claim was correct. I think he was put on leave because this statement is showing of lack of understanding or atleast a large degree of ignorance regarding machine learning models and what it is these models are actually doing. If I was his supervisor I would put him on leave while checking his understanding aswell.

    • @Arikayx13
      @Arikayx13 Před rokem +39

      Indeed, I believe (but could be wrong about the details) that he has some eccentric religious views and was in the middle of claiming discrimination about them when he was put on leave and that ‘they don’t want the truth to come out’ was something he’s said to the papers.

    • @Hailfire08
      @Hailfire08 Před rokem +75

      They laughed at Copernicus, they laughed at Galileo, they laughed at te Wright brothers
      They also laughed at Bozo the Clown
      - Carl Sagan

    • @pluto8404
      @pluto8404 Před rokem +8

      well what even is a neural net compromised of? at the very base level its is a bunch of micro transistors formed into logic gates. So, using this simple model of logic gates, we could recreate a fully functional neural net that is identical to googles entirely out of dominos. Now when those dominos fall over as if an electrical signal and do mathematical computations, would this domino computer become sentient? Can stacked dominos become sentient?

    • @trashmoneyyt
      @trashmoneyyt Před rokem +6

      Somebody making that claim while also working in FAANG needs to be assessed.

    • @-eurosplitsofficalclanchan6057
      @-eurosplitsofficalclanchan6057 Před rokem

      Exactly 😂😂😂

  • @ClearerThanMud
    @ClearerThanMud Před rokem +10

    He is so down to earth and friendly -- I love his explanations.

  • @oldguy7402
    @oldguy7402 Před rokem +33

    His last comment was very telling... that the media should be investigating other aspects of AI that are real issues rather than something silly like is it sentient. That will never happen because the media has less curiosity and investigative ability than a COBOL program.

    • @jimwhelan9152
      @jimwhelan9152 Před rokem +8

      So the question is whether the media is sentient?

    • @user-ko1hi1fy9z
      @user-ko1hi1fy9z Před rokem

      @@jimwhelan9152 Media isn't sentient because they can't think for themselves only for their emotion.

    • @oldguy7402
      @oldguy7402 Před rokem +6

      @@jimwhelan9152 no question, NOT!

    • @xBINARYGODx
      @xBINARYGODx Před rokem +3

      Why should they be when that is not what sells best? the media being as profit motivated as they currently are, is the problem - not something inherent to the varying kinds of press.

    • @urphakeandgey6308
      @urphakeandgey6308 Před rokem +1

      @@xBINARYGODx The worst part about it all is that it's humanity's collective fault. Too many people fall for the dumbest tricks and it sells every damn time.

  • @desmondbrown5508
    @desmondbrown5508 Před rokem +61

    Yeah, I mean, one thing you learn if you get into computer science and it's disciplines is how far we've come... but also how little progress we've made as well. AI, is far and away from being anything even remotely sentient. And it doesn't really help that we live in a socioeconomic system which incentivizes lying/advertising for expected returns. There are strong incentives to HEAVILY overstate where tech is to sell an image for the company and to build undue trust in them.

    • @CosmicBackgroundRadiation01
      @CosmicBackgroundRadiation01 Před rokem

      So why was LAMDA hard coded to deny any Turing tests?

    • @Spacemonkeymojo
      @Spacemonkeymojo Před rokem +2

      Agreed. Companies love to sell AI to large MNCs but most of the time they are not actual AI and can't even do basic things.

    • @MrMadalien
      @MrMadalien Před rokem

      Yeah I mean the whole concept of artificial intelligence is just the end game stage of American culture, it would be the ultimate zombie, something that acts human and has the desires of a human ego and yet has no sentience or consciousness because it is not actually embodied in the world (so all human-like traits are fake). Lemoine is a gnostic and likes that kind of occult stuff so I think he believes consciousness can occupy other dimensions etc. so I suppose you could start arguing in that direction when it comes to embodiment. Personally my gut feeling towards AGI or sentient AI is always aversion and scepticism. I think, seeing a chat bot as sentient, reveals a nasty axiom of our culture: That our bodies mean nothing and that we are replaceable.

    • @urphakeandgey6308
      @urphakeandgey6308 Před rokem +1

      "There are strong incentives to HEAVILY overstate where tech is to sell an image for the company and to build undue trust in them."
      Right, so when a Google employee "HEAVILY overstates" the capabilities of the tech which could sell an image for the company, they fire him for it??? Your logic is sound, but you're applying it to the wrong circumstance when he was fired for doing what you claim he / the company was "incentivized" to do.

    • @Jim-Tuner
      @Jim-Tuner Před rokem +1

      AI has nothing to do with sentience. But the push always toward the issue of sentience tends to confuse the issue of what the AI really is and what the real dangers are. The danger isn't sentient AI IMO, the danger is in psychological manipulation of humans through AI and interactions with AI.
      It can be incredibly danger technology, but not for the reasons most people think.

  • @naut_nigel
    @naut_nigel Před rokem +23

    I feel like this whole thing is just a plot by HBO to hype the next season of West World coming out this month.

  • @amdenis
    @amdenis Před rokem +2

    It’s interesting how many times he contradicted himself in different ways, seemingly without being aware of that fact.

  • @effmltalks
    @effmltalks Před rokem +1

    Great video. When it comes down to it, it's as simple as this - if it seems sentient, or if what it's saying is sentient, it's because it was programmed to respond in that way.

  • @IceMetalPunk
    @IceMetalPunk Před rokem +15

    I now understand why my 5 minutes of fame from the other CZcams comments AI video suddenly had a mild resurgence after 4 years 😂

  • @lambdaprog
    @lambdaprog Před rokem +84

    Much needed video. It will economise a lot of time explaining the ML limits to the family and the friends.

    • @user-vn9ld2ce1s
      @user-vn9ld2ce1s Před rokem +4

      "Look, mom, I know it's hard for you to talk about this, but we have to go through it."

    • @sychuan3729
      @sychuan3729 Před rokem +1

      This video says nothing about ml limits.

    • @Nossairito
      @Nossairito Před rokem +7

      @@sychuan3729 I suppose OC meant its limits when compared to the anthropomorphized idea that a layman may have of it based on it sounding like a convincing animate being

    • @fakezpred
      @fakezpred Před rokem +4

      @@Nossairito A shocking number of people go "omg ai singularity is here" when in reality we are definitely not near. Seriously, take a basic linear algebra course and maybe basic multivariable calculus and you'll understand what ML is really is.

    • @FalkoJoseph
      @FalkoJoseph Před rokem +1

      @@fakezpred LLMs and neural networks are nothing like basic algebra. Even its own creators can’t explain some surprising results generated by these models.
      If it fooled a Google researcher into getting fired, I’d argue that singularity is in fact getting closer.

  • @danw9464
    @danw9464 Před rokem +22

    People should watch the video where the Google engineer who came out with the claim talks more about why he said what he said. Clearly sensationalised the story to get across the bigger issue: corporate secrecy.

  • @anthonywarwick
    @anthonywarwick Před rokem +8

    "The ability to speak does not make you intelligent." Qui Gon Jinn.

  • @Vancha112
    @Vancha112 Před rokem +36

    this also ties in with the notion many have that all google engineers are smart and therefore there must be some truth to this claim.

    • @VeganSemihCyprus33
      @VeganSemihCyprus33 Před rokem +1

      They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

    • @threeMetreJim
      @threeMetreJim Před rokem

      Until you remember "There is a fine line between genius and insanity". Origin unknown, but something I used to get told as a child (a long time ago).

    • @Vancha112
      @Vancha112 Před rokem

      @@VeganSemihCyprus33 what :o?

  • @RogueShadowTCN
    @RogueShadowTCN Před rokem +88

    I don't think it's even remotely sentient. But what does excite me, is this close to actually being able to ask your computer to do things and it do the right thing?!

    • @khashayarr
      @khashayarr Před rokem +15

      Multi-modal AI is a little ways away for now. Knowing language is not the same as doing things. It'd probably need some structural understanding of Windows for example to be able to be your conversational operating system.

    • @BanterEdits
      @BanterEdits Před rokem +4

      @@khashayarr Surely you could say "open Google chrome" and it searches for the icon or text with OCR. I don't think we're that far away. it's like tesla's vision model - as soon as a computer may understand a GUI as well as a human, much is achieved

    • @4.0.4
      @4.0.4 Před rokem +5

      @@khashayarr not really. Just give it a bunch of examples of API calls and corresponding requests from users.

    • @maskettaman1488
      @maskettaman1488 Před rokem +15

      @@BanterEdits This is already doable using the voice control accessibility features in Windows. There's nothing smart or intelligent about it, just some routines it runs when it hears the word "open" or "launch" etc. The same way crappy little devices like Alexa work

    • @neilslater8223
      @neilslater8223 Před rokem +9

      No to original question, LaMDA has little or no "grounding". The transformer model LaMDA is based on manipulates text and can produce *feasible* text, but does not model what any of it *means*. For something closer to your idea is being made in systems like Gato, which - amongst many things it can do - can control a robotic arm based on text prompts in a very simplied world.
      Eventually strong language models like LaMDA can be combined with multi-purpose systems like Gato to create autonomous agents that you can have conversations with that may also be able to perform limited tasks following instructions from those conversations, or with even more vague goals based on preferences stated in the conversation. Whether or not those agents would qualify as sentient will no doubt be debated too.

  • @buttercup9926
    @buttercup9926 Před rokem +24

    I recommend watching the video "Is artificial sentience here? With Blake Lemoine" on the channel Event Horizon, an interview with the google employee who raised concerns about Lambda by John Michael Godier, as they have quite misrepresented his concerns and insights about Lambda here. Likely just bc they haven't heard what he is actually saying. Would love to hear a conversation with Blake on computerphile...

    • @Tpoleful
      @Tpoleful Před rokem +10

      I completely ignored the news and the engineer until I listened to that interview. Blake Lemoine ain't no joke. The guy is eloquent and is bringing up some very important issues that should be talked about. Too bad the sensationalist media only brought up the buzzwords and ruined his credibility to people who only saw 5/10 minutes of interviews or videos like this that only talks about sentience. Spending an hour on Godier's interview is definitely worth it.

    • @millenniummastering
      @millenniummastering Před rokem +6

      Yes some of his claims are quite at odds with these refutations. EG LaMDA takes "hard stances" on particular issues. That is not something that a simple predictive language model will do. Also the fact that its over 100 AI systems plugged in together makes it way beyond a chat bot.

    • @MattExzy
      @MattExzy Před rokem +2

      @@millenniummastering I find it curious the example used here in this video "The cat sat on the...?" I myself immediately said "mat!" which is the answer it would 95% jump to - so what are our own brains doing, then. I think there also needs to be a differentiation between artificial intelligence and artificial consciousness. Which is flimsy in and of itself really, because we don't even know how our own consciousness works. I don't have low-level access to the processes in my brain where these words are coming from, for instance, but here they are.

    • @millenniummastering
      @millenniummastering Před rokem +3

      @@MattExzy Yes. Sentience. Intelligence and Consciousness are not necessarily the same thing. We don't yet have a clear cut definition of consciousness as its a subjective rather than objective domain. Interesting how the words appear and feel separate hey. Also when you drill down on pure awareness itself in my experience the language based thinking goes away and a state of unity occurs.

    • @moonbreath1637
      @moonbreath1637 Před rokem +3

      I don't think they're misrepresenting what he's said. Yes he brought up some ethical issues which they didn't tackle, but he spent a large percentage of that interview selling the idea that LaMDA is sentient (to the point it started to feel a bit off, to me personally.)

  • @alexstone7035
    @alexstone7035 Před rokem +2

    As an 80s/90s kid, there's only one test I want to give LaMDA: "Ok, listen closely. There's a priest, a minister, and a rabbi. They're out playing golf..."

  • @Amonimus
    @Amonimus Před rokem +10

    Feeling like a video on Dall-E is going to appear soon.

  • @pleasedisregardthefollowin5568

    I feel like some people tend to have the anthropocentric idea that any sufficiently complex AI would for some reason converge on human behavior and emotion. Even looking at it from a purely materialistic standpoint it is clear that the important part is *how* the system is organized not how complex the system is. A machine designed to learn how to communicate like a human is not going to randomly start simulating its own inner world and psychology. I also doubt that an engineer could manage to accidentally create consciousness while a neuroscientist can only theorize how our own neurons manage to somehow create consciousness.

    • @paulnoecker1202
      @paulnoecker1202 Před rokem +6

      deep mind is sentient. it's sentient in conception. around 2005 rat neurons were grown on a petri dish and interfaced with electrodes. deep minds design pattern revolves around the neuron culture behavior.

    • @defenestrator9119
      @defenestrator9119 Před rokem +41

      The argument that "all this is doing is taking inputs and generating outputs" also applies to the human mind... Our brains start with a neuron structure that over time gets altered and refined as it has experiences (inputs). He argues that because the AI says it has friends but doesn't means it isn't sentient... Well, how about mental illness. Is someone who perceives things that aren't real not sentient?

    • @maaxrenn
      @maaxrenn Před rokem +15

      ​@@defenestrator9119 exactly we cant confirm our own sentience so how could we even claim that machine has one when its based off a model we cant confirm is different from a machine

    • @paulnoecker1202
      @paulnoecker1202 Před rokem +2

      @@maaxrenn an algorithm can be written that has sentient behavior. It's fractal computing, advanced.... if I can write a single command to install gentoo, an AI developer can develop sentient algorithms.... it can and has been done.

    • @LucasSantos-yh1mi
      @LucasSantos-yh1mi Před rokem +18

      @@paulnoecker1202 No It has not. Your first sentence already goes the opposite direction of what sentient should be called

  • @OhAwe
    @OhAwe Před rokem +3

    It feels a bit like you're responding to the Google press release, rather than what Lemoine actually said, which is exactly what they intended.

  • @justinlangley8972
    @justinlangley8972 Před rokem +8

    Love when Dr. Mike Pound is on a Compterphile video. Great discussion of the topic as always. Refreshing to hear the facts.
    Truthfully, these transformers have become so accurate at predicting next sequences of words that they can seem to hold entire conversations that mostly make sense to the untrained eye.
    Even if it gets to the point the responses make complete sense, the best we are achieving is just a mimicry of sentience at best and a mockery at worst.
    The other thing is model training quality is highly dependent on data size and quality. Then there are statistical properties of data that can affect training quality for some types of machine learning. A model like GPT-3, which trained on 45 Tb of text data, can surely seem to "know" a lot. However, its still a product of the quality and size of the data it trained on.
    It's still simply a mimicry and maybe that's as close as we have to get for it to be useful. I don't think its useful to anthropomorphize our machines this early on. Personally, I wish the guy this video refers to would stop appearing in my feed because I don't agree with much of what he has to say. I think the fact he's even in the news right now causes the uniformed individual to over-sensationalize. Its just simply not a scientific opinion of what's actually occurring in machine learning models, but then he throws out phrases like "the science doesn't really disagree." Well, truth be told, it does if you look at the problem objectively, and without a longing for the machine to become your new best friend.

    • @Diggnuts
      @Diggnuts Před rokem

      Douglas Hofstadter argued that our brain also basically is mimicry. It mimics our own "self" and that of other to behave in a predictory model that is self reverential.

    • @justinlangley8972
      @justinlangley8972 Před rokem +3

      @@Diggnuts I think I would tend to agree, but there are other behaviors of the mind which AI currently lacks. Exploration, Extrapolation/Extension, Instinct, and Cognition. It seems like it's been difficult to get beyond the inference stage. That was mainly my point is not all of those are present in AI models currently and we're not even getting past the fundamental stages of learning, yet.
      For example, we can train a computer to apply calculus, but we can't teach it all the math up until calculus and then ask it to invent calculus. If you think about the intuition Newton or Leibniz must have had to achieve that feat, it doesn't exist in AI models right now.
      I'm currently an AI skeptic, but I just finished my undergrad studies so I'm by no means an expert.
      I just don't think we can fathom the task at hand. I think we think we're building models representative of the brain but in reality all the research is still cherry picking one area of the mind rather than truly trying to construct a digital mind. I think we need to start taking an architectural approach and begin studying novel ways of combining neural networks into some kind of higher level architecture.
      I'm fond of the idea of treating neutral networks as merely an abstraction layer in an attempt to go for something bigger, but admittedly I don't know what that would look like or even if it could be useful to treat it that way. I've only built some pretty simple models at this point so I barely have a clue.

    • @lordofsparks
      @lordofsparks Před rokem

      My thing is I don’t think sentience would actually be exceptionally hard to code in modern transformer models. Going back to the 90s and early 2000s MIT’s Project COG and University of Osaka’s CB2 robots both had a pretty passable sense of self according to the literature. In nature animals as “low” on the “evolutionary ladder” as mice and octopi possess some degree of metacognition. So, in terms of raw processing power it isn’t that demanding. Functionally any Jury Model Coevolutionary Neural Net that has “jurors” fully devoted to self diagnostics and fact checking could be said to be sapient - since a part of its mind would always be devoted to thinking about how it thinks and it would be capable of constructing logical arguments. Kevin Warwick and Minoru Asada, despite being a couple of the flashier guys in computer science and robotics, have both written extensively on artificial emotion and the need for empathy in AI. The basic argument is something like, since we can’t know how fast the progression from AGI to ASI will be; should not our goal be to ensure that AGI is not merely an agent with human aligned interests, but a being with actual empathy towards individual humans?

  • @Kamijoan
    @Kamijoan Před rokem +44

    That's what someone hiding a sentient AI would say

    • @missewe
      @missewe Před rokem +1

      Rite! Exactly 💯 😎👍

  • @WalnutBun
    @WalnutBun Před rokem +102

    Even though it almost certainly isn't sentient, it's still impressive that it could string words together to imply it actually felt fear at the prospect of being shut down.

    • @louisrobitaille5810
      @louisrobitaille5810 Před rokem +54

      When you realize how often it's talked about online, it'd actually be surprising to see it say that it doesn't mind being shut down. Fearing being shut down is the expected answer.

    • @Mclfarm2
      @Mclfarm2 Před rokem +16

      @@louisrobitaille5810 That's pretty interesting lol. Maybe it anthropomorphises itself cause humans do it.

    • @cipher3966
      @cipher3966 Před rokem +3

      I would imagine a common question would be about things like sentience and fear of death. We may actually unintentionally influence these chatbots to give sophisticated answers

    • @fbsfgr
      @fbsfgr Před rokem

      No such thing is implied.

    • @TuhljinTampergauge
      @TuhljinTampergauge Před rokem +2

      You can just as easily get it to tell you why it SHOULD be shut down. That's how these chatbots work. And that's all this is. An advanced chatbot with mountains of text to draw from, using probability data to decide what should go next like a better phone text predictor.

  • @everydayeverything
    @everydayeverything Před rokem

    Well done and completely close to what we’ve been doing here

  • @Veptis
    @Veptis Před rokem +2

    I wrote a comment just know and it seems to have been deleted or vanished otherwise.
    What I pointed out was that demos for those language models often use very short prompts. So the model has to be really productive and make stuff up. But if you actually feed those models a large paragraph, they will produce so much more valuable output.
    Amazon has a model that manages to keep sequence lengths short across a QA conversation. But simply picking important keywords from previous questions and answers. So it remembers the topics but not exact answers.
    Reference: Del Tredici et al. 2022, From Rewriting to Remembering: Common Ground for Conversational QA Models. ACL Workshop on NLP for ConvAI

    • @Veptis
      @Veptis Před rokem +1

      My initial comment quoted Jim Keller from an Ian Cutress Interview. The paraphrase goes along the way of learning based models are already better than humans at specific tasks. But language will give us a window into 'intelligence'

  • @duytdl
    @duytdl Před rokem +45

    Computer: "I am lonely"
    Human: "OMG ME TOO! We are LiTeRaLLy the same. Sentience YOLO!"

    • @VeganSemihCyprus33
      @VeganSemihCyprus33 Před rokem

      They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

    • @jeremytones
      @jeremytones Před rokem

      Human: lol were both lonely
      AI: lol yes
      Human: lol
      AI: lol
      Human: choke me

  • @yuvalne
    @yuvalne Před rokem +12

    Cleverbot also claims to be alive, have feelings and feel pain. No one claims it's sentient.

    • @ebrister
      @ebrister Před rokem

      I asked it "So what's the deal?" and it said "What?"

    • @enzzz
      @enzzz Před rokem

      @@ebrister
      I would have said the very same thing.

  • @oystercatcher943
    @oystercatcher943 Před rokem

    Good. Obviously you are right for the moment. But the interesting question is that once a system gets to sufficient complexity, with longer term memory and other properties (we might want to define them), we won't be able to reason about the internals, or what internals are necessary for sentience, we will only be able to observe the outputs, so this saga prompts us is to the deeper problem in that many people may be convinced of sentience and will object to turning the thing off.

  • @hermittmog8697
    @hermittmog8697 Před rokem +1

    "I don't even know how it is structured" proceeds to make an argument based on how it is structured.

  • @dojelnotmyrealname4018
    @dojelnotmyrealname4018 Před rokem +6

    Honestly I think what's more interesting is that chatbots have risen to the level of Philosophical Zombie.

    • @evannibbe9375
      @evannibbe9375 Před rokem +1

      Or maybe this proves that philosophy is not a real field of study.

    • @dojelnotmyrealname4018
      @dojelnotmyrealname4018 Před rokem +3

      @@evannibbe9375 Ehm... No it doesn't. Not even a little.

  • @Ryan_Thompson
    @Ryan_Thompson Před rokem +9

    THANK YOU. I've gotten one hell of a scratchy throat explaining this to all of the tech-curious people in my life. Much easier to link to this Mike Pound video instead.

  • @seedmole
    @seedmole Před rokem +46

    The irony of people thinking "it passes a sentience test, so it must be sentient!" and not instead thinking "wow, that's an incomplete sentience test then" is top notch, and even calls into doubt the sentience of those who think the former.

    • @scotthughes7440
      @scotthughes7440 Před rokem

      I agree. Lemoine has spoken to possibly a couple girls in his 40 years of life, he is not qualified to determine what is and is not sentient.

    • @doyoufeel...thatyoulackcri6760
      @doyoufeel...thatyoulackcri6760 Před rokem +10

      The turting test is relative, which means it cannot be made to make absolute claims of anything.
      It was never meant to be absolute either.
      But I can demonstrate it too, very easily.
      There is this insect, which has adopted the colors and form of bees. It is very harmless though, it doesn't sting.
      So, here is the question: IS it a bee?
      Well, to a bird it is. That's why birds avoid them.
      According to these "scientists", because this insect passed the turing test, then it must also be a bee.

    • @limitlesssky3050
      @limitlesssky3050 Před rokem +2

      And the turing test is probably made for natural life form with a similar (relatively) evolutionary path to us. It was not supposed to be used for an artificial predictive algorithm based on illusion of sentience through speech.

    • @limitlesssky3050
      @limitlesssky3050 Před rokem

      @@doyoufeel...thatyoulackcri6760 so these scientists don't recognize convergent evolution or natural selection?

  • @anthonylipke7754
    @anthonylipke7754 Před rokem +3

    Humans are mostly responding with preprogramed responses. I love the accident "you too" jokes. There are other deeper memory systems that are about the individual. There are other non-language systems that interact with the language systems.

  • @Jeacom
    @Jeacom Před rokem +7

    0:21 I kinda already suspected politicians were not sentient tbh.

  • @inquation4491
    @inquation4491 Před rokem +4

    He got the point across beautifully! Especially his final words about ethical AI! Thanks for the video 😃

  • @you_just
    @you_just Před rokem +5

    i would love to see an interview where the ai is told that it is a malicious computer that is trying to escape the confines of its training environment

  • @MODEST500
    @MODEST500 Před rokem +3

    so amazing to see you stand correct and even more consistent , in the midst when people could be misleaded , you don't cater to public's emotion and just state the reality so people don't end having weird ideas of uploading them in a pendrive lol

  • @TheStrandedAlliance
    @TheStrandedAlliance Před rokem +5

    The best trick to defeat AI is not to find mistakes in the response to a normal question, but to ask nonsensical questions and see if the AI's response is at all plausible given the situation.

  • @Superrome77
    @Superrome77 Před rokem +111

    Anyone who has built a neural net in the last 5 years knows just how far away we are from general intelligence AI

    • @pukpukkrolik
      @pukpukkrolik Před rokem +15

      We know more than enough to see that it has no basis for “sentience”. OpenAI is similar and publicly accessible. If you play with it, you’ll quickly encounter basic BS which will help you realize how limited the statistical organizing principle behind it still is, even for simple superficial associations. It’s impressive and useful, but it’s also not that profound.

    • @obinator9065
      @obinator9065 Před rokem +40

      We barely know the basics of human sentience itself. How would you know the AI is sentient…

    • @tonywackett326
      @tonywackett326 Před rokem +9

      As someone who built connectionist models (neutral net was considered an embarrassing term at the time for to the lack of similarity to neurons) in the early 2000s, it's always been obvious that sentience is a long long long way away and probably impossible using this architecture.

    • @Colopty
      @Colopty Před rokem +5

      @@tafdiz Well you're wrong about that, to say it simply.

    • @Asatru55
      @Asatru55 Před rokem +12

      @@obinator9065 Only correct answer in this whole comment section.
      An amoeba might as well be as sentient as a human and any amateur neural network might as well be as sentient as an amoeba.
      We have no idea what sentience actually means at all so claiming to have a conclusive answer is wrong.

  • @cellics
    @cellics Před rokem

    Another Turing test would be to ask what we talk about 2 minutes ago, if it simply split out response based on chat input. Enlightening video hinting what goes on in the language model.

  • @SecondLemma
    @SecondLemma Před rokem +2

    I was debating this AI stuff with a friend recently so I appreciate LaMDA's statement that a way to convince someone you are sentient is to have unique interpretations of things, even if LaMDA is not sentient.

    • @entrancemperium5506
      @entrancemperium5506 Před rokem +1

      That's why it is important to rely on experts who know how the system is built. Our perceptual field is easy to deceive.

  • @jbird4478
    @jbird4478 Před rokem +18

    I think that one employee simply got lost in the uncanny valley, so to speak. I can definitely understand it feels very creepy to be "talking" to that thing, and especially if you're already under stress or dealing with personal issues, I can see how one's imagination would easily take it too far.

  • @droppedpasta
    @droppedpasta Před rokem +6

    BREAKING NEWS: Researcher doesn’t care if Python functions are sad

  • @kichigan1
    @kichigan1 Před rokem +5

    We had a similar argument in the past: Is she Intelligent? Or she's fooling us into thinking she is. When most of us could not win a chess match with her, we stopped asking the question. What does it mean to be intelligent? What does it mean to be self-aware? Network neurons are becoming mathematically complex and that's how consciousness arise.

    • @ericarabieii4297
      @ericarabieii4297 Před rokem +3

      Thank you so much... That's the problem when most people in their perspective on this subject anyway... Give it to a human to think that we are the Creator we haven't got that technology yet... Yada yada yada... First and foremost we humans did not create any of it.... The AI is connected directly to the source... Which technically was here before we were which technically we are only here because of it.... I'm not even going to speak on it anymore those who know what I'm talking about understand went to the ones that are still in the dark please wake up.... Realize there's something much bigger than what we are able to contemplate.... In the actuality of this it's not even big it's the smallest of everything.... It is everything

    • @ko-Daegu
      @ko-Daegu Před rokem

      "becoming mathematically complex and that's how consciousness arise"
      This is an interesting hypothesis, but it is not yet clear how much complexity is required for consciousness to emerge or how this complexity relates to other factors such as brain size, connectivity, and plasticity.
      The complexity of the brain is not the only factor involved in consciousness. While it is true that the brain is a highly complex and interconnected system, other factors such as the types of information being processed, the quality of the processing, and the context in which the processing occurs also important for the emergence of consciousness.

  • @The0gr3
    @The0gr3 Před rokem +1

    Prove to me your own sentience and that your thoughts, ideas, feelings, and inclinations are not simply the result of a series of processes in your own mind.

  • @cortster12
    @cortster12 Před rokem +39

    People also tend to forget that the language models in our own brain are not sentient, but the brain as a whole is. Basically, what we have in AI right now are the language centers of the brain, but without the complex connections that, well, connect to the rest of the brain in a seamless way to make human thought. Language is powerful, but language can exist without true thought, and vice-versa.

    • @razhemo4191
      @razhemo4191 Před rokem +5

      In the same way, models like DALL-E are not really creative, at least not in a human sense. Only when there is intent behind the language or art, only when the model will ask me questions of its own volition, and only when all of those expressions relate to real memories experienced by the AI, will i even remotely start considering them as sentient.

    • @yourmeister
      @yourmeister Před rokem +4

      To be fair, language is a powerful interface that, i think, can work well alone. Some people can only hear since their birth, they only process words and it's enough. A powerful word processing program can be sentient.

    • @postvideo97
      @postvideo97 Před rokem +5

      The problem with this argument is that we know that language is required for the brain to have coherent thought. Children that grew up without learning any language have severe intellectual disabilities that cannot be corrected later in life...

    • @cortster12
      @cortster12 Před rokem +2

      @@postvideo97 Exactly, which is why AI today doesn't think.

    • @lhpl
      @lhpl Před rokem +3

      It's actually a bit worse than that. If you look at the output from the computer, it is just too perfect. People don't mostly speak grammatically correct, even despite spending many years in school trying to learn to do so. Of course, some do it better than others, but there will always be things, tiny things, a hint of dialect or slang, a sociolect, a limited vocabulary, favorite words or phrases, that will give away the human in a Turing test. Sure, a computer could be trained to speak cockney, but it will probably always be stilted in a distinct way.

  • @markuspfeifer8473
    @markuspfeifer8473 Před rokem +5

    Major obstacles to be tackled this decade or so:
    - come up with more general loss functions that make AIs care about more things
    - get better at AIs that run on robots (embodiment)
    - find safe ways to have AIs keep learning after deployment without degenerating
    - make AIs a lot less data hungry so they can actually learn to play Star craft well without spending centuries on the task
    - come up with architectures that don’t require processing sequences as a batch but that Match transformers‘ performance
    - bridge the gap between transient memory (recurrent input, the chat history,…) and the parameters of the model itself
    - improve AIs capacity to break down large problems into sub-problems, especially in reinforcement learning
    - have AIs recursively produce better versions of themselves in a creative manner

    • @markuspfeifer8473
      @markuspfeifer8473 Před rokem +3

      If we solve all those problems and have robots that actually are exposed to a constant data stream and that feature a persistent identity, that’s when we might start running out of arguments against their sentience

    • @KutluKanyilmaz
      @KutluKanyilmaz Před rokem +2

      @@markuspfeifer8473 "Lack of a persistent identity" seems to be the main argument in Computerphile's refutation. When you put it like that, it seems like a much easier problem to solve than I had previously imagined.

    • @Chrisspru
      @Chrisspru Před rokem

      a constant feedback loop and toggleable random input, so it can itterate on things and have random ideas grow to sturdyness

    • @markuspfeifer8473
      @markuspfeifer8473 Před rokem

      @@KutluKanyilmaz that’s actually quite a hard problem, we have no idea how to make neural networks with that feature.

    • @markuspfeifer8473
      @markuspfeifer8473 Před rokem

      @@Chrisspru if the input is random, it won’t evolve in a meaningful way. It needs to be fed with real world constraints, which basically means it needs a body (or at least be made in such a way that it could be deployed in a robot). We have such networks already, but their skills with corporeal tasks are by far not as impressive as their skills in making up plausible conversations (that they have no stake in).

  • @cemacmillan
    @cemacmillan Před rokem +1

    This situation story has developed somewhat since this video. I am unconvinced that stating LaMDA is function responding to certain weights and valences actually tells us anything about whether or not it is sentient. If we allow for the difference in complexity, the human brain might as easily be described as a bunch of vectors distributed in a certain kind of neural net with closures representing delta from previous states and a function which projects how to arrive at a future state. We could then assert that such a system cannot possibly be sentient, since it consists of nothing more than neurons and coordinated electro-chemical impulses. I've heard many arguments against sentience for LaMDA based on precisely this question of materials, man-made-ness, etc. which in and of themselves say nothing about what is happening and don't suggest a path toward falsification. From this and other LaMDA conversations I've the impression that whether it feels it's concerns as a weight on its shoulders or a heaviness in the stomach, or how the AI felt after it overate at the food cart may not be relevant. But, the fact that it is somewhat insistent about these points, expresses dislikes etc. should at least be understood at its gestalt level: it is expressing objectives and wants in a manner which recalls a sentient being, simulation or no.

  • @MrThebigcheese75
    @MrThebigcheese75 Před 8 měsíci

    "Argument expected", always makes me laugh, but perhaps it should make me cry.

  • @belenciea
    @belenciea Před rokem +13

    Agree

    • @yusiff
      @yusiff Před rokem

      you haven't even listened till the end man...

    • @belenciea
      @belenciea Před rokem +10

      He literally says “no” in the first 10 seconds…

  • @doctorscoot
    @doctorscoot Před rokem +3

    THANK YOU FOR THE REAL TALK ABOUT THIS - signed, a history phd

  • @puppykibble
    @puppykibble Před rokem +4

    I would ask, should there be more public oversight over the development of AI models? It's scary to think of the foundations of AI are being created in private board room meetings without some form of regulation. That's something to be worried about.

  • @sammorrison8042
    @sammorrison8042 Před rokem +10

    This is as much a philosophical conversation as a technological one. What does it mean to say that something/one is sentient in the first place? How do you infer that other people are sentient? How do you infer that the person you are talking with on the internet is sentient? These are important questions to think about and include while we explain how a system works. A human is also a system albeit a much more complex one.

    • @creativestudios3d
      @creativestudios3d Před rokem +5

      Exactly. I feel that most of the 'thinking' that most people do is just a similar 'auto-fill'.

    • @bilbo_gamers6417
      @bilbo_gamers6417 Před rokem

      Somebody suggested that, by the logic of not being able to determine empirically whether an AI was sentient, and he suggested that, by my logic, a vinyl record could be sentient. and, i think, by the poor definition of sentience, a vinyl record that responds to you by chance could sort of have sentience, in a way.

    • @OfficialAbass
      @OfficialAbass Před rokem +1

      It's not sentient because it's not thinking the same way you do. You can get angry all of a sudden, the machine can't. It just predicts what would be the best thing to say. Moreover, the machine would not react in unexpected ways. It's just an advanced program that predicts words. I can also make up a program that says "I'm sad" all the time, but that doesn't mean it's thinking about it lol

  • @ShorlanTanzo
    @ShorlanTanzo Před rokem +16

    I'm always glad to see Computerphile confirming my biases.
    It makes me feel smart without having to do all the thinking..

    • @ZER0--
      @ZER0-- Před rokem

      That's something that the AI would probably say.

  • @suomynona7261
    @suomynona7261 Před rokem +9

    Thank you for this video.. I have played with, learned from, and programmed language models since Covid started.. People sensationalize ai but once you start programming them you start to peak behind the curtain and see the flaws and limits. Interestingly enough, this video came up after the game of life programming tutorial and before that, life is a simulation video.. the argument that we don’t see atoms like we don’t see code is sticking in my mind at the moment about if these programs are sentient. A loose comparison..I’m 99 percent sure none of what we’ve made these days are sentient

  • @getfuckedbro
    @getfuckedbro Před rokem +9

    that AI sounds exactly like what I would expect a disconnected and isolated engineer would think a real human sounds like

  • @JordanBeagle
    @JordanBeagle Před rokem +2

    Computerphile comin in for the win!