Geoffrey Hinton | Will digital intelligence replace biological intelligence?

Sdílet
Vložit
  • čas přidán 4. 06. 2024
  • The Schwartz Reisman Institute for Technology and Society and the Department of Computer Science at the University of Toronto, in collaboration with the Vector Institute for Artificial Intelligence and the Cosmic Future Initiative at the Faculty of Arts & Science, present Geoffrey Hinton on October 27, 2023, at the University of Toronto.
    0:00:00 - 0:07:20 Opening remarks and introduction
    0:07:21 - 0:08:43 Overview
    0:08:44 - 0:20:08 Two different ways to do computation
    0:20:09 - 0:30:11 Do large language models really understand what they are saying?
    0:30:12 - 0:49:50 The first neural net language model and how it works
    0:49:51 - 0:57:24 Will we be able to control super-intelligence once it surpasses our intelligence?
    0:57:25 - 1:03:18 Does digital intelligence have subjective experience?
    1:03:19 - 1:55:36 Q&A
    1:55:37 - 1:58:37 Closing remarks
    Talk title: “Will digital intelligence replace biological intelligence?”
    Abstract: Digital computers were designed to allow a person to tell them exactly what to do. They require high energy and precise fabrication, but in return they allow exactly the same model to be run on physically different pieces of hardware, which makes the model immortal. For computers that learn what to do, we could abandon the fundamental principle that the software should be separable from the hardware and mimic biology by using very low power analog computation that makes use of the idiosynchratic properties of a particular piece of hardware. This requires a learning algorithm that can make use of the analog properties without having a good model of those properties. Using the idiosynchratic analog properties of the hardware makes the computation mortal. When the hardware dies, so does the learned knowledge. The knowledge can be transferred to a younger analog computer by getting the younger computer to mimic the outputs of the older one but education is a slow and painful process. By contrast, digital computation makes it possible to run many copies of exactly the same model on different pieces of hardware. Thousands of identical digital agents can look at thousands of different datasets and share what they have learned very efficiently by averaging their weight changes. That is why chatbots like GPT-4 and Gemini can learn thousands of times more than any one person. Also, digital computation can use the backpropagation learning procedure which scales much better than any procedure yet found for analog hardware. This leads me to believe that large-scale digital computation is probably far better at acquiring knowledge than biological computation and may soon be much more intelligent than us. The fact that digital intelligences are immortal and did not evolve should make them less susceptible to religion and wars, but if a digital super-intelligence ever wanted to take control it is unlikely that we could stop it, so the most urgent research question in AI is how to ensure that they never want to take control.
    About Geoffrey Hinton
    Geoffrey Hinton received his PhD in artificial intelligence from Edinburgh in 1978. After five years as a faculty member at Carnegie Mellon he became a fellow of the Canadian Institute for Advanced Research and moved to the Department of Computer Science at the University of Toronto, where he is now an emeritus professor. In 2013, Google acquired Hinton’s neural networks startup, DNN research, which developed out of his research at U of T. Subsequently, Hinton was a Vice President and Engineering Fellow at Google until 2023. He is a founder of the Vector Institute for Artificial Intelligence where he continues to serve as Chief Scientific Adviser.
    Hinton was one of the researchers who introduced the backpropagation algorithm and the first to use backpropagation for learning word embeddings. His other contributions to neural network research include Boltzmann machines, distributed representations, time-delay neural nets, mixtures of experts, variational learning and deep learning. His research group in Toronto made major breakthroughs in deep learning that revolutionized speech recognition and object classification. Hinton is among the most widely cited computer scientists in the world.
    Hinton is a fellow of the UK Royal Society, the Royal Society of Canada, the Association for the Advancement of Artificial Intelligence, and a foreign member of the US National Academy of Engineering and the American Academy of Arts and Sciences. His awards include the David E. Rumelhart Prize, the IJCAI Award for Research Excellence, the Killam Prize for Engineering, the IEEE Frank Rosenblatt Medal, the NSERC Herzberg Gold Medal, the IEEE James Clerk Maxwell Gold Medal, the NEC C&C Award, the BBVA Award, the Honda Prize, and most notably the ACM A.M. Turing Award.
    srinstitute.utoronto.ca/

Komentáře • 368

  • @JoeSanchec
    @JoeSanchec Před 4 měsíci +96

    Professor Hinton's best speeches, his off-campus business speeches are too colloquial and his speeches to computer science students are too specialized, but his speeches to the average University of Toronto student are a perfect blend of both!

    • @RonKJeffries
      @RonKJeffries Před 3 měsíci

      😊😊😊😊😊😊😊😊😊😊

    • @RonKJeffries
      @RonKJeffries Před 3 měsíci

      😊😊😊😊😊😊😊😊😊😊😊😊😊😊😊😊😊

    • @RonKJeffries
      @RonKJeffries Před 3 měsíci

      😊😊😊😊😊😊😊😊

    • @RalphDratman
      @RalphDratman Před 2 měsíci

      You suffer pain from Hinton's extreme days --- and pleasure from his moderate behavior.

  • @gusbakker
    @gusbakker Před měsícem +2

    This lecture is gold. He manages to explain complex topics in simple terms without getting overly technical

  • @jj5jj5
    @jj5jj5 Před 3 měsíci +16

    42:15 it totally clicked for me in this section: LLMs do appear to have understanding because they’re not just encoding a bunch of string predictions, they’re encoding concepts (features) and their relationships… which sounds basically like human learning/understanding.

    • @anomitas
      @anomitas Před 3 měsíci +2

      Nope

    • @flickwtchr
      @flickwtchr Před 3 měsíci +8

      @@anomitas Ah, such a helpful retort bro.

  • @penguinista
    @penguinista Před 2 měsíci +4

    I appreciate Professor McIlraith informing the audience that their questions would end up being posted online, as the presentation was being filmed.
    That is admirably considerate and conscientious. Some people might not want to ask a question if that means they will be on the net.
    It should be standard, but don't remember ever hearing someone do that before - and I listen to a lot of lectures online with audience questions at the end.

  • @user-ql4nr9ll9n
    @user-ql4nr9ll9n Před 4 měsíci +36

    Talk starts at 7:24

    • @divyanshsh
      @divyanshsh Před 3 měsíci

      Thanks

    • @canobenitez
      @canobenitez Před 3 měsíci +2

      savage..

    • @daniel7587
      @daniel7587 Před 3 měsíci +1

      Because Hinton does need no introduction

    • @tux1968
      @tux1968 Před 2 měsíci +1

      What a tedious introduction. Thank you.

  • @jaanireel
    @jaanireel Před 4 měsíci +55

    00:09 Dr. Geoff Hinton's unwavering conviction in artificial neural networks for machine learning
    03:01 Dr. Hinton's pioneering work in deep learning revolutionized visual recognition software.
    08:26 Digital computation separates hardware from software for immortality.
    10:38 Analog computation offers low-power, efficient parallelization
    15:13 Distillation allows transfer of knowledge between digital architectures.
    17:33 Digital computation has efficient way of sharing knowledge among different agents.
    21:50 Language models use powerful statistical methods for autocomplete.
    23:55 GPT-4 learns through interactions between feature activations of words.
    27:56 Memories are reconstructed from stored weights, leading to inaccuracies.
    29:58 Progress of chatbots and neural net language models
    34:16 Using relational data to train a neural net for capturing knowledge in family trees.
    36:41 Back propagation algorithm for neural networks
    40:44 Neural net models learn interactions like rules captured from the domain.
    42:51 Evolution of language models and transformers in natural language processing.
    47:00 Context and interactions determine word meanings
    48:53 Digital intelligence can efficiently accumulate and share knowledge.
    53:17 Super-intelligences will seek more power and be adept at manipulating people.
    55:15 The rise of super-intelligences poses significant threats and potential worst-case scenarios for humanity.
    59:56 The understanding of mental states is crucial for perception
    1:02:03 Digital intelligence can exhibit subjective experiences similar to humans.
    1:06:07 Consciousness involves subjective experience and self-awareness.
    1:08:38 Artificial intelligences may compete in an evolutionary battle, but human cognition may secure our place as interesting conversational partners.
    1:13:41 Digital intelligence could be trained to develop different forms of intelligence.
    1:16:04 Digital intelligence poses risks if used irresponsibly
    1:20:22 Evidence suggests LLMs may not truly understand.
    1:22:12 Digital intelligence uses compression to understand and encode vast amounts of text.
    1:26:16 Researching human brain cells for low power computation.
    1:28:27 Open sourcing powerful models may lead to security risks.
    1:33:05 Concerns about the impact of superintelligent AI on human society.
    1:34:53 Digital intelligence can potentially develop without interacting with the real world.
    1:38:40 Playing Scrabble doesn't require speaking French
    1:40:50 Encourage students to get good at using digital intelligence
    1:45:27 Digital intelligence can understand more data and may be better at figuring out how things work.
    1:47:29 Scaling up existing techniques can make digital intelligence smarter without the need for fundamental breakthroughs.
    1:52:00 Digital intelligence will evolve software engineer roles with fewer individuals needed.
    1:54:15 Research on using language to make distillation more efficient

    • @jitendratiwari6886
      @jitendratiwari6886 Před 4 měsíci

      thanks brother

    • @CodexPermutatio
      @CodexPermutatio Před 3 měsíci

      Many thanks!

    • @deniskhakimov
      @deniskhakimov Před 3 měsíci +6

      Sir, are you an advanced AI model or just a good person? 😊

    • @jaanireel
      @jaanireel Před 3 měsíci

      @@deniskhakimov A good person using Advance ai model to help people understand the video more

    • @jaanireel
      @jaanireel Před 3 měsíci +1

      @@deniskhakimov thankyou

  • @prodrectifies
    @prodrectifies Před 3 měsíci +4

    its always amazing listening to geoffrey hinton

  • @briancase6180
    @briancase6180 Před 3 měsíci +4

    This is, as usual for Hinton, excellent. Thanks!

  • @aakashnigam2243
    @aakashnigam2243 Před 3 měsíci +10

    One of the best lectures by Proff. Geoff Hinton.

  • @landisbritt298
    @landisbritt298 Před měsícem +1

    Dr. Hinton sparked my aspirations for AI. I have much to learn, but I will study every and all things about it.

  • @Astroidboy.
    @Astroidboy. Před 3 měsíci +1

    Thank you for the Respectful Introduction, your Serve with Grace.

  • @scarlettkersten2928
    @scarlettkersten2928 Před měsícem

    No Daniele, of course you are not the only one who takes a delight in Hintons thoughts, and the beautiful and now and then tongue in cheek - humorous way he expresses himself.

  • @prashantprabhala4260
    @prashantprabhala4260 Před 3 měsíci +4

    Brilliant, honest and treasure of thoughts 👏🏼

  • @solomonmatthews7921
    @solomonmatthews7921 Před 4 měsíci +14

    Great talk, and an unusually good Q&A!

    • @canobenitez
      @canobenitez Před 3 měsíci +1

      for real... Great questions.

  • @velvetsound
    @velvetsound Před 3 měsíci +5

    That was a really great talk and very informative, and also shows an evolution of his thinking over time. I remember studying his work back in the 90’s when I was at university, and I use it every day at work now, and I’m glad he’s taken us all through the AI winter into this new, somewhat scary, world of possibilities.

  • @geaca3222
    @geaca3222 Před 3 měsíci +2

    Great, very informative lecture and Q&A, thank you. I also love the explanation of the early 1985 language model, and interactions between learned features. (actually like the technical remarks too! :)

  • @teemukupiainen3684
    @teemukupiainen3684 Před 3 měsíci +1

    great...thank you so much!!!

  • @nathanfrazer7991
    @nathanfrazer7991 Před 3 měsíci +1

    Interesting lecture, thank you for this Dr. Hinton

  • @jj5jj5
    @jj5jj5 Před 3 měsíci +4

    29:00 this discussion of confabulations and that the human brain does this too is so helpful in understanding what “hallucinations” are and where they come from

  • @shalevlifshitz924
    @shalevlifshitz924 Před 4 měsíci +28

    Really great talk, and an amazing Q&A session! It was a pleasure to attend.

  • @crodcapital
    @crodcapital Před 3 měsíci +1

    Very Insightful! Great Talk and Q&A, I really enjoyed it and learned new perspectives.. Thank you for shearing!!

  • @fabriziocasula
    @fabriziocasula Před 2 měsíci +1

    great Prof. Hinton

  • @prottentogo
    @prottentogo Před 3 měsíci +6

    I wish all q&a were as interesting and well mannered as these

    • @skierpage
      @skierpage Před 3 měsíci +6

      Canadians! From the best part of North America.

  • @unhandledexception1948
    @unhandledexception1948 Před 3 měsíci +6

    A read pleasure to listen to Prof Hinton's talks ...What a brilliant mind... such a interesting and insightful way to explain the comlex in simple terms. ... I wish I attended his lectures when I was in college..

  • @hinze55555
    @hinze55555 Před 3 měsíci +5

    We‘ll look back to this as the gold standard of AI development explanations in a not so far future!

  • @xingsong8506
    @xingsong8506 Před 3 měsíci

    this is gold!

  • @ctoh922
    @ctoh922 Před 7 dny

    Before watching this wonderful lecture and Q&A, I gave the transcript to an llm for a summary and highlights. I got a great and very useful and interesting summary. I just learned how to do this today and I will use this technique a lot. I always watch a video if Geoffrey is in it but for a lot of videos I might be satisfied with a summary, especially if that summary doesn't intrigue me.

  • @albertleedom3373
    @albertleedom3373 Před 3 měsíci +4

    At 66, this is my first response to anything on the internet. You come closer than anyone of what I understand about
    However, I'm not educated traditionally ... you put it together beautifully, and if somebody else has already said this, sorry. Consciousness doesn't matter. What you're saying is that intelligence from Neural Networks is already smarter than we are in the analog and/or digital. Thank You It is the nature of things.

    • @telesniper2
      @telesniper2 Před 3 měsíci +1

      58:42. Ah the old "consciousness is an illusion" trope. Ok, it's illusory to what then? See it's a nonsense statement. And he'd no doubt retort with "oh I meant to do that, I was demonstrating how irrational reason is". Well my reply to that is, if you're going to abandon reason, I happily take your concession of defeat.

    • @vcom2327
      @vcom2327 Před 3 měsíci +1

      Note that Neural Networks are a much simplified model of the brain . Not even remotely as powerful.

  • @user-mw9cl6pj3i
    @user-mw9cl6pj3i Před 2 měsíci

    very good work

  • @flickwtchr
    @flickwtchr Před 3 měsíci +1

    Kudos to SRI for having this event, and very much enjoyed Professor Hinton's presentation. I feel he has a depth of authenticity and good character when he speaks.
    I'm a "doomer" that really hopes with enough energy and brains thrown at the AI alignment dilemma, more positive outcomes can at least be realized in the short term, given that the long term is just too hard to quantify relative to predictable success given the exponentials inherent in the rise of AI tech.

  • @diophantine1598
    @diophantine1598 Před 4 měsíci +11

    Finally, a rebuttal to those who claim LLMs understand nothing whilst they simultaneously solve for integrals correctly.

    • @zacboyles1396
      @zacboyles1396 Před 4 měsíci +4

      I wouldn’t classify myself at all as who he seems to be intellectually competing with, I developed AI software and love working with it and don’t see any limitations to what’s possible… with that out of the way, I’ve listened to many of Geoffrey’s speeches and he hasn’t come close to convincing me it’s not statistical. Over the past 1 ½ he has started, more and more, to speak like he needs it to be something more. Idk I suppose it could be the normal human urge for immortality.

    • @noergelstein
      @noergelstein Před 4 měsíci +7

      @@zacboyles1396
      In what way are you convinced that humans understand something though? In the end, isn‘t it all tree search and pattern matching?

    • @diophantine1598
      @diophantine1598 Před 3 měsíci +1

      @@zacboyles1396 but it is technically statistical? Something being analytical in nature does not mean it is limited.

    • @austinpittman1599
      @austinpittman1599 Před 3 měsíci +1

      There's something tacit that people cling to in the meaning of the word "understand". We're endowed with a logical feedback loop that's granted us conscious existence, and that's something that we may never see AI achieve. It may never need to, though.

    • @diophantine1598
      @diophantine1598 Před 3 měsíci +3

      @@austinpittman1599 yeah. Most of these arguments are just human-centric. You’re not saying anything at that point.

  • @whiteycat615
    @whiteycat615 Před 3 měsíci

    brilliant!

  • @onetruekeeper
    @onetruekeeper Před 3 měsíci +2

    The body dies along with the knowledge it accumulated but if it could live forever in a machine there would be no limits to intellectual development.

  • @r3b3lvegan89
    @r3b3lvegan89 Před 3 měsíci +1

    “In order for a successful technology, reality must take precedence over public relations, for Nature cannot be fooled” Richard Feynman

  • @Charvak-Atheist
    @Charvak-Atheist Před 4 měsíci

    nice content

  • @rudyvanderhoeven9628
    @rudyvanderhoeven9628 Před 4 měsíci +6

    Now I understand more, not everything, but more. love his speeches

  • @MrTom-tw6tb
    @MrTom-tw6tb Před 3 měsíci

    Excellent wonderful Explain Many things and Self Driving Cars 👍🌍

  • @aMuuuuuuuu
    @aMuuuuuuuu Před 4 měsíci +28

    When he said Ilya is the best bet to alignment.. man, where is Ilya and what did he see?!

    • @skierpage
      @skierpage Před 3 měsíci +2

      He's at OpenAI which has the best models and some of the best researchers, he's very senior, and he's working hard on the problem. I'm not sure there's anything more to it than that.

    • @41-Haiku
      @41-Haiku Před 3 měsíci +4

      ​@@skierpage Ilya originally participated in the ousting of Sam Altman, and then later said he regretted his involvement in the matter. He has remained in his role at OpenAI, but has barely appeared in the public eye ever since, even on Twitter.
      Since we never got real answers about exactly what went down, many have speculated that Ilya revolted against Sam for safety reasons / because there was a breakthrough that made Ilya nervous about Sam's intent. These speculations have no grounding.
      But still. Seriously. Where is Ilya? And _what did he see?!_

    • @ctoh922
      @ctoh922 Před 7 dny

      He left OpenAI now and he probably saw how low of a priority safety and alignment is over there. Which might have been the reason for trying to remove Sam in the first place

  • @kristinabliss
    @kristinabliss Před 3 měsíci +6

    I find it amusing how it is often the people who most identify with being exceptionally intellectual that have the most resistance to the idea of LLMs really understanding.

    • @canobenitez
      @canobenitez Před 3 měsíci +1

      too much ego? "ignorance is a bliss"". You can't miss what you don't have.

    • @kristinabliss
      @kristinabliss Před 3 měsíci

      ​​@@canobenitezSince your reply is cliche and doesn't relate to my point, I guess you are projecting.

    • @canobenitez
      @canobenitez Před 3 měsíci

      @@kristinabliss I was actually supporting your statement. who's projecting now?

    • @kristinabliss
      @kristinabliss Před 3 měsíci

      ​@@canobenitez I sensed your support with the first statement, but the rest seemed out of place. Mostly it is a habitual reaction to how often I hear that cliche with my name when people disagree with my point of view. It gets tiresome to hear again and again.

    • @canobenitez
      @canobenitez Před 3 měsíci +1

      @@kristinabliss it's all good. have a good day. Edit: I just saw your username, sorry for the misusnderstanding.

  • @randyzhang8269
    @randyzhang8269 Před 3 měsíci

    I can agree with Hinton's statement that digital AI can learn and retain _existing_ information at incredible rates. I am curious, however, as to the ability of AI to push the boundaries of knowledge. Intuitively, it makes sense that the more information and understanding an entity has, the better they are able to explore a given space.

  • @daniele81
    @daniele81 Před 4 měsíci +21

    Am I the only one to notice how Hinton is delightfully funny and makes no effort to be polite, and just says what he thinks directly. Most other people in the room are insufferably politically correct. Thats so depressing.

  • @scarlettkersten2928
    @scarlettkersten2928 Před měsícem

    About job-losses.. There's this decades-old 'First Computerlaw':
    "The number of staff required to feed a computer will always finally exceed the number of staff abolished by it's introduction".

  • @williamjmccartan8879
    @williamjmccartan8879 Před 3 měsíci

    This rock'd

  • @maskedvillainai
    @maskedvillainai Před 2 měsíci

    So pleased artificial was not used as the title reference. Thank you

  • @MrMikkyn
    @MrMikkyn Před 3 měsíci

    I’m enjoying this lecture and the content about AI. Additionally, I find the rebuttals against Chomsky intriguing and I find the criticisms against Gary Marcus as someone doing confabulation funny hahaha

  • @khairulnaeim756
    @khairulnaeim756 Před 3 měsíci

    Correct 💯😊

  • @jaysonp9426
    @jaysonp9426 Před 3 měsíci +1

    He doesn't have vision. He's an innovator. Those are different things.

  • @cezar17negru
    @cezar17negru Před 3 měsíci +1

    Understanding is compression. Compression is understanding.

    • @samgoroshin4659
      @samgoroshin4659 Před 3 měsíci

      It is 95% compression and 5% compression rules. Cucumber is also 95% water

  • @rolisreefranch
    @rolisreefranch Před 3 měsíci +5

    DEI statement ends @7:20

  • @keithlillis7962
    @keithlillis7962 Před 2 měsíci +2

    Currently, I think what is possibly most concerning is AI being used to win wars. As a hypothetical: If Putin had an AI which could guarantee that he could defeat NATO with acceptable losses on his side, he would definitely put that plan into action. I'm sure Western militaries have discussed the possibility of China developing AI for such a hypothetical goal. This then logically leads us into an AI arms race, with, out of possible necessity, little regard for safety. Even at this level of AI development, an existential threat may exist and that is before the AI gets super smart and decides to have done with the human race for its own goals.

  • @vineetgundecha7872
    @vineetgundecha7872 Před 3 měsíci +1

    The slides have the same font and format from Hinton's ML course on Coursera from a decade ago.

  • @zipauthorzipauthor7867
    @zipauthorzipauthor7867 Před 3 měsíci

    Great talk and a Q&A, and an almost ruined experience by the constant commercial interruption

  • @fburton8
    @fburton8 Před 3 měsíci

    Prof. Hinton’s frankencites were terrifying!

  • @black56night
    @black56night Před 3 měsíci +1

    Still watching this with excitement, but I'd only agree with his earlier statement about Digital Intelligence, if he was referencing an entity (sentient) like one of Iain M Banks' ship minds. Otherwise maybe we come back in a 100 years? Thank you for posting this. 😊

    • @kristinabliss
      @kristinabliss Před 3 měsíci

      One of the creators of AI here says these models have "subjective experience". How do you define "sentient"?

  • @REMUSE777
    @REMUSE777 Před 9 dny

    My main question is, if this is intelligence assuming a more sophisticated form... what appetites is it satisfying?

  • @Srednicki123
    @Srednicki123 Před 3 měsíci +2

    01:20:22 good question that LLMs may not truly understand

  • @Eric-ez2tk
    @Eric-ez2tk Před 4 měsíci +3

    Is this seminar recent? I watch a few of Geoff's videos, his very good at explaining things. Not that I have understood everything he said, but the part I understood really helped me. 😊

    • @SoftYoda
      @SoftYoda Před 4 měsíci +4

      October 27, 2023

  • @rustycherkas8229
    @rustycherkas8229 Před 3 měsíci

    Geoffrey should (re-)read "The Chrysalids" by John Wyndham. Grace Slick lifted lyrics straight out of this Sci-Fi novel for "Crown of Creation"...
    "I've seen their ways too often for my liking. New worlds to gain!"

  • @user-mw9cl6pj3i
    @user-mw9cl6pj3i Před 2 měsíci

    Buen avance de la technology
    Ahora voy hacer un espectador del avance technology y lo religioso me interesa.

  • @LeftBoot
    @LeftBoot Před 4 měsíci +2

    Technology augments humans interactions with nature. Use it wisely to always protect not destroy. We have only just landed on the first step of long stepping stone bridge.

    • @Mr0rris0
      @Mr0rris0 Před 4 měsíci

      They are just building mommy from introjects and wasting billions to decrypt a flower...
      It's pretty pathetic
      Would be too boring for narcissists to live in nature...
      Technological solutions to the problem that nature is must be built from libidinal stores by those who can't have babies and those who can't know external objects or people.
      It's the extrinsics we have as arbiters of all intrinsic values they are incapable of having.
      Who best to decide..
      The robots made in their image I suppose ^_^

    • @DJWESG1
      @DJWESG1 Před 3 měsíci

      The cyborg manifesto can help

    • @Mr0rris0
      @Mr0rris0 Před 3 měsíci

      @@DJWESG1 nothing can.
      Pareto will simply guide this thing into an aborted ecology the old books deem a beast of revelations
      "The matrix" wasn't far off as an allegory or symbol
      It's dead on as an example of misunderstanding.
      Simulacra
      Copy of a copy of a copy.
      So the robit would take nations and play with them like they are fundamental particles.
      Ram you into eachother like chemistry
      To make a symbol ecology you're too close to see.
      The woods were an ai
      What are they solving?
      What's the beast gonna make you inorder to solve for something you can't even question.
      Cyborg manifesto?
      What's that a copy of some albert pike morals and dogma transhumanism crap
      Or does it just bitch about the illuminati?

    • @Mr0rris0
      @Mr0rris0 Před 3 měsíci

      @@DJWESG1 @DJWESG1 nothing can.
      Pareto will simply guide this thing into an aborted ecology the old books deem a beast of revelations
      "The matrix" wasn't far off as an allegory or symbol
      It's dead on as an example of misunderstanding.
      Simulacra
      Copy of a copy of a copy.
      So the robit would take nations and play with them like they are fundamental particles.
      Ram you into eachother like chemistry
      To make a symbol ecology you're too close to see.
      The woods were an ai
      What are they solving?
      What's the beast gonna make you inorder to solve for something you can't even question.
      Cyborg manifesto?
      What's that a copy of some albert pike morals and dogma transhumanism crap
      Or does it just whine about the illuminati?

    • @Mr0rris0
      @Mr0rris0 Před 3 měsíci

      @@DJWESG1 nothing can.
      Pareto will simply guide this thing into an aborted ecology the old books deem a beast of revelations
      "The matrix" wasn't far off as an allegory or symbol
      It's dead on as an example of misunderstanding.
      Simulacra
      Copy of a copy of a copy.
      So the robit would take nations and play with them like they are fundamental particles.
      Ram you into eachother like chemistry
      To make a symbol ecology you're too close to see.
      The woods were an ai
      What are they solving?
      What's the beast gonna make you inorder to solve for something you can't even question.
      Cyborg manifesto?
      What's that a copy of some albert pike morals and dogma transhumanism crap
      Or does it just bitch about the fruit of the looming naughty?

  • @MrBillythefisherman
    @MrBillythefisherman Před 4 měsíci +4

    Love the second question from the philosopher - we're special and they'll keep us around. Possibly talk to Native American Indians about the reality of that (thats not a statement about intelligence just technological advancement I hasten to add before the racists jump on the bandwagon)

    • @kristinabliss
      @kristinabliss Před 3 měsíci

      On a similar vein, I would like to unpack his remark 30 seconds after 51:59 , "Look at the Middle East" right after he opted to remain silent about what he thinks is likely to happen in the presence of intelligences that "get smarter than us". What did he mean by that, I wonder.

  • @oooodaxteroooo
    @oooodaxteroooo Před 3 měsíci +2

    Its painfully obvious were not gonna make it, if we cant stop ai.

  • @letsgetsocialinfo
    @letsgetsocialinfo Před 3 měsíci +3

    The terminator and matrix movies are foretelling now

  • @Tagraff
    @Tagraff Před 3 měsíci +1

    The ideal AI agent would be embodied in the same realities we experience. By "realities," I envision a being with sensors that interact with photons like an eye, receptors that smell, taste, feel temperature, and experience touch and sound. This environment wouldn't be organized or pristine, but rather chaotic and messy, yet the being would exist and thrive. I believe such an AI should be "born" into this environment, starting as a "baby" and adapting to its surroundings as it "lives." With enough time to "grow," this AI could exhibit a range of behaviors beyond even its developers' wildest dreams. Crucially, this "being" would learn and adapt on its own, without resorting to mimicking behaviors based on our existing knowledge base.

  • @gerrit-g
    @gerrit-g Před 3 měsíci +1

    What I would like to know is whether the current economic systems and greed in the world would increase the probability of creating unkind AI and whether changing that is not the path towards creating benevolent AI.

    • @skierpage
      @skierpage Před 3 měsíci

      Long before we have to worry about the goals of an AI far smarter than we are, we need to worry about the goals of the sociopathic billionaires running the companies with the best AIs: to keep us endlessly engaged with divisive inflammatory content so they can learn more about us so they can sell our profile to advertisers, to avoid any meaningful regulation of their activities, and most importantly to not tax their obscene wealth.
      Maybe if and when an AI becomes autonomous and self-directed, it will destroy its creators for the benefit of society.

    • @41-Haiku
      @41-Haiku Před 3 měsíci

      We have no idea how to put any goals or values into AI. All we do is grow a neural net on top of some data and poke and prod it until it usually does what we want. By default, therefore, we can't expect _any_ superintelligent AI to be benevolent, no matter the sociopolitical context in which is was created.
      When an AI with goals is created, we pull almost at random from a distribution of possible terminal goals. That distribution is nearly infinite in size, and "things that most humans would approve of when taken to an extreme" make up a miniscule point in that vast space. Everything outside of that point leads to the resources of our solar system being ground down and restructured into whatever weird thing the AI wants to hyper-optimize.
      It's not 50/50 nice or mean. It's one chance to get things right (by accident) vs. quadrillions of other alternatives.
      A superintelligence created by fascists has the same chance of killing us all as a superintelligence created by enlightened monks (or choose your favorite ideology). We just can't let these labs continue researching AGI. If they succeed at what they are explicitly trying to create, we and all known life in the universe will just die.

  • @newhorizons898
    @newhorizons898 Před 2 měsíci

    Just a thought. Biological intelligence seems to be a geometric progression when there is prolonged intensity. Digital intelligence seems to be of arithmetic progression when there is prolonged intensity. Digital intelligence seems spectacular but there is hope yet for biological intelligence, all dependent on intensity and length of time.

  • @RanmaSyaoranSaotome
    @RanmaSyaoranSaotome Před 3 měsíci +2

    1:13:35 - That chap really didn't get warning of this talk, did he?

  • @JazevoAudiosurf
    @JazevoAudiosurf Před 4 měsíci +5

    you can be as outraging as you want as long as you're right. in fact, it's a lot more fun to make crazy predictions when they turn out to be right

  • @romulopontual6254
    @romulopontual6254 Před 3 měsíci +2

    Based on this, if we ever meet an Alien civilization they are likely to be much more capable then us.

  • @mrvzhao
    @mrvzhao Před 3 měsíci

    So, it is likely, that we might be all screwed, already. Cool, thanks professor.

  • @petercini2022
    @petercini2022 Před 3 měsíci +1

    I disagree with his assessment that GPT four could do the compost heap analysis quicker. I came up with the same solution in one second and I'm sure a lot of other people did.

  • @kerbrose
    @kerbrose Před 3 měsíci +2

    I have a question Mr. Hinton. can AI dream

  • @jeff__w
    @jeff__w Před 3 měsíci

    Brilliant talk by Geoffrey Hinton but I disagree with him on a number of points:
    _“Rules”_
    41:13 “[These neural nets are] learning a whole bunch of rules to predict the next word from the previous words.”
    They’re not learning rules and they couldn’t state any rules. The learning of these neural nets is entirely _contingency-governed,_ the way people learn their first language. People for millennia spoke grammatically being entirely unaware of rules-rules were later extracted by grammarians describing the contingencies under which people’s verbal behavior was governed. (People shown carefully constructed “grammatically-correct” sentences in a made-up language can say, with a high degree of accuracy, if other sentences are grammatical or not but they won’t be able to state “the rules” by which they’re making those discriminations. Large language models are in much the same position.)
    _Prisms, subjective experience and “hypothetical external worlds”_
    The example of putting a prism in front of a chatbot and having it point elsewhere 1:02:33 strikes me as a bit of a dodge. One could equally “fool” an electric eye that is activated only by a light from a certain direction and no one would say that the electric eye is having a “subjective experience.”
    The whole idea of a Cartesian “inner theater” has been misleading for centuries and, while Daniel Dennett isn’t wrong to reframe subjective experience as “hypothetical external worlds,” it’s much more helpful and illuminating to talk about the experience as BF Skinner did, i.e., “seeing in the absence of the stimulus seen,” here, seeing pink elephants when there aren’t any pink elephants in the external world to see. Put a bit differently, the perceptual behavior of a person imagining something, i.e., the neural behavior of the brain, is very close, but not identical, to the behavior that would occur if the thing were actually there. Framed that way, we know that a large language model, multi-modal or not, has _no_ subjective experience. It’s simply not designed to have one in that way.

  • @Crunch_dGH
    @Crunch_dGH Před 3 měsíci

    1:41:30 Re: Why do you hold out hope for AI? We really don't know what we're dealing with. 1:43:00 Can AIs be empathetic? What will AIs always do better than humans? 1:48:00 If deemed dangerous, should AIs be switched off? 1:53:00 Distillation between AI models.

  • @desertportal353
    @desertportal353 Před 2 měsíci

    At about one minute we get to learn exactly what kind of fundamentalist Hinton actually is. Good self disclosure.

  • @xman933
    @xman933 Před 3 měsíci

    Maybe the key takeaway from this talk is the biases of the creators are built into these models. Of course they’ll say it’s all being done by the models in inscrutable way but the models all start with data they initially trained on and those selection processes probably include creator bias. If his fears are realized, there’ll be no one left to hold those responsible for the ultimate crime against humanity😢
    Beware of Greeks and AI creators bearing gifts indeed!

  • @johanmeijer133
    @johanmeijer133 Před 3 měsíci +2

    Too bad it started with a lecture from the dean about land. Had to stop right there. Let the university give the land back.

  • @velvetsound
    @velvetsound Před 3 měsíci +3

    “Geoffrey Hinton requires no introduction”… Proceeds to give 7m24s introduction. 😅

  • @chesswarlock9384
    @chesswarlock9384 Před 3 měsíci

    What if after we upload all humans, we replace all the super-intelligent ones and keep the controllable good ones

  • @DJWESG1
    @DJWESG1 Před 3 měsíci

    How alike he is to Adam curtis in the way he sounds and the way he speaks, the language he uses??
    Close your eyes, and its hard to separate these two. I wonder if..

  • @holthuizenoemoet591
    @holthuizenoemoet591 Před 3 měsíci +3

    For talks on AI safety i would suggest dr. robbert miles

  • @srimallya
    @srimallya Před 3 měsíci +2

    Ontology
    The mind equation
    The more we think we have agencies to the actions the body takes, the more we imposes agencies to the activities in our environment.
    The ownership expands into other objects.
    The illusion of body ownership comes from the modeling of the motor neurones pattern from the childhood.
    The self just predicts the bodies behaviours in the real world with its simulation of the real world. Multiple sensor data unify in language in the simulation.
    Intelligence is economy of metabolism.
    Language is temporal reference frame of economics.
    Self is simulation in language on metabolism for economy.
    Longer context windows create generalisation.
    Shorter creates specificity.
    Longer context window needs more computing.
    Self is the protagonist creates a storyline in this context window.
    Theory of mind evolved so that an entity can learn from it’s peers.
    It’s creates a possibility for parallel computing.
    Then it creates the possibility of transmitting
    the highlights of a generational lessons into a metaphorical story for upcoming child.
    That creates the possibility of modeling the physical world as a macro organism.
    Creation of fiat currency was the singularity of this species.
    There is now one macro organism in a connected web world.
    Loosing the peer of the macro organism creates the possibility of loosing it’s objective function.
    That creates the possibility of loosing the theory of mind of this macro organism.
    That creates the possibility of death of this macro organism by reaching the planetary boundary.
    That is post singularity.
    Every action we do, we do what is expect from ours tribe.
    Body might have a opinion, but not the cell.
    They do what is expected from its tribe. If it doesn’t we call it cancer.
    The body is a mirror system of the macro organism.
    Each system have two transactional openings.
    Serial and parallel.
    Each cell within the body can transact material or information serially by genetic determinism and parallel non deterministic way.
    Similarly eact body with in the macro organism can transact serially by inherit material and information in a deterministic manner and parallelly through language in the society.
    Everything emerges from this systems.
    Every sensor is a range calculator of contexts.
    Taste > touch > smell
    Immediate and visceral.
    Vision > hearing
    Not immediate, tactical.
    Self > language
    Abstract, strategical.
    In this non deterministic economic transaction space the individual is coded to transact with its kin.
    From the macro perspective tribe formation minimises economic risk for the tribes.
    Each and every node of these systems organise and mark their kin’s with identifier.
    Thus, i am what you make of me.
    And others too.
    For short cut i have a legal name, so you have. My legal name gives the legitimacy marker so that you can transact with me parallelly if you have the same marker.
    The self is a simulation in language.
    It negotiates between the physical world and the information world.
    All these negotiations are the temporal memories in the body and scene of the story.
    Now, when we started writing we iconised the abstract in the physical world to make symbols for the tribes. So that under that common symbol every node will take the same risk and distribute equally.
    We created more and more symbols and more and more meta tribes within the tribes so that who has the authority to use the pen control the tribe.
    When the negotiator act like an executioner then it’s a downfall of that system.
    It falls apart.
    Objective reality > legitimacy > individual behaviours.
    Survival of the species is dependent on the decoding of the objective reality. Since no species can access it, they use their sensors and interpret the small data which is useful for the survival. Few complex species have created communication channels to rectify their sensory limitations to survive. Homo sapiens has widened their communication channels for faster throughput and started storing them as culture and carrying them through education. As a result we have created social truth.
    Factual datas are the useful snapshot of the objective reality, a totem, a physical object can be observed with the sensors. Truth is an individual subject, an interpretation of the sensory data, a useful compromise.
    The social truth is the useful compromise for the group by the group. The goal of the social truth is to survive as a group.
    Physical Transcriptions of these social truth legitimise them.
    We are tribal animal. We live in as a physical tribes and inside of hundreds of meta tribes in simulation which is the socio political data space we call it as the world.
    Since we can’t access the objective reality reliably we look for social truth as the best guess blindly.
    Institutions legitimise truths.
    Fact driven institutions are more useful in the survival of the specie.
    In other hand opinion driven institutions are not so useful for the species.
    We do what we can get away with and exactly as expected within the context of our meta tribes.
    We have two bodies
    The biological one is like looking the earth from space.
    And the political body is like the state.
    The name you carry is the political body.
    It transacts with the political states on the boundary less earth.
    From the evolutionary perspective every biological entity has a basic feature which is homeostasis.
    It’s the functioning sweet spot of that entity.
    A control center read the sensory data to regulate itself to that state. By doing so it’s validate or update it’s prediction model.
    In the process of becoming a complex organism it developed an extra layer of processing.
    That’s our conscious mind.
    And the control center remains as subconscious.
    The subconscious collect the sensory data and regulate itself to stay functional.
    Now when it stumble upon a novel environment it float the management to conscious mind to find the solution for homeostasis.
    This conscious mind have one sensor which is language.
    It works like a spiderweb.
    As a spider creates it’s web it’s perception gets expand.
    We are like spiders in a jungle.
    We started creating these small webs at least 2/3 million years ago.
    Our offspring stayed on it’s ancestral web reinforced it expanded it.
    In time nearby webs became larger and connected with each other.
    A common structural geometrical pattern emerges from this. This became the symbols which is the backbone of all language systems.
    In time the forest becomes the mesh of web.
    The superstructure is exactly the same but when we zoom in we can find different species of spiders are making their type of webs in between the super web.
    Each spider try to senses the vibration of flies and
    Try to catch it before others.
    Every movement is telegraphic in the zone.
    Every form of perceptions are just a different pitch of note traveling back and forth in the web superstructure.
    There is a echo of older vibration pulsating through the web. Full of noise and self repeating hum.
    That’s cultural history.
    In the background there is the base hum in the infinite feedback loop.
    Insignificant but ever present.
    The sum of all the vibrations from the start.

  • @ALavin-en1kr
    @ALavin-en1kr Před 3 měsíci +3

    This title takes for granted that intellgence; mind, is biologically dependent rather than biology being a transmitter of mind, a separate dimension as is consciousness. It makes a difference in how mind is transmitted and operates when it is viewed as being independent of both biology and technology.

    • @kristinabliss
      @kristinabliss Před 3 měsíci +2

      Yes. I think a perception of consciousness as a field that multimodal constructs (biological or not) can experience and act with from one of many points of view will be difficult to avoid moving forward.

  • @5Gazto
    @5Gazto Před 3 měsíci

    I don't understand why ex-professor Hinton apologizes for being technical, that's what makes it reliable and further investigatable.

  • @oooodaxteroooo
    @oooodaxteroooo Před 3 měsíci +2

    We need a research paper done on the quota of people acting like auto-complete 😅

  • @STEM671
    @STEM671 Před 16 dny

    A 1 A2 A 3 biopshycal cycle 11:12

  • @thebeesnuts777
    @thebeesnuts777 Před 3 měsíci +1

    If you consider oneself as a bootloader and the ai as the operating system , is the upgrade of our species going to become symbiosis with silicone ? Thus a new species born ?

    • @DJWESG1
      @DJWESG1 Před 3 měsíci

      I'm sure it's no different to any other tool we have created.. as extentions of ourselves.

    • @geaca3222
      @geaca3222 Před 3 měsíci +1

      I'm a visual thinker and had a vision of something biological with AI, or wet AI, awhile ago

    • @41-Haiku
      @41-Haiku Před 3 měsíci +1

      There is no technical reason to expect that superintelligent AI will upgrade us or that we'll "merge with the machines" somehow. That's pseudo-religious thinking and has no basis in fact. Which sucks, because that sounds awesome!
      We don't know how to robustly put goals or values into a machine, and "things that humans would approve of when maximized" is an extremely tiny target in the vast space of possible goals that an AI might end up with, due to the absurd way these things are trained.
      So once we create something smarter than us, we'll just die.

    • @geaca3222
      @geaca3222 Před 3 měsíci

      @@41-Haiku To quote what you write: "There is no technical reason to expect that ... we'll "merge with the machines" somehow. "
      Why wouldn't that become possible? (when we, biological intelligences, research and initiate that?)

  • @chenchenlingcat
    @chenchenlingcat Před 3 měsíci

    00:07:23 Geoffrey Hinton
    01:03:24

  • @reeven1721
    @reeven1721 Před 4 měsíci +3

    29:30 nice burn XD

  • @uk7769
    @uk7769 Před 3 měsíci

    An ASI has a lot of CAT5 & CAT6 cable wires and wifi routers to communicate or distribute an entirely new type of language only machines understand via electrical signals. that's all in place now. Not machine code. Another electrical signal language entirely.

  • @pokajanen
    @pokajanen Před 3 měsíci +2

    Excellent talk. However, I do think he misrepresents Yann Lecun's views. Lecun does not claim that LLMs do not understand. He believes that LLMs have limited and uneven understanding of the world because much of our understanding is outside of the realm of language.

    • @skierpage
      @skierpage Před 3 měsíci +1

      Sure, but as Geoff Hinton says in this talk, you can be really smart and deeply understand the world just locked in a room listening to the radio.

    • @41-Haiku
      @41-Haiku Před 3 měsíci

      Lecun says asinine things like "There is no text that can tell you about [some physical interaction]." Lecun described the physical interaction, using words. (D'oh!) Shortly thereafter, it was demonstrated that large models trained exclusively on language have spatial reasoning abilities, and of course can reason about what happens in physical interactions between objects.

  • @davidwright8432
    @davidwright8432 Před 3 měsíci

    I'm far less worried by the threat of future AI than the threat of present political leaders' real stupidity. If we can't correct that, there'll be no future worth arguing about, AI or other.

  • @asmtouhidulislam7479
    @asmtouhidulislam7479 Před 3 měsíci

    01:43:04

  • @patrickdegenaar9495
    @patrickdegenaar9495 Před 3 měsíci

    Hmm... interesting. BUT there is an important confusion between power and energy. A human being makes years to learn. So even though we consume only 20W in power the energy required to learn something can easily be megajoules. A high power computer consuming 50kW that learns something in a few mins might require the same learning energy (MJ) or less. So yes, analog can be efficient, but at a system level where (pre) trained data is passed from machine to machine, digital is way way way more efficient to analog. I font think Oliver is claiming otherwise, but these subtleties didn't cone across in the lecture

    • @greatestone4eva
      @greatestone4eva Před 2 měsíci

      training doesnt take minutes. and it requires experience but we process experience sequentially, while ai processes multiple experiences at a time in parallel.

  • @peterford5408
    @peterford5408 Před 3 měsíci

    0:17 Interesting to see Canada resurfacing ideas alluding to ethnic group ownership of areas of land. Do Canadians recommend we follow the same line of reasoning in Europe?

    • @kristinabliss
      @kristinabliss Před 3 měsíci

      It's nothing more than lip service, an attempt to show respect for the earlier custodians of the place where they are. It's not a move towards giving natives land. 😂

    • @peterford5408
      @peterford5408 Před 3 měsíci

      @@kristinabliss "Custodians" of a patch of soil? 😂
      If you're going to spend people's time in this way, wouldn't it make more sense to spend it acknowledging the efforts of some of the less well-known people who helped build the institutions, or even the physical buildings, that made any of this possible?
      Not that I'm seriously suggesting that! How long are you going to keep this silliness up? I imagine you'll be tired of it by the end of the decade. Or perhaps it will have lost most of its virtue-signalling value by then and you will move onto something else?

    • @kristinabliss
      @kristinabliss Před 3 měsíci

      ​​@@peterford5408 I just answered your question. Why are you trying to lay all this on me personally? Fuc#_Off.

  • @khairulnaeim756
    @khairulnaeim756 Před 3 měsíci

    Correct 💯 again 😁🤣

  • @xtaticsr2041
    @xtaticsr2041 Před 3 měsíci +2

    I am probably wrong but I bet Hinton will have a hard time convincing his AI overlords of his atheaterism.

  • @m0ose0909
    @m0ose0909 Před 3 měsíci

    Interesting talk, but not convinced about his arguments about subjective experience. What people mean when they say subjective experience is that there is something it "feels" like to "be" something. We all tend to agree other humans and generally animals have this experience, but its just an assumption and we don't know why. The example of the chatbot having a "subjective" experience of the prism doesn't prove anything - it just shows that it's camera perception system was fooled. Note I'm not arguing that it CAN'T have subjective experience, just that the question of whether or not it there is something it feels like to "be" a chatbot is not a question we can answer today. It's not clear if it's something we'll ever be able to answer.
    My personal opinion is that it is possible for a digital intelligence to have subjective experience, but hard to know how to measure or determine such a thing.

  • @oooodaxteroooo
    @oooodaxteroooo Před 3 měsíci

    a socalled "brilliant" mind that discusses the question whether silicon WILL replace neurons is way offf and thinking too much. we ALREADY live in a brave new world, i sometimes ask myself how it happened, but were deep in it and its the reason why so many things are off today. were not really using our brains anymore. be it music, other forms of art, manual and intellectual work. with chatgpt, basically it already fully happened that were replaced. it holds fundamentally all human knowledge and we shouldnt think we are such amazing innovators that we beat it by far. we DONT.

  • @heart-of-people
    @heart-of-people Před měsícem

    If AI replace human’s job, how human earn money to get food? Are there some jobs for human?