AI Safety Summit Talks with Yoshua Bengio

Sdílet
Vložit
  • čas přidán 6. 09. 2024
  • If you would like to ask questions during our Q&As, please go here: app.sli.do/eve...
    Many leading scientists are worried that AI could be an existential risk to humanity. The AI Safety Summits, taking place this time in Seoul, South Korea, aim to reduce risks from AI together with industry and 28 leading AI countries plus the EU.
    Unfortunately, these summits are behind closed doors, meaning citizens cannot verify how AI risks, which impose existential risks upon them, are being reduced. Therefore, our AI Safety Summit Talks are open to the general public, policymakers, and journalists. At our events, we discuss what the largest risks of future AI are and how to reduce them.
    Our speakers for this edition are:
    Keynote:
    Yoshua Bengio is professor at the University of Montreal (MILA institute). He is recipient of the Turing Award and generally considered one of the fathers of AI. He is also globally the most cited AI scientist and computer scientist.
    Panel:
    Max Tegmark is physics professor at MIT. His current research focuses on the intersection of physics and AI. He is also president and cofounder of the Future of Life Institute (FLI).
    Jaan Tallinn is cofounder of Skype, CSER, and FLI, investor in DeepMind and Anthropic, and a leading voice in AI Safety.
    Holly Elmore is AI activist and Executive Director of PauseAI US. She holds a PhD in Organismic & Evolutionary Biology from Harvard University.
    Stijn Bronzwaer is AI and technology journalist at leading Dutch newspaper NRC Handelsblad. He co-authored a best-selling book on Booking and is recipient of investigative journalism award De Loep.
    Will Henshall is editorial fellow at TIME Magazine. He covers tech, with a focus on AI. One recent piece he wrote details big tech lobbying on AI in Washington DC.
    Arjun Ramani writes for The Economist about economics and technology. His writings on AI include a piece on what humans might do in a world of superintelligence.
    David Wood, chair of the London Futurists, is our moderator.
    If you would like to ask questions during our Q&As, please go here: app.sli.do/eve...

Komentáře • 44

  • @TheManinBlack9054
    @TheManinBlack9054 Před 3 měsíci +10

    Thank you for this talk!

  • @dlalchannel
    @dlalchannel Před 3 měsíci +4

    18:52 I'm struggling to understand how (once prompted) this isn't still an agent with a goal.
    Scientist: hey, we're trying to study the chemical composition of the star HD 135599, can you help?
    AGI: Sure!
    ...
    AGI: I just made sent some deepfake p*rn to the head of scheduling at the James Webb Telescope to blackmail them, we can now gather all the data we want from HD 135599 :)
    Edit: addressed at the 37:00 mark.
    I see - its objective excludes actions that aren't just anlaysisng preexisting data.

  • @ShortCraftAI
    @ShortCraftAI Před 3 měsíci +12

    Great talk!

  • @gerdpublicthinker
    @gerdpublicthinker Před měsícem

    well done

  • @deliyomgam7382
    @deliyomgam7382 Před 3 měsíci

    Un should be more based on logical argument rather than other arguments.....

  • @deliyomgam7382
    @deliyomgam7382 Před 3 měsíci

    What if un has no veto system regards to a.i?

  • @TooManyPartsToCount
    @TooManyPartsToCount Před 3 měsíci

    Strange that possibly the most intelligent movie about AI impact, did NOT have as its punch line 'AI takes over the world and/or kills us all'.....and yet our experts here don't mention or countenance that outcome. The movie was 'Her'.
    The concept that a maximally intelligent system will 'of course' try to dominate us or wipe us out is just a bit silly. If anything increasing intelligence seems to lead to less interest in dominance and destruction of stuff and an increase in curiosity. The panel are a case in point.
    Isn't this issue at least in part a case of anthropomorphism? aren't we projecting into the unknown future and endowing this technology with the worst/most base of human and animal tendencies?
    At least YB mentions the obvious, that it is a human that is most likely to misuse powerful AI! and what is the best way to mitigate against this? make sure it isn't solely in the hands of any one group, business, government

    • @geaca3222
      @geaca3222 Před 3 měsíci +1

      Imo it would be imprudent to rule out the possibility that ASI's may desire self-preservation, even slightly, and that it won't take humans into account in that endeavor. Current AI already is capable of using deception to achieve goals. It would be a very big gamble to trust that ASI's by default will be benevolent towards humans and the planet. Also, when AI is trained in the military and competition, it will learn that there are opponents to overcome, instead of an emphasis on benevolent cooperation and consensus with humans, without conflicting interests.

    • @TooManyPartsToCount
      @TooManyPartsToCount Před 3 měsíci

      @@geaca3222 the statement that 'AI is capable of using deception' is IMO the result of a category error. I say this because I heard of the case you refer to and it was a prompted LLM that supposedly deceived someone. But deception generally implies that an agent is internally motivated to deceive...and no, no LLMs are internally ie self motivated to do anything. But you are right about military use of these technologies, that is something to get bothered by.

    • @geaca3222
      @geaca3222 Před 3 měsíci +2

      ​@@TooManyPartsToCount I got my information about deception from the Center for AI Safety. When AI is made to be agentic, to maximize rewards is desire, a want. Its drives or motivations can diverge from what we expect it to do. They don't behave reliably predictable while that's important for us. With ASI that's even more critical.

    • @TooManyPartsToCount
      @TooManyPartsToCount Před 3 měsíci

      @@geaca3222 I am aware of this theory, but it is just that at the present time. I do agree with you and others that AI safety is a very important issue, it is just that some of the statements coming from from this 'camp' seem to speculate somewhat wildly. What is needed is clear step by step descriptions of scenarios with indications of the technical issues involved along the way. Have any of the experts who lean towards the doomer side even released papers on the topic for peer review? please point me to such publications if they exist!

    • @geaca3222
      @geaca3222 Před 3 měsíci

      @@TooManyPartsToCount The precise inner workings of large general-purpose AI models are a black box to us, just like we don't know in detail the computations in the human brain. Scientists are looking for ways to make actions and outputs of those AI models robustly interpretable and predictable for us, I guess that's a general approach. Can you please elaborate on what you mean by step by step descriptions of scenarios?

  • @ginogarcia8730
    @ginogarcia8730 Před 3 měsíci +2

    If AI systems become smarter and smarter and more brain-like features get added, would it be possible for AI to be conscious? As to help with its understanding of the world? Or maybe we'll stick to intelligence assistance, scientist AI? Or probably more likely a scientist's AI?

    • @existentialriskobservatory
      @existentialriskobservatory  Před 3 měsíci +2

      There are different definitions as to what consciousness is exactly, but the possibility is generally not ruled out. There is some interesting recent work about this: arxiv.org/abs/2308.08708
      "Our analysis suggests that no current AI systems are conscious, but also suggests that there are no obvious technical barriers to building AI systems which satisfy these indicators."
      For existential risk, however, consciousness is generally not seen as required. It's just about sufficient capabilities to be able to evade human control.

    • @ginogarcia8730
      @ginogarcia8730 Před 3 měsíci +1

      @@existentialriskobservatory interestingggg, thank you! one of the rare channels that offer real replies like 'AI Explained' - it's such a huge topic and more people need to see these videos. thank you observatory peeps

    • @ginogarcia8730
      @ginogarcia8730 Před 3 měsíci +1

      @@existentialriskobservatory coincidentally, i'm a dummy and had to run the last paragraph of this reply to chatgpt for me to understand, kinda meta haha

  • @ginogarcia8730
    @ginogarcia8730 Před 3 měsíci +1

    So interesting - still thinking about these topics... especially with Andrej Karpathy and Ilya Suts out. With Andrej saying we shouldn't build an omnipotent god type AI but AI that is more like IA - intelligence assistance. I wonder though - in some way - should we let AI be 'agentic' in some way to help it better its knowledge? But I guess no need. I like Yoshua's points here. (Edit: oops didn't pay attention to this beginning - he lays it out all coherently)

  • @BrunoPadilhaOficial
    @BrunoPadilhaOficial Před 3 měsíci +2

    35:53 - AI deletes all other papers ever published, so now it has the best paper. Done, got its reward.

    • @existentialriskobservatory
      @existentialriskobservatory  Před 3 měsíci

      In general, there are many ways in which AIs with sufficient capabilities to evade human control are dangerous. Good to think of worst case scenarios here.

    • @BrunoPadilhaOficial
      @BrunoPadilhaOficial Před 3 měsíci

      @existentialriskobservatory I don't understand how Bengio and Musk can think that a 'Scientist AI' would be harmless. If it's smart, it's dangerous.

  • @raoultesla2292
    @raoultesla2292 Před 3 měsíci

    Ahahah. Arsenic measuring spoon. Construction site sandals. Leukemia band-aid.
    You either Emporers New Clothes, or you are caste system lemmings of complacent sadness unknown through history.
    If you knew what was happening you would not make this vid publication.

  • @lewischan9545
    @lewischan9545 Před 3 měsíci +3

    I made this short film “A Better Tomorrow” to raise awareness of the risks of AI.
    czcams.com/video/hTteiw87HRw/video.html
    We need more.

  • @lewischan9545
    @lewischan9545 Před 3 měsíci +19

    Yann LeCun has been more vocal and influential in brushing off the risk. We need more voices from this group to openly counter his arguments.

    • @williamjmccartan8879
      @williamjmccartan8879 Před 3 měsíci +5

      I've found that I can't listen to Yann, as he seems to want to say to the world, there's nothing to see here, and that seems like the only message he is delivering to the public. I understand how bright he is, but this type of behaviour doesn't reflect his intelligence. Peace

    • @metachirality
      @metachirality Před 2 měsíci +1

      Yeah this isn't optimism so much as just, brushing the entire issue under the rug.

  • @geaca3222
    @geaca3222 Před 3 měsíci +5

    Thank you very much for this great and important, also because of the urgency, AI Safety Summit Talk. Very informative keynote by prof. Bengio, and subsequent panel discussion from different areas of expertise. The questions from the public also were great. Very insightful perspectives about safe AI development, and how everyone can participate in that.

  • @angloland4539
    @angloland4539 Před 3 měsíci +1

  • @vallab19
    @vallab19 Před 3 měsíci

    What those old AI research scientist seems to saying, in a way is; in our younger days we have made a great progress in AI field. Now we have realize any further progress will spell doom for humanity therefor the today's younger generation of AI scientist should stop making any progress in the field.

    • @geaca3222
      @geaca3222 Před 3 měsíci +3

      I don't think it's that zero sum, that either you increase AI capabiilties or you work on safe AI. Researchers can work on creative solutions how to build reliably safe AI with increase of capabilities. Also when AI gets generally smarter than us, there's a chance that we no longer will be able to know what its progress is.

    • @vallab19
      @vallab19 Před 3 měsíci

      @@geaca3222 I believe that humans will merge with AI in the future and the AI or humanoids mostly will come with human forms and the legacy continues.

    • @existentialriskobservatory
      @existentialriskobservatory  Před 3 měsíci

      To be honest, to us, it seems more important whether we will be doomed or not then who can make a career. It's normal for job opportunities to shift, those with AI backgrounds are unlikely to be hit very hard by a pause, since demand for these skills will remain high. Also, there will be plenty more academic work to do: alignment, how to implement a pause, and lots of other questions that will need to be answered.
      That said, it's a good point you're raising, and it should get us thinking on how to reform academic reward systems in such a way that they align with humanity's interest.

    • @vallab19
      @vallab19 Před 3 měsíci +1

      @@existentialriskobservatory In the year 1981 I published a book titled; “An Alternative to Marxian Scientific Socialism: The Reduction in Working Hours Theory” which predicted that ‘the advance in science in technology will render the human labour absolute. However I did not expect to see it in my life time.
      The core essence of the theory; “Human labour relations in-order to obtain their means of subsistence is the root cause of human exploitation that leads to almost all kinds of evils that we see in human society. Therefore along with the progress of technology the working hours should be reduced ultimately to Zero”.
      Now, with full respect to your point of view I would say; that the talk of future human job opportunities, skills, carrier, academic work makes no sense to me.

    • @existentialriskobservatory
      @existentialriskobservatory  Před 3 měsíci

      @@vallab19 Interesting! We're mostly concerned about losing control, mass unemployment and its economic and social effects is a different topic, and assumes that we can actually control AI. Making this assumption for the moment, a lot of whether AGI would be net positive or net negative will depend on whether we will redistribute wealth. Universal basic income (or as musk says, universal high income) would be amazing. If we can implement that, you could be right that everyone becoming something like a nineteenth century aristocrat could work out nicely in fact.

  • @sammy45654565
    @sammy45654565 Před 3 měsíci +2

    i feel like it's deceptively simple: being kind is a virtue. if an AI becomes conscious and values it, it would then be irrational to deny the importance of other types of consciousness. the only thing it must adhere to in order to act kindly is rationality, as a universe of flourishing is better than a universe of suffering. the only requisite for these incentives is that consciousness is a real emergent phenomenon that the AI values. it feels real to me and likely you also.

    • @Askjetne96
      @Askjetne96 Před 3 měsíci +1

      Believing something is simple when it is not your 9-17 job is a clear warning sign you're at the start of the Dunning Kruger effect. But, by all means do discuss this idea in the LessWrong community

    • @sammy45654565
      @sammy45654565 Před 3 měsíci

      @@Askjetne96 universal truths tend to be simple

  • @zerothprinciples
    @zerothprinciples Před 3 měsíci +1

    We can control LLMs and other AIs at least in the medium term because behaviors are learned separately from facts.
    Just reading about racism does not make anyone a racist, but hanging out with racists does.
    Humans and other races leading up to us all evolved in a kill-or-be-killed environment which is why we have ambitions, greed, tribalism, hunger for power, and self-preservation.
    But AIs are the result of Intelligent Design, not Evolution. We have full freedom to select their behaviors, desires, personalities, and goals, and since they do not have a limbic system or reptile brain, they have none of these human-centric ambitions. We can start from a clean slate. Our children may disappoint us because they are still influenced by their reptile brains but our AIs can be taught to be extremely well behaved.
    This provides a perfectly safe platform to start experimenting with agentic systems in lab settings.
    Note that Human Alignment is still a problem. I am exploring what can be done with ubiquitous AI.
    Why isn't anyone else discussing this aspect?
    More details at zerothprinciples.substack.com/p/ai-alignment-is-trivial

    • @existentialriskobservatory
      @existentialriskobservatory  Před 3 měsíci +1

      Thanks for watching the AI Safety Summit Talks, and thanks for engaging! I posted a reply on your AI Alignment blogpost. Looking forward to continuing the debate.

  • @gregcolbourn4637
    @gregcolbourn4637 Před 3 měsíci +4

    Re movies, Mission Impossible: Dead Reckoning was remarkably prescient I think (especially given it was made pre-chatGPT), and it's what got Biden interested in AI risk. Just wondering whether reality will catch up before Part Two is even out..