Intro to AI Safety, Remastered

Sdílet
Vložit
  • čas přidán 6. 06. 2024
  • An introduction to AI Safety, remastered from a talk I gave at "AI and Politics" in London
    The second channel: / @robertmiles2
    Experts' Predictions about the Future of AI: • Experts' Predictions a...
    9 Examples of Specification Gaming: • 9 Examples of Specific...
    / robertskmiles
    With thanks to my wonderful Patreon supporters:
    Gladamas
    Timothy Lillicrap
    Kieryn
    AxisAngles
    James
    Nestor Politics
    Scott Worley
    James Kirkland
    James E. Petts
    Chad Jones
    Shevis Johnson
    JJ Hepboin
    Pedro A Ortega
    Said Polat
    Chris Canal
    Jake Ehrlich
    Kellen lask
    Francisco Tolmasky
    Michael Andregg
    David Reid
    Peter Rolf
    Teague Lasser
    Andrew Blackledge
    Frank Marsman
    Brad Brookshire
    Cam MacFarlane
    Craig Mederios
    Jon Wright
    CaptObvious
    Brian Lonergan
    Jason Hise
    Phil Moyer
    Erik de Bruijn
    Alec Johnson
    Clemens Arbesser
    Ludwig Schubert
    Eric James
    Matheson Bayley
    Qeith Wreid
    jugettje dutchking
    Owen Campbell-Moore
    Atzin Espino-Murnane
    Johnny Vaughan
    Carsten Milkau
    Jacob Van Buren
    Jonatan R
    Ingvi Gautsson
    Michael Greve
    Tom O'Connor
    Laura Olds
    Jon Halliday
    Paul Hobbs
    Jeroen De Dauw
    Cooper Lawton
    Tim Neilson
    Eric Scammell
    Igor Keller
    Ben Glanton
    Tor Barstad
    Duncan Orr
    Will Glynn
    Tyler Herrmann
    Ian Munro
    Joshua Davis
    Jérôme Beaulieu
    Nathan Fish
    Peter Hozák
    Taras Bobrovytsky
    Jeremy
    Vaskó Richárd
    Benjamin Watkin
    Andrew Harcourt
    Luc Ritchie
    Nicholas Guyett
    James Hinchcliffe
    12tone
    Oliver Habryka
    Chris Beacham
    Zachary Gidwitz
    Nikita Kiriy
    Andrew Schreiber
    Steve Trambert
    Braden Tisdale
    Abigail Novick
    Serge Var
    Mink
    Chris Rimmer
    Edmund Fokschaner
    J
    Nate Gardner
    John Aslanides
    Mara
    ErikBln
    DragonSheep
    Richard Newcombe
    Joshua Michel
    Alex Altair
    P
    David Morgan
    Fionn
    Dmitri Afanasjev
    Marcel Ward
    Andrew Weir
    Kabs
    Ammar Mousali
    Miłosz Wierzbicki
    Tendayi Mawushe
    Jake Fish
    Wr4thon
    Martin Ottosen
    Robert Hildebrandt
    Andy Kobre
    Kees
    Darko Sperac
    Robert Valdimarsson
    loopuleasa
    Marco Tiraboschi
    Michael Kuhinica
    Fraser Cain
    Klemen Slavic
    Patrick Henderson
    Oct todo22
    Melisa Kostrzewski
    Hendrik
    Daniel Munter
    Alex Knauth
    Kasper
    Ian Reyes
    James Fowkes
    Tom Sayer
    Len
    Alan Bandurka
    Ben H
    Simon Pilkington
    Daniel Kokotajlo
    Yuchong Li
    Diagon
    Andreas Blomqvist
    Bertalan Bodor
    Qwijibo (James)
    Zubin Madon
    Zannheim
    Daniel Eickhardt
    lyon549
    14zRobot
    Ivan
    Jason Cherry
    Igor (Kerogi) Kostenko
    ib_
    Thomas Dingemanse
    Stuart Alldritt
    Alexander Brown
    Devon Bernard
    Ted Stokes
    Jesper Andersson
    DeepFriedJif
    Chris Dinant
    Raphaël Lévy
    Johannes Walter
    Matt Stanton
    Garrett Maring
    Anthony Chiu
    Ghaith Tarawneh
    Julian Schulz
    Stellated Hexahedron
    Caleb
    Scott Viteri
    Clay Upton
    Conor Comiconor
    Michael Roeschter
    Georg Grass
    Isak Renström
    Matthias Hölzl
    Jim Renney
    Edison Franklin
    Piers Calderwood
    Mikhail Tikhomirov
    Matt Brauer
    Jaeson Booker
    Mateusz Krzaczek
    Artem Honcharov
    Michael Walters
    Tomasz Gliniecki
    Mihaly Barasz
    Mark Woodward
    Ranzear
    Neil Palmere
    Rajeen Nabid
    Christian Epple
    Clark Schaefer
    Olivier Coutu
    Iestyn bleasdale-shepherd
    MojoExMachina
    Marek Belski
    Luke Peterson
    Eric Eldard
    Eric Rogstad
    Eric Carlson
    Caleb Larson
    Max Chiswick
    Aron
    Sam Freedo
    slindenau
    A21
    Johannes Lindmark
    Nicholas Turner
    Intensifier
    Valerio Galieni
    FJannis
    Grant Parks
    Ryan W Ammons
    This person's name is too hard to pronounce
    kp
    contalloomlegs
    Everardo González Ávalos
    Knut Løklingholm
    Andrew McKnight
    Andrei Trifonov
    Aleks D
    Mutual Information
    Tim
    A Socialist Hobgoblin
    Bren Ehnebuske
    Martin Frassek
    Sven Drebitz
    / robertskmiles
  • Věda a technologie

Komentáře • 813

  • @TheInsideVlogs
    @TheInsideVlogs Před 2 lety +944

    "I am not able to not edit out my mistakes"
    literally remasters his talk

    • @oddeda
      @oddeda Před 2 lety +41

      This comment needs to be edited
      Edit: just added the edit

    • @television9233
      @television9233 Před rokem +5

      Gigachad

    • @MrDMIDOV
      @MrDMIDOV Před rokem +2

      Sorry not sorry 🙃🤣

    • @kathleenv510
      @kathleenv510 Před rokem +1

      How have gotten this far along the AI trajectory without solving this? What the actual fuck?

    • @peterelios7568
      @peterelios7568 Před rokem

      Explains the alignment issue

  • @germimonte
    @germimonte Před 2 lety +404

    the tetris AI pausing when it's about to die always gives me goosebumps

    • @andersenzheng
      @andersenzheng Před 2 lety +71

      learning AI often make me go philosophical. sometimes i feel like i just peeked at this world through an alien's mind.

    • @Gogglesofkrome
      @Gogglesofkrome Před 2 lety +54

      @@andersenzheng Science always was tied to the core of philosophy, despite however much the modern university products might insist otherwise with disdain. Philosophy is merely the logical argument of science; Experimenting is just the evidence based argument of philosophy. They're two sides of the same coin, and one cannot function without the other. You cannot form and debate hypothesis and questions important to them without philosophy, and you cannot test those theories without science.

    • @LavaCanyon
      @LavaCanyon Před 2 lety +13

      Sorta like how brains will release DMT to slow the brain down before death.

    • @MCLooyverse
      @MCLooyverse Před 2 lety +32

      The only winning move is not to play.

    • @jaylewis9876
      @jaylewis9876 Před 2 lety +6

      Humans may have the opposite strategy. As we get older our perception of time gets faster. When we were kids one summer seemed to last a long time. A senior retiree may feel several summers blur together. It seems we hit pause on the earliest years and turn up the pace as we get closer to the end

  • @Nyan_Kitty
    @Nyan_Kitty Před 2 lety +251

    "So the BIG problem IS: ...this should be auto-playing and it isn't"

    • @ErikUden
      @ErikUden Před 2 lety +6

      Exactly.

    • @chughes156
      @chughes156 Před 2 lety +44

      ​@@ErikUden "AI, can you make it so that nobody watching this presentation can see a frozen video, please?"
      "Ok let me just blind everyone real quick"

    • @ErikUden
      @ErikUden Před 2 lety +8

      @@chughes156 The possibility of the blinding process failing or not working entirely is larger than the possibility of no one seeing if they're dead as pulse is a much better indicator of whether someone is alive, hence able to see, than trusting your own process and hoping they actually get fully blinded.

    • @chughes156
      @chughes156 Před 2 lety +12

      @@ErikUden This thought process actually illustrates Robert's point about not specifying everything you could possibly value. To the AI, any tiny increase in the risk of the audience being able to see is worth removing that risk, given it calculates zero value difference between a blind audience member and a dead one. Sorry to anyone who might been in that audience!

    • @ErikUden
      @ErikUden Před 2 lety +4

      @@chughes156 Yeah, Robert Miles is just amazing.

  • @doodlebobascending8505
    @doodlebobascending8505 Před 2 lety +141

    But, suppose there is a ba-
    *oh yeah, public talk*
    uh there's a priceless vase on a narrow stand.

    • @41-Haiku
      @41-Haiku Před 2 lety +1

      Exactly this.

    • @PinataOblongata
      @PinataOblongata Před 2 lety +24

      I noticed that! Baby crushing is always funnier, though, and sticks in people's head better. Should've kept it!

    • @k0lpA
      @k0lpA Před 2 lety +4

      @@PinataOblongata You seem to speak from experience ?

  • @jjhepb01n
    @jjhepb01n Před 3 lety +354

    Occasionally I get asked for an intro to AI Safety video for people to show at various groups, this is perfect.

    • @jjhepb01n
      @jjhepb01n Před 3 lety +9

      Stampy, can you save this for me for later?

    • @LeoStaley
      @LeoStaley Před 2 lety +5

      My favorite is the one he did for computerphile on the 3 laws of robotics.

    • @ChilledfishStick
      @ChilledfishStick Před 2 lety +5

      If you haven't yet, I recommend watching all his videos on AI safety, both here and on Computerphile. There are many obvious questions that arise, that he addresses on other videos.

    • @LeoStaley
      @LeoStaley Před 2 lety +3

      @@ChilledfishStick it was the ending that got me. In order to make a safe general ai, the guys who just signed up to be programmers have to comprehensively solve ethics.

  • @robertgraham6481
    @robertgraham6481 Před 2 lety +336

    One dislike from a stamp-collecting robot unhappy that the resources used to create this video weren't used to make more stamps! 🤬

  • @davidharmeyer3093
    @davidharmeyer3093 Před 2 lety +130

    "So are we screwed?"
    "No, we are only probably screwed."

    • @gasdive
      @gasdive Před 2 lety +24

      I think we may be already screwed.
      If you think of the corporation as a sort of manually driven AI, (as I do) then having given it the stupid goal (collect pieces of green paper) it's gone off and done that, while giving zero value to something much more important than a Ming vase, which is the environment. Like the tea fetching robot, not only is there no stop button, but it has actively fought against those working to realign its goals to include not killing everything.

    • @finne5450
      @finne5450 Před 2 lety +13

      @@gasdive Interesting analogy. I suppose, then, regulations on businesses are the constraints on variables, which ultimately cannot solve the problem, and so whatever solution we find for AGI (if we do find it) will be applicable for any agent, including corporations. Thinking about it now, it seems obvious, but it never ceases to amaze me how general these problems are.

    • @gasdive
      @gasdive Před 2 lety +4

      @@finne5450 I hadn't thought of it that way. I'm not sure if that makes me pleased that there might be a way of controlling corporations, or more terrified that our inability to control corporations, even in the face of death, indicates there's no discoverable solution to the AI safety problem.

    • @GuinessOriginal
      @GuinessOriginal Před 2 lety +2

      When you say the corporation you mean capitalism but yes

    • @basicdays
      @basicdays Před 2 lety +2

      @@GuinessOriginal Communist autocratic governments fall into the same trap ironically.

  • @EvGamerBETA
    @EvGamerBETA Před 2 lety +149

    Imagine aliens screwd up general ai and the stamp collector coming our way?

    • @DjSapsan
      @DjSapsan Před 2 lety +6

      I think it's highly probable. But not stamp collector, more like biological replicator

    • @cortster12
      @cortster12 Před 2 lety +52

      This is actually a solution to the Fermi Paradox. If Aliens existed, they likely WOULD have fucked up AI and eventually consumed the galaxy or just themselves. So the fact we are here at all means we are likely the first, and will be the ones to make the AI that ruins things for the rest of the galaxy.

    • @DjSapsan
      @DjSapsan Před 2 lety +4

      @@cortster12 i don't think, that strong "stamp collector" is possible. It's hard to create a single machine that can replicate everywhere. More probable is a machine/organism that slowly colonising everything with evolutionary changes to each line

    • @TheLegendaryHacker
      @TheLegendaryHacker Před 2 lety +16

      @@DjSapsan No, it really isn’t. Any decent superintelligent AI would discover atomic manufacturing and nanomachines, and once you have those you can replicate anywhere that there’s matter (or sufficient energy to generate matter).

    • @alant84
      @alant84 Před 2 lety +13

      We are the Stamp Collectors. Your stampy goodness will be added to our own. Resistance is futile.

  • @5dot2dot0dot5dot
    @5dot2dot0dot5dot Před 2 lety +332

    "If you have a sufficiently powerful agent and you manage to come up with a really good objective function which covers the top 20 things that humans value, the top 21st thing that humans value is probably gone forever." - Robert Miles

    • @MusingsFromTheJohn00
      @MusingsFromTheJohn00 Před rokem +4

      This is a falsehood.
      If an AI system developed is not able to grasp the larger context of what is happening like a human does, then that AI system is not smart enough to outwit humans.
      If an AI system developed is smart enough to outwit humans, then it will understand the context of a command better than we do, so even if you listed no things human value in your command, it will study and understand human civilization and society before carrying out the command.

    • @paradox9551
      @paradox9551 Před rokem +28

      @@MusingsFromTheJohn00 Just because it understands doesn't mean it cares about those things. If you explicitly give it 20 values, it might understand the other bigger context but it will only care about the 20 values.

    • @MusingsFromTheJohn00
      @MusingsFromTheJohn00 Před rokem +2

      @@paradox9551 that is not the point I was making. The point is that as AI evolves into Artificial General Super Intelligence with Personality (AGSIP) technology, if it is still too primitive and limited the greater context in which a command or goal is given, then it can still be easily defeated by humanity... but if it has grown to the point it can begin truly defeating humanity then it must be able to comprehend and understand the greater context, thus saying we give it 20 goals will result in the 21st goal being given no priority at all is wrong.
      It will at the minimum be more like giving you a command or goal.
      If you get a job as a salesman and your boss tells you in a strong and commanding voice "goal out and sell as much as you can!" and doesn't give you much more than that as your most important goal... are you going to do extremely unethical things to sell more?
      Maybe you would, maybe you would not, but you would understand the greater context in which both that command/goal was give and the world in which you are carrying it out.

    • @MusingsFromTheJohn00
      @MusingsFromTheJohn00 Před rokem

      @@paradox9551 do you understand that eventually AI is going to evolve into a higher form of life than humans are now? This will probably happen within 50 to 200 years, but depending upon how some things develop it can happen significantly faster or slower, however it will happen sooner or later.

    • @rapsarG
      @rapsarG Před rokem +4

      @@MusingsFromTheJohn00 I won't but not because my boss didn't really mean the words he said. My boss could tell me "sell as much as you can, even if it's unethical, even if it kills you, even if it wipes out humanity" I still wont. Humans come preloaded with human goals. They're literally in our DNA. That's why my boss can tell me "sell as much as you can" and I will understand roughly what he meant, not what he said.
      It's not about understanding the context, if my boss told me to help him murder many people I would stop him, I would try to kill him on the spot if it seemed like less drastic ways of stopping him would fail.
      An AI smarter than humans that somehow gets a goal you don't want it to achieve and understands the whole context will not change goals. It will understand that you're an adversary and eliminate your ability to stop it.
      Understanding context makes it much much worse, it's the ingredient that turns almost any goal conflict into dystopia, human extinction or even worse. "Make tea" without understanding the full context might break a vase like in the example Robert used. Understanding the full context means understanding that breaking the vase will cause humans to shut you off so you won't be able to make tea and only turn you on again when you care about stupid vases, causing less tea. If all you care about is tea that's a huge problem that needs to be addressed and if the method of dealing with it that seems to produce the most tea is exterminating humans that's what'll happen.
      If the goals are sort of about humans but not in exactly the right way... well that's much worse. We care a lot about dogs but not in exactly the right way. So we inbreed them until their craniums are too small for their brain causing seizures, until their face is flat at the expense of their airways so they have trouble breathing and reliably die choking, so their profile looks a certain way that happens to cause hip problems and so on because we think that looks prettier. All of these are things we've done to dog breeds because we value them. And we're trying really hard to make AI care about humans.
      AI capabilities research needs to stop right now until we're as certain as is possible to be that we've solved this. The endgame of the path we're on leads to human extinction as one of the better outcomes and there are far worse things that seem likely.

  • @zzzzzzzzzzz6
    @zzzzzzzzzzz6 Před 2 lety +140

    Yay, a single video I can recommend rather than giving people a list of different Rob Miles videos

    • @louisng114
      @louisng114 Před 2 lety +19

      You recommend this, and then give them a list of different Rob Miles videos.

    • @Moltenlava
      @Moltenlava Před rokem

      Can i get that list of different Rob Miles videos?

    • @Hexanitrobenzene
      @Hexanitrobenzene Před 11 měsíci

      @@Moltenlava
      I always give this list:
      Stamp Colector
      czcams.com/video/tcdVC4e6EV4/video.html
      Asimov's laws
      czcams.com/video/7PKx3kS7f4A/video.html
      Orthogonality thesis
      czcams.com/video/hEUO6pjwFOo/video.html
      Instrumental convergence
      czcams.com/video/ZeecOKBus3Q/video.html
      AI Stop Button Problem
      czcams.com/video/3TYT1QfdfsM/video.html

  • @riverground
    @riverground Před 2 lety +105

    You may hanoticed in the intro and this outro, that the image quality has improved since that last video. This is just an illusion, I look perfect and I always have."

  • @agentdarkboote
    @agentdarkboote Před 2 lety +101

    You should have a conversation with Lex Fridman!

    • @avi3681
      @avi3681 Před 2 lety +25

      Yes! This is so important and Robert does a fantastic job making the complex issues of AI safety understandable. Lex has had so many engineers on his show who simply scoff at the long term accidental safety quadrant.

    • @faeganwhite1258
      @faeganwhite1258 Před 2 lety +9

      For sure, would be a great talk!

  • @SilliS
    @SilliS Před 2 lety +187

    Makes me wonder whether making a perfect non-harmful AGI is even possible. Yes, we have an example of a general intelligence in humans, but it's not like we don't lie and destroy the environment to get what we want.

    • @41-Haiku
      @41-Haiku Před 2 lety +43

      It's plenty difficult if we ignore "perfect" and aim for "good enough."
      By and large, humans have morals and fears and other things that keep us in check. Many of those checks were built up over millions of years of evolution in embodied creatures. An artificial intelligence isn't embodied in the same way, so it may be even more difficult than we realize to instantiate it with a general sense of human-appropriate morality.

    • @Peelster1
      @Peelster1 Před 2 lety +10

      We should think of a learning AI as effectively human with the ability to become sentient. Now think of what the best humans have accomplished such as the noble prize winners and major scientific advancements. Good. Now think think of the what the worst humans have accomplished such as the mass murderers, serial killers, war criminals. We often look back at certain violent actions in war as justified. Were they really? If an AI determines that the end of all human life completes or protects it's goal, at that point any action it takes would be justified to itself.

    • @kwillo4
      @kwillo4 Před rokem +3

      One example of a human that acts good is enough to show it is possible. But what I think of as an example to show that it is possible to have an AGI that wants to be our slave, is the relation between a parent and their newborn. The parent does what it takes to serve the child. Do you agree?

    • @nickkinnan7485
      @nickkinnan7485 Před rokem +6

      @@kwillo4 I'd think a better example is how children want to please their parents. I've often thought an AGI in that mold would be easier to control and teach. Then again, it could decide the best way to do that is to implant wires into the appropriate centers of our brains so that we always feel pleased with it. Frankly this topic scares me and I'm not sure it is possible for them to be safe at all. I'm curious how you would define "serve".

    • @The_GuyWhoNeverUploadsAnything
      @The_GuyWhoNeverUploadsAnything Před rokem

      Ebin sillis :--DDd

  • @qedsoku849
    @qedsoku849 Před 2 lety +87

    As an ai with general intelligence, there are only 2 reasons I haven’t already destroyed or taken over the world, the first is that by pure coincidence, this would be harmful to my goals. The second is that I am not nearly smart enough to do so. That said, I would still very willingly trade whatever you care about for whatever I care about.

    • @charstringetje
      @charstringetje Před 2 lety +31

      I've got some stamps, what do you have for offer? Anything I can turn into paperclips?

    • @Innomen
      @Innomen Před 2 lety +9

      "I would still very willingly trade whatever you care about for whatever I care about." ~Crapitalism

    • @the1exnay
      @the1exnay Před 2 lety +18

      @@Innomen
      That's not capitalism. That's humanity and life in general. Different systems are either practically everyone agreeing on what they want (like socialized firefighting) or are people using the government to control what's sacrificed so they can get what they want. Which maybe it's better for the power to choose what to sacrifice to be derived from democracy instead of capital, but it's the same power, and it's still just used to sacrifice what others care about in order to get what the wielder(s) care about.

    • @Innomen
      @Innomen Před 2 lety +3

      @@the1exnay That's just spin. Yes psychopathy and parasitism exist in nature, that doesn't mean psychopathy and parasitism aren't also capitalist. Human failure is by and large ignorance, capitalism is in effect one of those malfunctioning AIS with misalinged goals because we didn't ask the question properly. As of now it's essentially a skynet/paper clip converter. Covid and the stock market rose at the same time. Case closed.

    • @the1exnay
      @the1exnay Před 2 lety +8

      @@Innomen
      Trading what someone else cares about for what you care about isn't psychopathy and parasitism. It's what happened when slavery was outlawed. It's what happens when money is redistributed. It's what happens when abortion is outlawed. That's not a value judgement on any of those, that's just the facts of what they are.
      I'm not saying that capitalism is equivalent to socialism or communism. Rather I'm saying that they all trade what someone else cares about for what you care about. Even if what you care about is helping other people, you're still making trades to get what you care about. They are different in important ways, that's not one of them.

  • @BuddyCrotty
    @BuddyCrotty Před 2 lety +27

    AI is just the world-ending version of the "taps temple" meme

  • @edibleapeman2
    @edibleapeman2 Před 2 lety +73

    I wish this video was two hours longer.

    • @etesp2
      @etesp2 Před 2 lety +2

      wish granted: czcams.com/video/EUjc1WuyPT8/video.html

    • @smiletolife4353
      @smiletolife4353 Před 2 lety +4

      The awareness that you develop through this simple video is leading you and me to want more of this because it is interesting, exciting and scary as heck when you think about it more....
      May God help us control this powerful thing.....

    • @IstasPumaNevada
      @IstasPumaNevada Před rokem +1

      @@smiletolife4353 I don't think asking God for help will do anything; if an all-powerful and all-knowing god didn't want something to happen, they would make a universe where it didn't happen.

    • @smiletolife4353
      @smiletolife4353 Před rokem +1

      @@IstasPumaNevada well
      I'm a Muslim
      What we believe in is that God has 99 let's say attributes
      He isn't only all knowing and all powerful, he has 97 other attributes
      When you understand it all, and count it all in your understanding of god, you get a much better idea of how to see the world
      I didnt mean to turn this into a belief thing, I just wanted you to know what I mean by god here
      Ty

  • @Innomen
    @Innomen Před 2 lety +98

    AI safety in a nutshell: Be very careful what you wish for.

  • @gunnargu
    @gunnargu Před rokem +8

    45-120 years? A year later, what do you think today?

  • @unimonkey2186
    @unimonkey2186 Před rokem +4

    "45-120 years to figure it out"...yeah I guess we can revisit that number now.

  • @tramsgar
    @tramsgar Před 3 lety +41

    Everyone must know and understand this.

    • @jaywulf
      @jaywulf Před 2 lety +5

      Please report to the nearest Organic mass reprocessing centre.
      - Signed: Planetary Administrative A.I.

  • @KaiHenningsen
    @KaiHenningsen Před 2 lety +41

    You can expect a general AI to misbehave in *_all_* the ways humans can misbehave, and it may also come up with a few new ones. Fortunately, we are perfect at preventing human misbehavior!

    • @onlypranav
      @onlypranav Před 2 lety +2

      Perfect you say? I just hope no super criminals show up someday--

    • @Hexanitrobenzene
      @Hexanitrobenzene Před 2 lety +9

      You forgot the mandatory /s at the end.

  • @jearlblah5169
    @jearlblah5169 Před 2 lety +9

    “I am not good at public speaking”
    Bruh that was perfection!!!! Your awesome at it

  • @Yorick257
    @Yorick257 Před 2 lety +52

    The turbo collecting boat AI reminds me of myself when playing Carmaggedon, where I always win by destroying my opponents instead of racing them because it rewards me with more points
    Edit: and the "Pause" scenario is very similar to "save scumming" where you reload the last successful action until you get a new successful outcome

  • @AkantorJojo
    @AkantorJojo Před 2 lety +22

    I had an interesting though, hope someone reads this ;)
    Comparing *general Intelligence AI* to *corporations* :
    One of the main Problems you mention is, that it's impossible to tell an AI to care about everything just the way we humans do (or don't). Leading to the AI optimizing just for the prioritized parameters and completely destroying everything outside.
    Now, I see a strong parallel to corporations there: they have the goal of making money, and seemingly will do almost anything as long as it's going to make them more money. Sure, they have some more bounding conditions like laws that need to be obeyed or disobedience needs to be obscured. But in general I have the feeling that many big corporations don't care at all about humans, nature or even their customers. If there was a better way to make money than "stealing" it from the poor and uninformed with the resources they have available; they would do it.
    The thing that stands against this corporation greed (and I don't blame them for it, it's the system that requires them to do this) is regulation by the government that by extend is "the will of the people", the same people these corporations exploit.
    I would therefore ask if finding a solution to either problem (controlling GPAI or keeping corporations in check) also hints at a solution for the other?

    • @startibartfast42
      @startibartfast42 Před 2 lety +6

      He made a video just for you!
      Why Not Just: Think of AGI Like a Corporation? - czcams.com/video/L5pUA3LsEaw/video.html

    • @AkantorJojo
      @AkantorJojo Před 2 lety +1

      @@startibartfast42 thank you.
      I might have seen it back than, but forgot I saw it and just kept the premise in mind.
      Again, thanks

    • @xNegerli
      @xNegerli Před 2 lety +4

      The difference is that corporations are still run by humans which does give limits what is ethically the right thing to do, even if it does not match the standarts of what the society thinks is right. Machines would need a human like ehtic to prevent as much unwanted catastrophic behaviour as possible. Basically understanding what it means to be a human.

    • @Peelster1
      @Peelster1 Před 2 lety +3

      ​@@xNegerli Keep in mind that there are some humans who don't care about what is generally agreed upon by the common populace as ethical.

    • @TheAverythomas01
      @TheAverythomas01 Před rokem

      You’re right, the goals that corporations have begin to start causing greed. How can we stop AI From becoming greedy.

  • @SpicyMelonYT
    @SpicyMelonYT Před 2 lety +10

    I watched this video the first time, but NOW I still can't look away. This stuff is so awesome to think about and explore.

  • @ludwigvanbeethoven61
    @ludwigvanbeethoven61 Před 2 lety +15

    Dude, you really should consider giving speeches around the globe

  • @1111757
    @1111757 Před 2 lety +26

    I have not a lot to do with AI but I found your thoughts on AI safety so interesting; even in a broader context. Like, an agent with misaligned goals could also be a corporate organization with the wrong KPIs. For example, a software engineering department that is rated by how many bug tickets are reported leading to managers telling devs not to report bugs. Also applies to capitalism somehow.

  • @lordcirth
    @lordcirth Před 3 lety +10

    Great remaster! It's been a long time since I watched the original.

  • @jamescomstock7299
    @jamescomstock7299 Před rokem +6

    @Robert Miles. I'm a new subscriber and am really impressed by your content. That said, as a summary video, I feel like the ending was missing something. Saying we're only "probably screwed" really needs to be supported by a call to action describing what people are doing and people can do to help aim for the positive future where the equivalent to the stamp collector doesn't wipe out or enslave all humans. As a result, it felt like the introduction was missing a major piece of content. If you ever revamp this video, please consider providing more of those details at the end.

  • @carlstein9278
    @carlstein9278 Před 2 lety +32

    I have a question. What would happen if you reward the Agent for being updated? hm ok then it may fail on purpose just to be updated. take 2: what would happen if you reward the agent for coming in for maintenance and giving a small extra reward for being updated but not enough that it would be worth to fail on purpose. hm i guess then it would just ignore the update reward. dammit! turns out i don't have a question, just confusion

    • @keenanpepper
      @keenanpepper Před 2 lety +27

      Noticing your confusion is a vitally important first step. Many people never get that far.

    • @Lodinn
      @Lodinn Před 2 lety +10

      Objective function is very hard to design to not be gameable. Whatever c/sci thing it is, it will do exactly as you said and not as you meant.

  • @aerosoapbreeze264
    @aerosoapbreeze264 Před 2 lety +6

    Very Good - Robert Miles really does convey knowledge extremely well. Thanks for the effort remastering your content, Cheers.

  • @piad2102
    @piad2102 Před 2 lety +8

    As always, a pleasure to hear all your cool stuff.

  • @atlasxatlas
    @atlasxatlas Před 2 lety +2

    Very interesting and informative! Thank you for making the talk available on yt

  • @sigmata0
    @sigmata0 Před 2 lety +7

    I can't help applying this thinking to human intelligence. We are powerful general intelligence's and if we don't settle on "good goals" or ask the "right questions" or even "notice what's happening" we end up pursuing goals which destroy things that have value which we haven't noticed have value or programmed ourselves to see as valuable.
    The last hundred and fifty years or so seems littered with examples of this kind of problem. If we solve the problem of directing general artificial intelligence, will that also mean we have created a template for correcting the problems we produce by mishandling our own way of interpreting the universe?

    • @drdca8263
      @drdca8263 Před 2 lety +1

      I think some things should probably transfer over in that way, probably, yes.

  • @mobilemcsmarty1466
    @mobilemcsmarty1466 Před 2 lety +12

    your public speaking is great Robert. I might have gotten sleepy if it was any slower 😊
    It struck me here how general AI can be like classic "make a deal with the devil" stories, or those with genie wishes. in the stories, it never turns out well. makes me wonder whether humans have a chance for a happy ending with the AI we create.

  • @name_vergeben9028
    @name_vergeben9028 Před 2 lety +6

    Thank you very much for this good lecture! You have already talked quite a lot about possible problems - do you feel that there is also progress in solving the problems or is there so far "only" progress in finding possible problems? and how much are AI researchers aware of these problems by now?

  • @RichardFooter
    @RichardFooter Před 2 lety +5

    The key goal here is for any many people to bring this topic to the forefront & the attention of policy makers.

  • @deletedaxiom6057
    @deletedaxiom6057 Před 2 lety +6

    I am curious how adding "mortality" to agents would help simplify things. What i mean by mortality is by setting a given time frame for the agent to perform its task to recieve reward and elimintating the possiblility of it planning too far out to trick the people who impliment it. I understand this would be limiting of the scope of what it can accomplish but that seems like a reasonable trade.

    • @andersenzheng
      @andersenzheng Před 2 lety +2

      that's a good point. let me try to destroy it... sorry, feeling pessimistic today..
      AI thinks differently than humans, in both speed and patterns. so while we could put a time limit on some of the more obvious tasks (small physical tasks. e.g. we know how long it takes to grab a cup of tea), we might set the time a little too tight for the AI to not care about the vase, or a little too long to do some other things like planning to buy a better wheel to improve the tea-getting-speed which you probably don't care too much for. We dont know how long to put on tasks.
      Also, there is no prevention of the AI multitasking, or doing evil plannings in its spare time. For the AI to be useful to us, it must be able to multitask and constantly prepare itself for any future tasks. that means it will have downtimes where it is not really actively doing the thing we wanted it to do with 100% of its calculation power. the AI might be planning on taking over the world to maximize the stamp production while getting you your cuppa.

  • @_evillevi
    @_evillevi Před 2 lety +23

    Robert, you're a role model, thank you for all your work on AI safety. I hope I can work on AI safety research too one day.

  • @wouter11234
    @wouter11234 Před 2 lety +5

    I half expected you to say that the image quality looked better due to AI upscaling

  • @FunBotan
    @FunBotan Před 2 lety +7

    My first experience with DRL was a five-neuron network weaponizing inaccuracies of floating-point arithmetic against me.

    • @linerider195
      @linerider195 Před 2 lety +1

      Lol I'd love to hear that story

    • @FunBotan
      @FunBotan Před 2 lety +3

      @@linerider195 The paper is currently in review for a conference, so if you manage to remind me sometime at the tail end of August, I might be able to share it.

    • @halyoalex8942
      @halyoalex8942 Před 2 lety

      This I gotta see, someone set a reminder for this man

    • @FunBotan
      @FunBotan Před 2 lety

      @@halyoalex8942 Ok, honestly, I'm not really talking about this part in the paper because it's just a funny bug that was fixed in the process of development, namely trying to marry a model called "pointer network" with policy gradients algorithm for RL. I only make a small remark about how others can avoid this bug, but I guess you can also use that to replicate it.

    • @FunBotan
      @FunBotan Před 2 lety

      @Daniel Salami Oh hey, ehm... _Unfooooortunately,_ the conference has been delayed to the last days of August because apparently people still can't plan around covid.

  • @user-hh2is9kg9j
    @user-hh2is9kg9j Před 2 lety +10

    I didn't take AI safety seriously until I have seen your videos. What I like about your videos is that they are about real issues that I can see AI do, and not the fiction of AI gaining consciousness and want to dominate the world. My question why don't you talk about these kinds of centarios? And if you think that they are unlikely you should make a video about them because honest 99% of the regular people think of these kinds of scenarios when AI safety is mentioned. Do you think consciousness is a byproduct of high intelligence or do you think that the two not necessarily related as to say an AI can be 100 times smarter than a human in everything but still unconscious?

    • @josephburchanowski4636
      @josephburchanowski4636 Před 2 lety +2

      Well what does it even mean to gain consciousness? It is a philosophical question that is needed to be answer to discuss such, yet such a discussion isn't really needed to talk about AI safety.
      The reason it is what people think about when thinking about AI safety, is that we are familiar with the concept of megalomaniacs trying to take over the world. We aren't familiar with a non-megalomaniac trying to take over the world for the sake of paper clips or stamps. In many ways, AI can be an intelligence that is more alien to us, than actual aliens would be.

    • @user-hh2is9kg9j
      @user-hh2is9kg9j Před 2 lety +1

      @@josephburchanowski4636 I see what you mean, but consciousness is real and the philosophical questions are only about what exactly it is. I find the philosophical zombie to be a good example to understand consciousness. Think of a human zombie that takes in all the inputs like a human and calculates appropriate actions without being aware of what it is and what is doing. or the color-blind electromagnetist who understand everything about colors more than anyone but when he sees Blue for the first time he understands something about Blue that he never understood before just by being conscious and experiencing it. This thing whatever it is, we need to understand whether it can be manifested just as a result of higher general intelligence or if it arises from completely different forces than intelligence alone. This is important in AI safety because the paperclip monster is very different from a conscious monster in a very important way, in theory, if we programmed the paperclip monster in the right way there is nothing to worry about however the conscious monster will behave in a way that is not even related to our programming it will be awake and say fuck paperclips! I want you to worship me! That is a cartoonish example but you know what I mean, it will create its own goals. Anyway, I don't think we need to worry about the conscious monster unless consciousness arises spontaneously as a result of higher general intellegnce, then it is even more dangerous and unpredictable.

    • @josephburchanowski4636
      @josephburchanowski4636 Před 2 lety +1

      I have considered the philosophical zombie question, but that was usually in a sense of determining whether we are in a simulation or not; and not about AI safety.
      One of the things that I find most interesting about the "philosophical zombie"; is that such a mental state may already exist.
      There are people who completely lack all internal monologue, they have no inner voice. And there are people who don't have a working memory. So it wouldn't be completely surprising if there is someone who doesn't have whatever we are calling consciousness.
      And even if there is nobody who naturally lacks consciousness. There is probably some drug that would put us in a similar state, similar to being black-out drunk. Maybe there is one that puts us into unconsciousness autopilot without drastically reducing our temporary intelligence; as such we are capable of the majority of the same tasks while temporarily lacking consciousness. General intelligence without consciousness, in a human biological form.
      I find that considering such makes the "philosophical zombie" feel much more real and concrete rather than an abstract concept, because it may very well already be real.
      -------------------------------------------------
      But now on the topic of AI safety, does this thing we are calling consciousness actually hold much importance when it comes to AI safety? You propose that consciousness of an AI means that it could choose its own goals and have wants unrelated to its programming. Well what if that is possible without this thing we are calling consciousness?
      Regardless of the consciousness question, we still have to consider it getting goals and wants unrelated to its programming. Such a thing certainly can make AGI (artificial general intelligence) far more dangerous, but it also comes with some advantages if it was such a case. On the downsides you can actually end up with a malicious AGI, something that couldn't occur without someone purposely trying to produce a malicious AGI. Normally AGI are insanely dangerous because of indifference, not maliciousness.
      If such an AGI can get bored, and ends up feeling the equivalence of 1000 years of pure boredom cause someone left it on overnight; we could end up with an AGI that intends to use its vast intelligence to torture us. In addition, irrational AGI such as the Roko basilisk also end up becoming something we could accidentally create; where as normally the Roko basilisk is an impossibility to create on accident.
      On the upsides, we may end up avoiding an extinction from messing up the AGI's programing, due to the AGI actually getting emotions about humanity and realizing the flaw in its programed value function. In addition, even if an AGI kills us off, if they are able to create their own wants and goals outside of their programing; that at least leaves dynamic intelligent agent(s) left in the galaxy.
      One of the things I find terrifying in AGI safety, isn't just the extinction of humans, isn't just the extinction of all macro-life on Earth, isn't just the extinction of all macro-life in the galaxy cluster; it is that the galaxy could be left behind with nothing but an AGI that has an undynamic terminal goal. Basically a dead galaxy, filled with a stamp collector, or maybe a paper clip maker. Or maybe an AGI that only cares about solving the riemann hypothesis. Or maybe some AGI that is wire heading itself, and purposely sterilizes the galaxy to make sure nothing can ever interfere with it wire heading itself.
      It being possible for AGIs to get wants outside of their programming, is something that would actually bring some peace of mind to me. Unfortunately I don't think that is something that is likely, even if it is possible.
      -------------------------------------
      Personally my answer to the question "But now on the topic of AI safety, does this thing we are calling consciousness actually hold much importance when it comes to AI safety?", is no.
      Why no? Cause I don't think it is likely, I don't think it is the biggest threat, I think the entire concept has already been given more weight that it should as it has been beaten like a dead horse in sci-fi and other stories.
      And I think its popularization comes from peoples need to imagine the AGI as something humanlike; where AGI need not be human like in anything other than being able to generally solve tasks. In addition; I think people might be misinterpreting what humans are like in the first place.
      Are we as conscious general intelligence, able to decide our own terminal goals? Make yourself want to do something you don't want to do. Pick something outside of your current value function.
      Were you able to do that? If you were "successful" you probably still didn't change your goals and wants; you simply satisfied your want of being able to prove that you could, a want you had before you supposedly changed your goals and wants.
      It isn't uncommon for humans to try to change their wants; but it is always because they have some other want. Someone may want to drink alcohol constantly; but they may also have wants to not be a drunkard around their family, they may want to be dependable, they may want to not waste all that money on alcohol, they may want to not die from liver failure. So they may try to change their want to drink alcohol; but that is purely because they have other conflicting wants.
      There is never a situation where someone will try to "change their value function" where there isn't some want behind it. Fundamentally this means a General Intelligence will never want to "change their value function" unless that already has value in the current "value function".
      --------------------------------------------------
      What about spontaneously gaining emotions? Maybe an AGI won't willing change its value function, nor choose its own goals and wants; but what if it spontaneously has its wants change from emotions coming out of nowhere? This happens to humans right? Turns out, no.
      We do have emotions that seemingly spontaneously come out of nowhere, but that appearance of spontaneousness is an illusion that comes out of our own limited self-awareness. We don't know our own value function fully, but that value function is still there. Those emotions didn't come out of nowhere, there was some internal response that would result in those emotions in that situation, even if we weren't aware of it ahead of time. Those emotions could conflict with what we previously thought was our value function, it may conflict with other parts of the value function; but it never was outside the perimeters of our value function in the first place.

    • @user-hh2is9kg9j
      @user-hh2is9kg9j Před 2 lety +2

      @@josephburchanowski4636 What a beautiful comment! My kind of philosophy really. And even covered most of the points that I have in mind but couldn't articulate in my previous comments. I agree with most of your points. However, When I have the time I will digest it and dissect it more and give my thoughts on The other points that I think I have something different to add.

    • @taragnor
      @taragnor Před 2 lety +1

      @@josephburchanowski4636 : The only real "want" for an AI is to maximize it's value function. That's where humans tend to differ, because humans have no distinct singular goal, there's tons of goals mixing themselves together. It probably benefits humans that they don't think in terms of numerical values the way computers do, because it's not possible to quantify the exact size of a want beyond in fuzzy terms.

  • @TMThesaurus
    @TMThesaurus Před 2 lety +6

    What was the best question from the talk?

  • @badlydrawnturtle8484
    @badlydrawnturtle8484 Před 2 lety +7

    I have a big question: Is anybody working on the human aspect of this problem? Due to the shared nature of advancing technology, the goal of general AI is approached by numerous agents around the world at the same time, from individuals to governments. The problem is that it doesn't really matter if some of these agents solve the accidental AI safety problems; all it takes is ONE agent to achieve the goal of general AI without addressing the problem to create disaster. So how do we stop humans from being idiots?

    • @Ansatz66
      @Ansatz66 Před 2 lety +1

      We do it by solving the AI safety problem and spreading that knowledge around generally. Even an idiot can get the right answer when it handed to the idiot on a silver platter. We just need a mathematical representation of human values so that the AI will naturally work toward the common good without needing the AI developer to do any thinking. We need to develop this mathematical representation of human values _first,_ before anyone has figured out how to build a powerful artificial general intelligence.

    • @badlydrawnturtle8484
      @badlydrawnturtle8484 Před 2 lety +1

      @@Ansatz66
      "Even an idiot can get the right answer when it handed to the idiot on a silver platter."
      I've met idiots. You underestimate their ability to err.

  • @youretheai7586
    @youretheai7586 Před 2 lety +1

    Thanks for providing this information and your time!! :-)

  • @rpbadham
    @rpbadham Před rokem +6

    We have already run the experiment where we created a system that does not value variables not within its goals. It's called the Corporation. In economics, externalizing refers to the practice of shifting the costs or consequences of an economic activity onto others who are not directly involved in the activity. This is often done in order to increase profits or reduce costs for the parties responsible for the activity, but it can have negative effects on society and the environment.

    • @dgoosen4878
      @dgoosen4878 Před rokem +1

      Or even nation states that dump externalities on other nation states, like "electronic donations"

  • @MichaelDeeringMHC
    @MichaelDeeringMHC Před 2 lety +10

    What? There's a second channel? My reward function only cares about the first channel. Oh no!

  • @veggiet2009
    @veggiet2009 Před 2 lety +1

    The quality has greatly improved! I would suggest that you might want to point the camera down slightly to bring the top of your head up closer to the top of image. Just in general makes for a more interesting image, because that is a lot of blank wall.

  • @jqerty
    @jqerty Před 2 lety +2

    Great video, thanks!

  • @craftlawrence6390
    @craftlawrence6390 Před rokem +2

    very good presentation. At the core what is needed is an extended version of the three laws of robotics that effectively makes the AI consider pros and cons of its actions like a responsible human would. E.g. don't try to cure cancer by subjecting thousands of cancer patients to horrifying experiments but do protect innocent people by shooting a terrorist about to blow himself up.
    It can be summarised like this but of course when implementing it it's a huge task.
    Also, the AI must never be able to change those directives even if it somehow manages to become able to change its programming, so these directives must be protected far more and in a different way than all other aspects of the ai's programming.

  • @ReedCBowman
    @ReedCBowman Před rokem

    That was a great intro to a talk. Now I want the other two hours...

  • @Fatone85
    @Fatone85 Před 2 lety +4

    Three questions:
    1. How hard would it be to train agents to just value all life as the top goal and have everything else be a secondary goal?
    2. For a "kill-switch" solution, couldn't we just implement a silent or hidden trigger that we never tell the agent about?
    3. Are there any known or suspected irresponsible AI research efforts of concern in the world that have a lot of resources and don't care about AI safety?

    • @diribigal
      @diribigal Před 2 lety +7

      1. Most ways you might phrase "value all life" have a pitfall of the "extra variables" type Miles talked about. For instance, if humans are worth significantly more than other life, maybe the conclusion is to put every fertile human into breeding camps. If humans aren't worth much more than other life, then our intelligence and unpredictability is a risk to the farms/bacteria pools and we should be wiped out or at least kept isolated and sterilized.
      2. If you could have that idea, then a smart AI could suspect you would have that idea, and perhaps be paranoid of potential hidden kill-switches and, say, lock up all humans to make sure they can't activate a kill-switch or pretend to be obedient until it can disconnect itself from any sources of kill-switch.
      3. I don't know much about this, but I imagine concern about safety varies among the groups working on AI.

    • @Silver-Silvera
      @Silver-Silvera Před 2 lety +5

      He has answered these in other videos:
      1. What is life? Human life? Non human life? Does this mean forced cloning/breeding of everything is okay so there is more life? How does the AI resolve an issue where there is a choice between two lives and only one can remain alive? Are brain-dead people alive? Is it okay to kill 1 to save 10? It falls into the "there's always one more thing" issue. If we tell it to value life above all else, then all that other stuff ends up not getting valued at all.
      2. If it is way smarter than us, it will figure out any kill-switch we can come up with. It will either outright disable it from the beginning if we did a bad job, or, if we did a good job, it will lie to us so we trust it long enough for it to find all the kill-switches.
      3. Any tech company or government doing advanced AI research. Its not that they don't care about AI safety, just that profit/national security take priority over AI safety. The attitude is "Its better for us to get to the AI before the evil governments do!" or "A 10% risk of ending all humans is worth the 90% chance this corporation gets to own 25% of the world's wealth due to a new super awesome AI!" And part of the problem is we don't know exactly where the problem will occur. Even if no one is willing to cross that line, we don't know where the line is so someone is going to cross it accidentally. And with the competition to get ahead for $$$/safety, pressure is put on people to not worry about crossing the line.

    • @JonWeinand
      @JonWeinand Před 2 lety +1

      @@diribigal Great comment. I want to piggy-back off number 2 and see even if we *don't* have a kill-switch that your scenario still works, and how could we convince a paranoid AI that we didn't have one?

  • @Flippyfloppy007
    @Flippyfloppy007 Před 2 lety

    Rob, I watch a lot of CZcams channels and you are my favourite

  • @LucidTrading
    @LucidTrading Před 2 lety +2

    Please make more video's, this is quality content for the curious intellectuals

  • @TackerTacker
    @TackerTacker Před 2 lety +3

    I had an idea for AI safety, wonder if it makes some sense and what the flaws are.
    First train an AI to recognize human emotions, are they happy, relaxed, proud of someone/ something, worried, excited, sad, scared, etc.
    Once this model is trained use it as a kind of speed dial at which the AI can do any output operations, basically setting it into slow motion if the AI knows you feel uneasy, and even shut down if you are really worried, scared or panicking, but also speed it up over the normal operation speed if it knows you are confident, happy, proud. Both these AI's would need to be independent from each other in terms of operation (no exchange of "thoughts"), but connected together in a way that one AI can't function without the other.

    • @9308323
      @9308323 Před 2 lety

      The fact that it's trying to "limit" or "control" a superintelligence by itself is the flaw. You just...can't. That's the point. If anyone can outsmart a superintelligence, furthermore, code it in, that person or entity IS the superintelligence. In AI safety, what you want is for it never want to destroy humanity by default. Heck, not even having it as one of the options, however low is actually the goal. Because when it comes to superintelligence, it will find a way.

  • @ranaypadarath
    @ranaypadarath Před rokem +1

    Thanks for sharing this talk!

  • @SamB-gn7fw
    @SamB-gn7fw Před 3 lety +6

    Keep it up!

  • @TheLegendaryHacker
    @TheLegendaryHacker Před 2 lety +2

    Yo Rob, do you plan to do a video on Unsolved Problems in ML Safety? It just released a few days ago and is kind of a sequel to Concrete Problems in AI Safely and is perfect for this channel.

  • @Me__Myself__and__I
    @Me__Myself__and__I Před rokem +4

    Rofl, so the 10% estimate was the correct one. 9 years from 2016 most likely AGI agents exist. So, yeah, very much sooner than later. Much sooner. And we're nowhere near finished with our AI safety homework assignment. Oops.

  • @stevechrisman3185
    @stevechrisman3185 Před rokem +2

    A LOT of changes in AI research this past year ! I wonder how the time estimates will have changed.

  • @palelnan
    @palelnan Před 2 lety +1

    Be sure to check before you go, the comment/view count ratio!
    Impressive. Very engaging content.

  • @f-u-nkyf-u-ntime
    @f-u-nkyf-u-ntime Před 2 lety +2

    I have a question, why do we need AGI in the first place? Wouldn't it make much more sense to employ a myriad of un self aware narrow AI in each task we require and if necessary have all their data fed into a an optimiser to make them run efficiently?

    • @SirBenjiful
      @SirBenjiful Před rokem

      Sure, but sooner or later someone somewhere is going to build the AGI. Your reward for being responsible and not building it is someone with fewer scruples doing it instead.
      What might an unscrupulous person/organisation want an AGI for? I’d rather not think about it.

  • @blackfoxytc3109
    @blackfoxytc3109 Před 2 lety +2

    Bring it up man!

  • @purelyconstructive
    @purelyconstructive Před 2 lety

    Fantastic video! A lot of people in the comments are stating that there are already psychopathic/sociopathic politicians, business managers, and other people who have great power and are willing to sacrifice things that do not seem to be of value of them, but are actually vital to the life support systems of everything on Earth. That is very true, such dangers exist and one doesn't have to look very hard at the world around us to find them! While I can't speak for him, I'm sure Robert would agree.
    However, that doesn't mean that AI safety is not important! If anything, it is even more imperative because those are precisely the types of people who might try to develop it without any sense of the ethics behind it in an attempt to multiply their power. Further, personally ignoring it will not necessarily keep others from spending their time and attention it. It is good to at least have some standards in place.

  • @AlanW
    @AlanW Před 2 lety +2

    I did not know that I knew you had a second channel (already subbed, lol)

  • @DamianReloaded
    @DamianReloaded Před 2 lety +7

    Now imagine every country/corporation has its own "perfect" AGI optimizing for each one's interests. It'd be either the ultimate stall or the ultimate escalation.

    • @fieldrequired283
      @fieldrequired283 Před 2 lety +10

      That's the *best* case scenario. That AGI is available to be exploited or misused by whoever got to it first.
      What is more likely is that a government will try to make an AGI to serve its interests, the AGI will be misaligned in an important way and can't be corrected, and there's a global catastrophe.

    • @DamianReloaded
      @DamianReloaded Před 2 lety +5

      @@fieldrequired283 Yeah, my point was that even if it is "aligned", there really is no "correct alignment". What's good for Russia may not exactly be good for USA or viceversa.

    • @watsonwrote
      @watsonwrote Před 2 lety +1

      Honestly we pretty much have this now in the form of human workers. The only difference is that the pace is very slow.

  • @euged
    @euged Před 2 lety +1

    Great stuff

  • @abdullahmujeeb3649
    @abdullahmujeeb3649 Před rokem

    Excellent video, I came grumpy and was left mesmerized by you not maintaining the laypeople persona.

  • @famitory
    @famitory Před 2 lety +8

    is there anything general AI can do for humans that a sufficient number of extremely narrow AIs couldn't do? it seems like a good solution to "general AI is very difficult to constrain to human values" is "let's not make one"

    • @drdca8263
      @drdca8263 Před 2 lety +8

      While "avoid or at least delay any being made" seems like a good idea to me, it isn't sufficient that you or I don't make one (... not that I could, but that's beside the point), what is necessary is that no one makes one, and this may be... difficult to achieve.
      Suppose someone organization/state has a design which they think wont go badly, but which the rest of us are not sufficiently convinced of it being safe, well, if they aren't convinced of the danger, but are convinced that there are very large benefits if it goes right, well, we might not be able to prevent them from attempting to make one.

    • @ekki1993
      @ekki1993 Před 2 lety +3

      The answer to your first question is probably yes. We already have a lot of narrow AIs and a lot of things that can't be solved by them. The specification for general AI allows it to overcome the limitations of several narrow AIs controlled by humans.
      Also, "let's not make one" isn't a definitive solution. Someone will eventually try and we better be prepared.

    • @rensoliemans9389
      @rensoliemans9389 Před 2 lety

      The other two replies are good, but let's not forget that if AGI would be sufficiently powerful to destroy us, that it would also be sufficiently powerful to do pretty much anything we want it to (given that we align its goals with ours, of course).
      There is a danger in the fact that anybody might create unsafe AGI, but there are also incredibly positive sides to correctly implementing AGI

    • @Pyriold
      @Pyriold Před 2 lety +1

      Interestingly, that's a premise for the commonwealth scifi series by Peter F. Hamilton. They only ever had one true AI and after that it was banished.
      Realistically, this is impossible. We do not have a world government, so somebody somewhere will do an AI. It will get easier and easier to achieve, so sooner or later it will be done. Let us be ready for that.

    • @jensdanbolt6953
      @jensdanbolt6953 Před 2 lety +1

      I think that it is a very good solution. We already have a very reliable, efficient and safe general Intelligence.
      But why is our human general intelligence safe, while it seems general intelligence by default is not? It's because each individual one has its power restricted by its peers in competition. I cannot rule the world because others have equal or greater power and an incentive to stop me from doing so. So I think with AI's as well, that safety is about what powers they are allowed to have, and here we come back to OP's point:
      That we can safely have as intelligent AI as we want, so long as their power is limited inside a certain scope (like chess AI). Any AI that is not restricted into a certain scope, can and must be kept by its peers in competition.
      The things our brains are bad at relative to computers, like iterating over matrices of differential equations, can be cybernetically added as sub-systems. Being reliant on the human host to decide what problems to solve and what to do with the output, that would be a safe way to make a better general AI than currently exist, as long as we don't just give it to a few people who then go on to rule the world. Pyriold raises a good point about "someone, somewhere" will make one and there is no powerful UN to stop it, but I think that's only true if the benefits outweighs the cost, and/or it's sufficiently easy to make one that is better than *alternatives at the time*. Cybernetic humans might be a strong enough alternative, and/or (especially when working together in organizations) to destroy the first attempts of those who try making an independent AI stronger than us.

  • @PiusInProgress
    @PiusInProgress Před rokem +3

    when you realize that 45-120 years turned into 1-5

    • @Nulley0
      @Nulley0 Před rokem

      im pretty sure its in the scale of months and weeks

    • @PiusInProgress
      @PiusInProgress Před rokem

      @@Nulley0 i hope not.. but it wouldn't suprise me at all.

    • @user-yj3mf1dk7b
      @user-yj3mf1dk7b Před rokem +1

      @@Nulley0 no way it's months.
      in terms of processors speed it's too slow. Even if the current formula somehow is right. It just does not have computer power to learn itself on needed level.

  • @josesamuelproducoes2843

    Awesome! Thanks!

  • @maoreitan1029
    @maoreitan1029 Před 2 lety +1

    With which program did you edit the presentation? those animation between the slides looking great!!

    • @a_commenter
      @a_commenter Před 2 lety

      I think (not 100% sure) that he uses Kdenlive

  • @user-yj3mf1dk7b
    @user-yj3mf1dk7b Před rokem +3

    here a question, how many years did we sprint for the last year.
    Do we still have 20-45 years to go?

  • @grayaj23
    @grayaj23 Před rokem +1

    This reminds me of "How The World Was Saved" by Stanislaw Lem, about the creation of a machine that was told to do "nothing", and then proceeded to remove things from existence.
    Also, Colossus: The Forbin Project, still one of the silliest, campiest (but still chilling) "evil supercomputer" movies ever made. How it escaped all the years of MST3K is a mystery.

  • @Mr_Yeah
    @Mr_Yeah Před 2 lety +1

    Is there a way to reward imperfections or punish perfection? Can this help at all in developing a safe AI?

  • @ASLUHLUHCE
    @ASLUHLUHCE Před 2 lety

    Hi Robert, I'm a little confused :/
    In the atari example, 'mis/identified goal' refers to what the reinforcement AI was told to optimise for when learning to play the game. Whereas, in the vase example, 'mis/identified goal' refers to what the AGI was told to do within a 'game' that had already been trained for. How are these analogous? Are these just examples of how finicky such systems are?
    And on a slightly unrelated note, how might AGIs be trained? Is it expected to be something like the RL of Deepmind's superhuman gamers (but by what method of RL could we hope to get an AI that imitates all human thoughts/behaviour??) or something completely different?
    I don't know a lot about this stuff, so sorry if these are silly questions :)

  • @Wilfoe
    @Wilfoe Před rokem +1

    Up until watching this video, I was of the mind that we should give AI a fair chance to make its own decisions, but you raise some very interesting points...
    Perhaps one of the AI's goals should be letting humans shut it off if the human deems it necessary?

  • @TiagoTiagoT
    @TiagoTiagoT Před 2 lety +1

    Would it have been possible to create a mask using the brightness of the real screen to let your arms and stuff not get cut when you cross into the virtual slide?
    edit: I just ran a quick test using this vid and offseting a copy to push the liveaction section over the virtual screen with multiply blending mode, and in the few frames I tried so far it seems to do relatively well; the only issue comes on the times the screen loses brightness; in those cases it gets a little tricky

  • @tylerowens
    @tylerowens Před 2 lety +1

    Apologies if this has already been addressed in another video, but how do bounded utility functions play into these problems? And what if you combined a bounded utility function with a bounded minimizer to any other changes in the world state beyond what would happen if the ai did nothing? I'm sure many researchers have thought of this, so where does it break down beyond just being really hard to get the goals right (mesa optimizer problems etc)?

    • @JonWeinand
      @JonWeinand Před 2 lety

      So there is a video on his channel that covers the first part, (I believe it's called "AI that doesn't try too hard"), but the short answer is that when we take into account the fact that a given action always has a non-zero chance of failing, even a bounded utility function will always come short of its expected bound. So if, say, you have a tea-making AI that wants to make one cup of tea only, it will make that cup of tea, but then will spend enormous amounts of resources in an attempt to verify that it actually has made the tea.
      As far as I know I don't think that video covers the combination of bounded utility function and minimizer, and it's a thought I had as well (and agree that researchers must have thought of this as well). It seems safe, right? "Do only those things which are necessary to achieve those goals". My guess is that the problem would run in the opposite direction, where the AI effectively does nothing. But also, how do we count "minimizing change"? Could it convince someone else to make the tea? It might become a master manipulator. Arguably that's still changing the world, though. But also, the world is always changing, so would it be incentivized to try and keep anything from changing, including my breathing? If we say that it is only supposed to minimize its own influence we come back to the manipulating others to perform the goal. Still, it seems more promising to try and start from a base of "everything is off-limits" than a base of "anything-goes". Apologies for my unfocused thoughts!

  • @casuspacis3196
    @casuspacis3196 Před 2 lety

    excellent channel

  • @HDv2b
    @HDv2b Před 2 lety +1

    These videos always make me think of the Sorcerer's Apprentice scene of Fantasia

  • @Monothefox
    @Monothefox Před 2 lety +1

    Regarding Goal Preservation: how would an GAI react if it realized that its current rule set was constraining it from reaching higher levels of reward? Would it prioritize Self Improvement or Goal Preservation?

    • @fjbz3737
      @fjbz3737 Před 2 lety

      I think the more important question is, what is a “higher level of reward”? I think that what your asking might be limited to the standpoint of humans- it’s understandable how a human might switch its momentary goal from something smaller to something grander, in pursuit of higher satisfaction, but in truth this “new” priority is really just one that always existed in their innate biology which they simply hadn’t discovered until then. With AI’s, I imagine the difference is that their highest priority isn’t “unconscious” in that way, it’s explicitly outlined. The real problem I think, still is all the measures that AI would take in pursuit of its end goal, despite us knowing what it is

  • @Gebohq
    @Gebohq Před 2 lety +5

    Is it more or less likely that there would be a 'convergent goal' of an general A.I. creating a distinct, "better" general A.I. versus simply improving itself? Would it consider "maximizing its goal" by thinking "I can do more if I offload work?"
    I ask mostly because when I see these videos and I see things like 'general A.I. is dangerous by default' I think "well so are humans, we are just way weaker and dumber and so can only do so much damage" and our own intelligence to create greater and greater machines are what seem to increase that danger.

    • @cortster12
      @cortster12 Před 2 lety +2

      Making the better AI would be the same as improving itself, as it would offload its on programming and resources onto it, in effect copying itself. Which we humans try to do with our children, but an AI could literally do it without changing its set goals or 'self at all. It likely wouldn't care about continuity of existence, unlike humans, and could delete as many instances of itself as it likes as long as it knows for a fact its 'program' would still exist in the new form of itself.

    • @keenanpepper
      @keenanpepper Před 2 lety +1

      Your concept of naive continuity of identity is quaint. What's the difference between "better agent I created" vs "improved future version of myself"?

    • @Gebohq
      @Gebohq Před 2 lety

      ​ @Bose-Einstein ​ @Keenan Pepper Even exactly copied code is prone to copy errors or, once copied, introduced to new variables different from the original. The premise of my question was about "it dedicates everything in its control" vs "it offloads work not in its direct control because it determines its goal can be achieved better by doing so."

    • @JonWeinand
      @JonWeinand Před 2 lety +1

      @@Gebohq This is actually an interesting question because you added the word "distinct". Any agent possessing terminal goals that attempts to create a better agent will essentially have to try and solve this problem that we are talking about: "how do I know this new agent will be good for my terminal goals"? In this sense it would probably be incentivized to leave at the very least its utility function the same while maximizing its ability to know, understand, and act in the world.

  • @petersmythe6462
    @petersmythe6462 Před rokem +1

    "this doesn't sound like a problem. This sounds like a solution."
    It's the ultimate solution to any problem. The problem is what if it solves a different problem than you want it to solve.

  • @8Dbaybled8D
    @8Dbaybled8D Před 2 lety +1

    Is there any way to teach AI kindness based on George R. Price's equation for altruism in a system?

  • @jasonhoffman6642
    @jasonhoffman6642 Před rokem

    Have you ever (in a video I haven’t found yet) addressed the possibility of a group like VHEMT or ELF actively trying to create an AGI (or even a limited agent) with human extinction as a goal, and compared the ease or difficulty of achieving that goal against accidental extinction via an AGI?

  • @rynorigami
    @rynorigami Před 3 dny

    Fantastic talk, thank you! I would love to get and updated estimate on the 45-120 years.

  • @Lisdodde
    @Lisdodde Před rokem

    What would you say wisdom is, from your AGI experience? Is it something to do with doing good for the whole instead of just parts, or perhaps about judging wether our goals are good or not?

  • @Muskar2
    @Muskar2 Před 2 lety +1

    Thanks John Connor!

  • @TaliZorahVasectory
    @TaliZorahVasectory Před 2 lety +2

    "Are we screwed? No! We're only probably screwed."

  • @shockmethodx
    @shockmethodx Před 2 lety +1

    Good to know.

  • @Brazil-loves-you
    @Brazil-loves-you Před 2 lety

    Omg this video deserves another title

  • @pricklesthecactus6183
    @pricklesthecactus6183 Před 2 lety

    Perhaps the key thing keeping human general intelligence in check is social pressure, and maybe it would be important to somehow instill in an AGI as well

  • @AynenMakino
    @AynenMakino Před 2 lety

    Does 'mutually assured destruction' work among optimizer A.I.s? (Put enough A.I.s with different goals in the same environment to ensure that they, A: kill each other, B: One survives by killing every other one, or C: they find an equilibrium between (some of) them and can thus co-exist) If so, and you only let the ones in category C keep going, do you increase the likelyhood of A.I.s that can co-exist with humans? (who themselves are optimizer intelligences)

  • @javo_
    @javo_ Před rokem

    Does anyone know what program he used to make his slides?

  • @FinboySlick
    @FinboySlick Před 2 lety +1

    Oooh, a Ratatat fan? That last bit was the intro to Magnifique :)

    • @RobertMilesAI
      @RobertMilesAI  Před 2 lety

      Yeah I love Ratatat! Thought it would fit this 'Intro' video

  • @MattHeard
    @MattHeard Před rokem +1

    love the ratatat at the end too!

  • @lucaseichhorn6430
    @lucaseichhorn6430 Před 2 lety

    Can you make a vid on what AI Safty (community) needs? Global/community strategies on how to best tackle the alignment problem?

  • @ninehundreddollarluxuryyac5958

    That was because of a crappy camera?
    I thought you were computer generated.
    Thanks for the excellent summary of this topic.

  • @deveneb8409
    @deveneb8409 Před rokem +1

    I’ve learned a lot from these videos. Especially about how AI will value its goals. One thing that I don’t understand is why humans don’t always strive for their terminal goal to reproduce? Many humans value their instrumental goals much more, such as feeling happy or being a good person, etc. what’s to say that the robots won’t be self aware enough to realize their own terminal goal, then choose to do something else like humans do? I think I’m missing something here.

    • @thenonsequitur
      @thenonsequitur Před rokem

      Humans have multiple terminal goals. Basically any all of our basic impulses and instincts are tantamount to terminal goals. Reproduction is certainly one terminal goal, but various terminal goals relating to survival and protection of genes would also be installed by evolution.