The Alberta Plan for AI Research: Tea Time Talk with Richard S. Sutton

Sdílet
Vložit
  • čas přidán 10. 09. 2024
  • Artificial general intelligence (AGI) is one of the grand ambitions of much machine learning research - the benefits of an artificial intelligence agent that can learn and understand tasks at a near-human level could transform the world as we know it. It would be one of the greatest scientific achievements of all time, argues Amii Fellow, Canada CIFAR AI Chair & Chief Scientific Advisor Rich Sutton, and will happen sooner than we might think.
    Watch Sutton lay out the Alberta Plan - a twelve-stage roadmap that he things could lead to a fuller understanding of intelligence.

Komentáře • 13

  • @borntobemild-
    @borntobemild- Před rokem

    When you found the gizmo, it was a good metaphor on how you are freeing up the agent in the world with technology.

  • @Gabriel-oi6zb
    @Gabriel-oi6zb Před rokem +2

    Minute 11: Our interaction with the environment is not continual. There are special training periods: sleep -- a crucial step in all mammals, might even extend back to all vertebrates.

    • @erkinalp
      @erkinalp Před rokem +1

      There are total insomniacs who cannot sleep for years, but they do not exhibit significant learning-related disabilities. Hence sleep should not be considered the only factor for unlearning the falsehoods.

    • @Gabriel-oi6zb
      @Gabriel-oi6zb Před rokem +2

      ​@@erkinalp you might want to wiki that, total insomnia (also called fatal insomnia because you die from it) causes hallucinations.

    • @schok51
      @schok51 Před rokem +2

      ​@@erkinalp Sources? I thought sleep deprivation and disorders were pretty universally harmful to cognitive abilities. You cannot simply not sleep and be healthy and functional.

  • @howtobe7460
    @howtobe7460 Před rokem +4

    This entire comment section looks AI generated 😂😂

  • @judgeomega
    @judgeomega Před rokem +3

    it seems a contradiction to say you want a model with no domain knowledge yet still having a reward function. doesnt knowledge of a reward imply knowledge of the domain of that reward?
    the amount of knowledge in the universe is nigh infinite, and we need that reward to anchor our focus on just that which has utility with respect to our goals(rewards).

    • @schok51
      @schok51 Před rokem

      I guess that's just semantics and that the point is that the reward function should encode all that is relevant about the domain?

    • @LionKimbro
      @LionKimbro Před rokem

      I was wondering the very same thing. What's your reward function? With ChatGPT, the score comes from "did I predict the next word accurately?" I have no idea what this system is going to use. One possibility is -- is it going to be an auto-decoder? Don't know.

    • @ArtOfTheProblem
      @ArtOfTheProblem Před 7 měsíci +1

      here I believe he means the "value function" defines the reward, specificially is it getting better or worse. It's not inputting an external reward. reward is part of perception and is learned by the value function (if you understand TD learning)

  • @googm
    @googm Před rokem

    "paltry worries like THE ECONOMY IS IN TROUBLE".

  • @richsutton366
    @richsutton366 Před rokem +13

    The slides can be found on my web site richsutton.com.