L13/5 ResNet (Residual Networks)

Sdílet
Vložit
  • čas přidán 29. 08. 2024
  • Dive into Deep Learning
    UC Berkeley, STAT 157
    Slides are at
    courses.d2l.ai
    The book is at
    www.d2l.ai
    Residual Networks

Komentáře • 11

  • @nikhilpandey2364
    @nikhilpandey2364 Před 5 lety +27

    Hey Alex, please don't stop uploading. It is really hard for us engineers to understand the cutting edge. But this video made things so easy. Thanks

  • @paulieddd
    @paulieddd Před 4 lety +1

    Eventhough I still find it hard to really be able to explain this in my own words, this definitely helped me with my first understanding about resnets. Thank you for sharing this!

  • @puttatidam.1819
    @puttatidam.1819 Před 3 lety

    Ive been watching other youtube videos for about an hour and as much as I understood the mathematical and the research problems, they couldn't clearly explain intuitively why it is useful aside from, oh, it just use skip connections to give higher accuracy at deeper depth (ignoring what is the point of identity mapping and why we need skip connection and so forth). Your video blew me away and gave me that eureka moment. Thank you, sir.

  • @fractelet
    @fractelet Před 5 lety +7

    Thanks for sharing Alex

  • @petarulev6977
    @petarulev6977 Před 2 lety

    I don't understand the different function classes (generic and nested), the universal approx. th states that even a single-layer nn can learn any function.

  • @aghazi94
    @aghazi94 Před 4 lety

    sir your work is appreciated

  • @neocephalon
    @neocephalon Před 4 lety

    So... not much difference in terms of what you need to use for each then. I was hoping that resnext architecture would be more efficient and require less computational resources to achieve similar or higher accuracy results.

  • @Niels1234321
    @Niels1234321 Před 3 lety

    I am sceptical of this explanation, because convolutional layers and fully connected layer can both learn the identity function, so a network consisting of those falls in the nested function class category already without the residual trick.

    • @smolix
      @smolix  Před 3 lety +1

      Actually, learning the identity function might not be quite so easy due to the subsequent nonlinearlity. They can sort-of kind-of learn it but not easily. On the other hand, if you have an identity function directly available, it's a much better inductive bias.

  • @kallolroy5029
    @kallolroy5029 Před 4 lety

    Awesome

  • @broccoloodle
    @broccoloodle Před 2 lety

    How can a lecturer explain things but his words were never definite?
    Even the first figure is wrong, deeper networks can represent the shallower networks just fine. They just have different inductive bias