RMSprop Optimizer Explained in Detail | Deep Learning

Sdílet
Vložit
  • čas přidán 27. 07. 2024
  • RMSprop Optimizer Explained in Detail. RMSprop Optimizer is a technique that reduces the time taken to train a model in Deep Learning.
    The path of learning in mini-batch gradient descent is zig-zag, and not straight. Thus, some time gets wasted in moving in a zig-zag direction. RMSprop Optimizer increases the horizontal movement and reduced the vertical movement, thus making the zig-zag path straighter, and thus reducing the time taken to train the model.
    The concept of RMSprop Optimizer is difficult to understand. Thus in this video, I have done my best to provide you with a detailed Explanation of the RMSprop Optimizer.
    ➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
    ▶ Momentum Optimizer in Deep Learning: • Momentum Optimizer in ...
    ➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
    ▶ Watch Next Video on Adam Optimizer: • Adam Optimizer Explain...
    ➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
    ✔ Improving Neural Network Playlist: • Overfitting and Underf...
    ✔ Complete Neural Network Playlist: • Neural Network python ...
    ✔ Complete Logistic Regression Playlist: • Logistic Regression Ma...
    ✔ Complete Linear Regression Playlist: www.youtube.com/watch?v=mlk0r...
    ➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
    Timestamp:
    0:00 Agenda
    1:42 RMSprop Optimizer Explained
    5:37 End
    ➖➖➖➖➖➖➖➖➖➖➖➖➖➖➖
    Subscribe to my channel, because I upload a new Machine Learning video every week: czcams.com/channels/JFA.html...

Komentáře • 28

  • @user-tt1ox5ls2d
    @user-tt1ox5ls2d Před 7 měsíci

    thankyou so much for uploading these videos, your explanations are easily understandable

  • @jyotsanaj1425
    @jyotsanaj1425 Před rokem

    Such clear explanation

  • @syedalimoajiz1179
    @syedalimoajiz1179 Před rokem

    how to initialize value of Sdw and Sdb?

  • @christopherwashington9448

    Hello thanks for the info. But you didn't mention the purpose of the square for the gradient.

  • @vishalchandra1450
    @vishalchandra1450 Před 2 lety +3

    Hand's down,best explanation ever:)

  • @yahavx
    @yahavx Před rokem

    What is (dw)^2?

  • @keshavmaheshwari521
    @keshavmaheshwari521 Před 2 lety +2

    What is S?

  • @kasyapdharanikota8570
    @kasyapdharanikota8570 Před rokem +1

    your channel is highly underrated, it deserves a lot more audience

    • @CodingLane
      @CodingLane  Před rokem

      Thank you for this considerate comment 😇

  • @himtyagi9740
    @himtyagi9740 Před 2 lety +1

    waiting for SVM since you explain so nicely..thnks

    • @CodingLane
      @CodingLane  Před 2 lety

      Thank you! I will upload SVM video after finishing RNN series

  • @Vinay192
    @Vinay192 Před 2 lety +3

    Hi Sir, any plan of uploading videos on support vector machines? If yes then, please try to cover the mathematical background of SVM as much as you can ...
    Anyway your content is really appreciable...Thanks !

    • @CodingLane
      @CodingLane  Před 2 lety

      Thank you so much for your suggestion! Yes, I will be making video on SVM and covering mathematical details behind it.

  • @KorobkaAl
    @KorobkaAl Před 2 lety +1

    You are the best, thanks dude 🤙

  • @Maciek17PL
    @Maciek17PL Před 2 lety +1

    If situation with w and b would be opposite values of gradients on the vertical axis were small and values on horizontal axis where large would RMSprop slow down the training by making vertical axis values larger and horizontal axis values smaller?

    • @CodingLane
      @CodingLane  Před 2 lety +1

      No no… it will still make the training faster. Vertical horizontal is just an example i am giving. Realistically, it can be in any direction. In every direction, its gonna work the same way.

  • @mugomuiruri2313
    @mugomuiruri2313 Před 7 měsíci

    good

  • @ueslijtety
    @ueslijtety Před rokem +1

    Hi,Is it correct that you set the vertical coordinates to w and the horizontal coordinates to b? I think it should be the other way around.Because whether the goal can be reached in the end depends on w rather than b.

    • @CodingLane
      @CodingLane  Před rokem +1

      Hi… neither we set vertical to w nor b. Its just an example given… in a model.. there are many axis, not just x and y if we have more than 2 number of features. So a model can take any axis as any w or b. and it doesn’t matter as well which axis is for waht

    • @ueslijtety
      @ueslijtety Před rokem

      @@CodingLane thanks!So in practice this is not going to be a 2D planar image but a multidimensional image?And which parameters can determine the point of convergence in gradient descent?W OR b?

    • @minister1005
      @minister1005 Před 11 měsíci

      So i guess what he means is that if you get a high gradient, you will be updated a lower amount and if you get a low gradient, you will be updated a higher amount.

  • @marccasals6366
    @marccasals6366 Před 2 lety +1

    You're incredible

    • @CodingLane
      @CodingLane  Před 2 lety +1

      Thank You Marc! Glad you found my videos valuable.

  • @user-rx9kq2wi8n
    @user-rx9kq2wi8n Před 10 měsíci

    Explain ADMM also

  • @MrMadmaggot
    @MrMadmaggot Před rokem +1

    Man and what kind of LOSS should I use when training using RMSprop optimizer?