Video není dostupné.
Omlouváme se.

[DeepBayes2019]: Day 5, Lecture 3. Langevin dynamics for sampling and global optimization

Sdílet
Vložit
  • čas přidán 15. 08. 2024
  • Slides: docs.google.co...
    Materials: github.com/bay...
    Lecturer: Kirill Neklyudov

Komentáře • 20

  • @miladaghajohari2308
    @miladaghajohari2308 Před 3 lety +7

    great presentation. It gave me a lot of insight especially given the current developments in score-based models and generation with them.

  • @leonardohuang
    @leonardohuang Před 4 lety +1

    Fantastic talk! Clear, informative and vivid.

  • @Tryndamereization
    @Tryndamereization Před 2 lety +1

    Very good presentation. Thank you

  • @jfndfiunskj5299
    @jfndfiunskj5299 Před 4 lety +2

    Very good. Well done.

  • @PradeepBanerjeeKr
    @PradeepBanerjeeKr Před 3 lety +2

    Simply super!

  • @nadineca3325
    @nadineca3325 Před 4 lety

    Thank you for this informative discussion and amazing presentation!

  • @sayarerdi
    @sayarerdi Před 3 lety +4

    Could you also share the codes of your animations?

  • @MrDudugeda2
    @MrDudugeda2 Před 4 lety +1

    great talk!

  • @hakiim.jamaluddin
    @hakiim.jamaluddin Před 3 lety +1

    Super clear, thanks!

  • @user-zr4ns3hu6y
    @user-zr4ns3hu6y Před 3 měsíci

    Fantastic,,,, i want to go samsung

  • @bingbingsun6304
    @bingbingsun6304 Před rokem

    very good talk!

  • @qiuhaowang2063
    @qiuhaowang2063 Před 3 lety

    Helpful talk! Thank you . I just don't understand why there is a 'div' in the equation at 16:45. Could you show me some references?

    • @k_neklyudov
      @k_neklyudov Před 3 lety

      Remember that div(f) is just a convenient notation for \sum_i df_i/dx_i. If you'll try to take the jacobian of the formula for x' you will end up evaluating the determininant of a large matrix, and there you should see that some terms of this determinant are negligible: o(dt). Thus, in some sense, the diagonal of the matrix (which is 1+df_i/dx_i*dt) plays the main role and, thus, div appears.

    • @ivanmedri288
      @ivanmedri288 Před 2 lety

      I know it is old post but could you solve this? I am stuck on the same part.

  • @abhishekmaiti8332
    @abhishekmaiti8332 Před 3 lety

    At 34:01, are \thetas parameters of the model or the sampled data points from a model which is parameterised by \theta?

    • @k_neklyudov
      @k_neklyudov Před 3 lety +1

      thetas are the parameters of the model, which could be a regression or a classifier in this case. hence, here we sample the parameters of a model (say weights of linear regression), and then we should average the predictions over all these sampled parameterrs.

    • @abhishekmaiti8332
      @abhishekmaiti8332 Před 3 lety

      @@k_neklyudov I see, thanks a lot for the clarification.

  • @umountable
    @umountable Před 4 lety +5

    Maybe its useful to motivate a little more before showing only slides packed with equations for half an hour straight

    • @MNasirAziz
      @MNasirAziz Před 4 lety

      I understand you. But you can look who he uses final derived equation by just skipping derivation.

    • @k_neklyudov
      @k_neklyudov Před 4 lety +2

      Hi Stefan! Thanks a lot for your feedback! I totally agree that this talk taking separately lacks motivation and practical examples. The reason is that this talk was designed as a part of Day 5, which is about MCMC. You can also find the discussion of the Langevin dynamics in this talk czcams.com/video/Q_Bi2H9NKzc/video.html.