#032

Sdílet
Vložit
  • čas přidán 8. 09. 2024

Komentáře • 44

  • @oncedidactic
    @oncedidactic Před 3 lety +12

    Strikes me that someone sharing very engineering and problem-solving oriented statements can be both extremely useful from the standpoint of a practitioner and also touching on very profound, edge-of-understanding areas. You don't always get this dual experience, so I really appreciated hearing from a guest like that. :D

  • @snarkyboojum
    @snarkyboojum Před 19 dny

    I can't believe how much great content you've put out there on the internet. From one netizen to another - thank you!

  • @TheVigi99
    @TheVigi99 Před 3 lety +3

    As a speech researcher, it was insightful to hear dr simon's thoughts on simclr, following the same recipe on speech via simclr and similar frameworks, seems to a sample efficient way to learn across various modalities.

  • @drhilm
    @drhilm Před 3 lety +7

    So many practical insights. So fan to look under the hood of these popular papers.

  • @SawomirMucha
    @SawomirMucha Před 3 lety +1

    Brilliant content! On the topic of systematic way of finding the data augmentations, it feels like DA helps in overcoming "staticness" of the image domain (for natural images). As Dr. Simon noted, videos provide a lot of semantic information that's missing in the static image world. Looking at the objects, we naturally can find them "cropped" (obscured, behind other object) or in different colour schemes throughout the day. Knowing that the main bottleneck is compute, it makes the whole thing even more exciting (assuming the exponential compute growth keeps up). Big breakthroughs ahead of us!
    Thank you once again for this show, you're doing great job!

  • @snippletrap
    @snippletrap Před 3 lety +3

    I'm with Yann LeCun on this one -- I think regularized latent variable models are ultimately the way to go. They reduce the Kolmogorov complexity of the learned model, constraining the size of the solution space, whereas contrastive learning pushes on individual points. To shape the space you're going to need a lot of points...

  • @flamboyanta4993
    @flamboyanta4993 Před rokem

    Such a good discussion! Thanks!

  • @videowatching9576
    @videowatching9576 Před 2 lety

    Fascinating, I’m interested in hearing more about text to video generation

  • @smealzzon
    @smealzzon Před 3 lety +1

    Really enjoyed the short/condensed version. Looking forward to watching the long version to improve understanding. I really like this podcast format! Also great material!

  • @MrJorgeceja123
    @MrJorgeceja123 Před 3 lety +5

    Tim, your Whimsical notes are great! Can you release them public?

  • @abby5493
    @abby5493 Před 3 lety +2

    Another awesome and informative video. 😍

  • @Johan-qf8iv
    @Johan-qf8iv Před 3 lety

    Great insights. I really appreciate the relatable and grounded discussions during the whole session. Thank you!

  • @DavenH
    @DavenH Před 3 lety +2

    Thank you for this great segment. So many insights in those papers. I guess contrastive loss works best when we can use our own intuitions of the right inductive bias to augment the dataset, like Kilcher notes. I struggled with this part in an NLP project, which was concerned with learning sentence embeddings; how do you augment sentences? There aren't many simple surface-level tweaks you can do, analogous to changing the hue and brightness and cropping. You could swap synonyms and delete the odd adjective, but more than that what can you do without changing the meaning?

  • @quebono100
    @quebono100 Před 3 lety +4

    Wow, with every episode even better content :-O mind-blowing

    • @TimScarfe
      @TimScarfe Před 3 lety +1

      🔥🔥✌️✌️😊😊

  • @philipprenz9741
    @philipprenz9741 Před 3 lety

    1:23:04 Random sampling is not trickery. If you try to follow up the sequence "A cat is a" then suggestions with their probability are maybe ("is a", 0.4), ("cute animal", 0.3), ("awesome climber", 0.3).
    The first is a useless repetition and the latter two make sense. That means that most of the probability mass is concentrated on stuff that makes sense, but the most probable continuation doesn't. Sampling the most probable continuation ignores that most of the probablity rests on stuff that makes sense. When more outputs are available this effect can get more pronounced.

  • @ronjagring9691
    @ronjagring9691 Před 3 lety +1

    Thanks for this really nice podcast. Please continue :)

  • @Chr0nalis
    @Chr0nalis Před 3 lety

    The way I see it, data augmentation is a way of making up for the lack of suitable prior in our model (CNN). Data augmentation doesn't generate new data, but it does implicitly strengthen the weak prior of the CNN which can lead to better generalization, faster training, etc.

  • @alandolhasz7863
    @alandolhasz7863 Před 3 lety

    Great episode! Plenty of practical ideas to play with.

  • @crimythebold
    @crimythebold Před 3 lety

    Another fantastic one. I've got more papers to read about contrastive loss now

  • @georgepearse2680
    @georgepearse2680 Před 3 lety

    Could you create a series / episode on active learning? Hugely relevant to work in industry, particularly segmentation tasks the require expert labellers, but weirdly little research on it. Recently discovered the channel and love the work you do.

  • @kimchi_taco
    @kimchi_taco Před 3 lety +1

    Today is my lucky day. I found it!!

  • @fredguth1315
    @fredguth1315 Před rokem

    Please tell Simon that he can find type dispatching (function overloading) in fastcore from fastai. He is absolutely right that python is not the best language for de DL job. Anyway, I believe python will be default la gauge in the long run in the same way JavaScript is for web dev.

  • @shipper611
    @shipper611 Před 3 lety +1

    Stupid question maybe - if self similarity seems to be a pretty straightforward way to detect unnecessary layers and their exclusion would help - why isn’t a „self similarity check“ and corresponding auto deletion of layers not build into high level frameworks like Keras?

  • @alachance2010
    @alachance2010 Před 3 lety

    Loved it

  • @lethanh1122
    @lethanh1122 Před 3 lety +3

    Which tool to draw the graph at around 2:40 to 5:00 ?

  • @quebono100
    @quebono100 Před 3 lety +4

    I had some similar thought, that maybe the selfsimilarity that mandelbrot discovered has some implication in NN

  • @obsiyoutube4828
    @obsiyoutube4828 Před 2 lety

    amazing

  • @dr.mikeybee
    @dr.mikeybee Před 3 lety

    When there is no direct correlation in a layer, adding other layers may provide a combined feature correlation. Once the necessary combined correlations are created, there is nothing left to find. Additional layers just create unnecessary combinatoric burden.

    • @crimythebold
      @crimythebold Před 3 lety

      Agreed. Any volonteer to code that for next week ?

  • @florianhonicke5448
    @florianhonicke5448 Před 3 lety

    Thanks, you made my day! One usecase of the embeddings learned by contrastive learning is neural search (finding nearest neighbours).
    However, you can use many different attributes for contrastive learning. Would you just train a head on a pretrained network for each attribute? Or is the gain from fine-tuning the whole model worth it?

  • @SimonJackson13
    @SimonJackson13 Před 3 lety

    Not sure why you'd not use a mean/variance brightness contrast normalization on the colour channel histograms.

  • @marekglowacki2607
    @marekglowacki2607 Před 3 lety

    After representation learning there will be data augmentation learning ;)

  • @machinelearningdojowithtim2898

    First!! 🙌🙌

  • @quebono100
    @quebono100 Před 3 lety +1

    Second

  • @vladomie
    @vladomie Před 3 lety

    Please drop the background noise
    ... it's distracting and annoying.

  • @MarkoTintor
    @MarkoTintor Před 3 lety +2

    Frequent cut-in video snippets are distracting from the talk.

    • @machinelearningdojowithtim2898
      @machinelearningdojowithtim2898 Před 3 lety +1

      Thanks for the feedback! We are trying to achieve a certain format and are always experimenting - might have gone too far on this one. It's just a show teaser though, the full interview is always shown afterwards with no music or distractions, so just skip ahead.

    • @crimythebold
      @crimythebold Před 3 lety +4

      If it is the static picture inserts, I disagree. I dont always know what they are talking about, it helps to undertand

    • @Chr0nalis
      @Chr0nalis Před 3 lety +1

      I disagree as well. I think they stimulate the imagination / provide a canvas while you are thinking about the content that is being talked about.