VGG Deep Neural Network Explained with Pytorch

SdĂ­let
VloĆŸit
  • čas pƙidĂĄn 16. 06. 2024
  • Depth in neural networks is a very important parameter. Before ResNet, the VGG network was able to prove its importance by scaling up AlexNet to 16-19 layers.
    In this tutorial, we'll take a look at the theory behind the architecture as well as a Pytorch implementation from the official documentation.
    Table of Content
    - The Importance of Depth in Neural Networks: 0:00
    - VGG Network Architecture: 0:57
    - VGG Network Training Regiment: 4:23
    - VGG Network Result: 5:19
    - VGG Pytorch Code Walkthrough: 7:41
    - Conclusion: 15:39
    Important Links:
    📌 Paper: arxiv.org/pdf/1409.1556
    📌 Github: github.com/yacineMahdid/deep-...
    Abstract:
    "In this work we investigate the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting.
    Our main contribution is a thorough evaluation of networks of increasing depth using an architecture with very small (3x3) convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
    These findings were the basis of our ImageNet Challenge 2014 submission, where our team secured the first and the second places in the localisation and classification tracks respectively. We also show that our representations generalise well to other datasets, where they achieve state-of-the-art results.
    We have made our two best-performing ConvNet models publicly available to facilitate further research on the use of deep visual representations in computer vision. "
    ----
    Join the Discord for general discussion: / discord
    ----
    Follow Me Online Here:
    GitHub: github.com/yacineMahdid
    LinkedIn: / yacinemahdid
    ___
    Have a great week! 👋

Komentáƙe • 8

  • @kukfitta2258
    @kukfitta2258 Pƙed 24 dny +1

    very cool thank you for the knowledge

  • @AbhishekSaini03
    @AbhishekSaini03 Pƙed 24 dny +1

    Thanks , how can we use VGG for 1D signal? Is it possible to use VGG for regression instead of classification, how?

    • @machinelearningexplained
      @machinelearningexplained  Pƙed 22 dny

      Hmmm depends, what’s the 1D signal about? Is it visual?

    • @AbhishekSaini03
      @AbhishekSaini03 Pƙed 22 dny +1

      It’s acoustic signal.

    • @machinelearningexplained
      @machinelearningexplained  Pƙed 21 dnem

      Ah then no, VGG shouldn’t be your pick here. It was expressively designed for image classification.
      Take a look at the various model on PyTorch made specifically for audio signal:
      📌 pytorch.org/audio/stable/models.html

    • @AbhishekSaini03
      @AbhishekSaini03 Pƙed 21 dnem +1

      Can’t we change output layer, activation function to do regression?

    • @machinelearningexplained
      @machinelearningexplained  Pƙed 21 dnem

      Yes you can, but the internal of the model is tailor built for image.
      If you are able to express your 1D signal input as an image I would say it might be worth it to try.
      However, there are other models made specifically for audio.