ML For Nerds
ML For Nerds
  • 44
  • 139 141
YOLO-V4: MiWRC, CmBN, DROP BLOCK, CIOU, DIOU || YOLO OBJECT DETECTION SERIES
This video is about Yolo object detection family. This is about YoloV4 which is the most popular and widely used object detector in the industry. YoloV4 has the highest usage by industry for commercial purposes because of its optimal speed and accuracy. In this video, we discussed Multi-Input Weighted Residual Connections, Cross mini Batch Normalization, Drop Block Regularisation, types of IOU losses. These are all parts of Bag of Specials and Bag of Features in YoloV4.
YOLO Playlist:
czcams.com/play/PL1u-h-YIOL0sZJsku-vq7cUGbqDEeDK0a.html
Neural Networks From Scratch Playlist:
czcams.com/play/PL1u-h-YIOL0u7R6dg_d5O5M9HUj0SFjHE.html
Link to Papers:
YoloV4: arxiv.org/pdf/2004.10934.pdf
EfficientDet: arxiv.org/pdf/1911.09070.pdf
Cross Batch Norm: arxiv.org/pdf/2002.05712.pdf
DropBlock Regularization: arxiv.org/pdf/1810.12890.pdf
IOU Losses: arxiv.org/pdf/1911.08287.pdf
Chapters:
00:00 Introduction
02:00 Cross Mini-Batch Normalization
11:06 Multi-Input Weighted Residual Connections
17:50 Drop Block Regularization
25:57 IOU Loss
30:32 GIOU Loss
34:29 DIOU Loss
37:29 CIOU Loss
43:06 Conclusion
#yolo #yoloobjectdetection #objectdetection #yolov4 #yolov5 #yolov3 #yolov7 #computervision #imageclassification
zhlédnutí: 2 747

Video

Batch Normalization - Part 4: Python Implementation on MNIST dataset
zhlédnutí 509Před 11 měsíci
We have been discussing Batch Normalization in detail. We have seen why do we need Batch Normalization and we have dig deeper into how Batch Normalization works and also understood the significance of learnable parameters called Gamma and Beta which are Scaling and Shifting. We also saw the Backpropagation for Batch Normalization layer and how Batch Normalization works during inference without ...
Batch Normalization - Part 3: Backpropagation & Inference
zhlédnutí 1,4KPřed rokem
We have been discussing Batch Normalization in detail. We have seen why do we need Batch Normalization and we have dig deeper into how Batch Normalization works and also understood the significance of learnable parameters called Gamma and Beta which are Scaling and Shifting. In this video, we will see the Backpropagation for Batch Normalization layer and also see how Batch Normalization works d...
Batch Normalization - Part 2: How it works & Essence of Beta & Gamma
zhlédnutí 1,4KPřed rokem
We have been discussing Batch Normalization in detail. We have seen why do we need Batch Normalization in the previous video. In this video, we will dig deeper into how Batch Normalization works and also understand the significance of learnable parameters called Gamma and Beta which are Scaling and Shifting. Deep Learning Projects playlist: czcams.com/play/PL1u-h-YIOL0s2GYHiaemx7-o-iWrht_Lk.htm...
Batch Normalization - Part 1: Why BN, Internal Covariate Shift, BN Intro
zhlédnutí 3,5KPřed rokem
In this video, we dig deeper into “Why do we need Batch Normalization?” And Internal Covariate Shift. Deep Learning Projects playlist: czcams.com/play/PL1u-h-YIOL0s2GYHiaemx7-o-iWrht_Lk.html Neural Networks From Scratch in Python: czcams.com/play/PL1u-h-YIOL0u7R6dg_d5O5M9HUj0SFjHE.html Chapters: 00:00 Introduction 02:52 Issues with NN Training w/o BN 04:00 Internal Covariate Shift 04:28 What ar...
Neural Networks From Scratch - Lec 24 - Regression Losses - Mean Square Logarithmic Error
zhlédnutí 585Před rokem
Building Neural Networks from scratch in python. This is the twenty fourth video of the course - "Neural Networks From Scratch". This video covers most commonly used loss functions we use in regression problems. We discussed the important properties of log function and its significance to regression problems. We also saw Mean Square Log Error and its advantages and drawbacks. We also saw the py...
YOLO-V4: CSPDARKNET, SPP, FPN, PANET, SAM || YOLO OBJECT DETECTION SERIES
zhlédnutí 9KPřed rokem
This video is about Yolo object detection family. This is about YoloV4 which is the most popular and widely used object detector in the industry. YoloV4 has the highest usage by industry for commercial purposes because of its optimal speed and accuracy. In this video, we discussed about Backbone CSPDarknet-53, SPP, FPN, PANT and SAM modules. These are all parts of Bag of Specials in YoloV4. YOL...
YOLO-V4: Optimal Speed & Accuracy || YOLO OBJECT DETECTION SERIES
zhlédnutí 7KPřed rokem
This video is about Yolo object detection family. This is about YoloV4 which is the most popular and widely used object detector in the industry. YoloV4 has the highest usage by industry for commercial purposes because of it's optimal speed and accuracy. YOLO Playlist: czcams.com/play/PL1u-h-YIOL0sZJsku-vq7cUGbqDEeDK0a.html Neural Networks From Scratch Playlist: czcams.com/play/PL1u-h-YIOL0u7R6...
YOLO-V3: An Incremental Improvement || YOLO OBJECT DETECTION SERIES
zhlédnutí 8KPřed rokem
This video is about Yolo object detection family. In this video, we dig deeper into Yolo-v3 object detection model, which is an incremental update over YoloV2. This was the Sate of the art object detector and faster object detector back then when it was released. YOLO Playlist: czcams.com/play/PL1u-h-YIOL0sZJsku-vq7cUGbqDEeDK0a.html Neural Networks From Scratch Playlist: czcams.com/play/PL1u-h-...
YOLO-9000 - An Object Detector for 9000 classes || YOLO OBJECT DETECTION SERIES
zhlédnutí 3,6KPřed rokem
This video is about yolo object detection family. In this video, we will dig deeper into yolo9000 model which can detect objects from 9000 categories. This is an extension of YOLOv2 for detecting objects at largest scale. This uses Darknet 19 backbone for yolo. YOLO Playlist: czcams.com/play/PL1u-h-YIOL0sZJsku-vq7cUGbqDEeDK0a.html Neural Networks From Scratch Playlist: czcams.com/play/PL1u-h-YI...
YOLO V2 - Better, Faster & Stronger || YOLO OBJECT DETECTION SERIES || YOLO9000
zhlédnutí 12KPřed rokem
This is the second video of the series about YOLO object detection model family. This video digs deeper into YOLO-V2 paper which is an improvement over Yolo-V1. This version of yolo object detector is much more accurate and faster than yolo v1. They have made an new architecture named darknet 19 for backbone which is more accurate and less complex. YOLO Playlist: czcams.com/play/PL1u-h-YIOL0sZJ...
YOLO V1 - YOU ONLY LOOK ONCE || YOLO OBJECT DETECTION SERIES
zhlédnutí 33KPřed rokem
Hi Guys, I am starting a new series about YOLO object detection model family. This is not an overview series, we will dig deeper into every detail of these yolo object detectors. Everyone uses YOLO models, they are the state of the art models for object detection. Hope you learn something from these videos. YOLO Object Detection Series: czcams.com/play/PL1u-h-YIOL0sZJsku-vq7cUGbqDEeDK0a.html PD...
Neural Networks From Scratch - Lec 23 - Regression Losses - Smooth L1 Loss and Huber Loss Functions
zhlédnutí 1,5KPřed rokem
Building Neural Networks from scratch in python. This is the twenty third video of the course - "Neural Networks From Scratch". This video covers most commonly used loss functions we use in regression problems. We discussed Smooth L1 loss and Huber loss and their differences. Neural Networks From Scratch Playlist: czcams.com/users/playlist?list... Please like and subscribe to the channel for mo...
Remove the confusion once for all! Cost Function vs Loss Function vs Objective Function
zhlédnutí 1,1KPřed rokem
In this video, we have resolved the confusion between the most commonly used loss terms in machine learning. What is loss function? What is cost function? Are they same? Time stamps: 00:00 Introduction 00:22 What do you think? 00:35 Answer! Difference between them 02:02 Illustration with Example 03:10 One more difference 03:45 What is Objective Function then? 04:59 Conclusion #loss #objective #...
Neural Networks From Scratch - Lec 22 - MAE vs RMSE, Comparison with an Example
zhlédnutí 964Před rokem
Building Neural Networks from scratch in python. This is the twenty second video of the course - "Neural Networks From Scratch". This video covers the similarities an important differences between MAE loss and RMSE loss functions, How do we interpret them and which one to prefer. Code Link: github.com/MLForNerds/Neural_network_from_scratch/blob/main/MAE_RMSE.ipynb Neural Networks From Scratch P...
What is Numpy? Why Numpy arrays are faster than python lists?
zhlédnutí 570Před rokem
What is Numpy? Why Numpy arrays are faster than python lists?
MNIST Classification: Hands-on Project in PyTorch 1.12
zhlédnutí 525Před rokem
MNIST Classification: Hands-on Project in PyTorch 1.12
PyTorch Vs Tensorflow: Jobs, Research and Industries. Who is the winner in 2022?
zhlédnutí 1,6KPřed rokem
PyTorch Vs Tensorflow: Jobs, Research and Industries. Who is the winner in 2022?
MNIST Classification: Hands-on Project in Tensorflow 2.8
zhlédnutí 429Před 2 lety
MNIST Classification: Hands-on Project in Tensorflow 2.8
Building a Neural Network from scratch: MNIST Project (No Tensorflow/Pytorch, Just Numpy)
zhlédnutí 13KPřed 2 lety
Building a Neural Network from scratch: MNIST Project (No Tensorflow/Pytorch, Just Numpy)
Neural Networks From Scratch - Lec 21 - Regression Losses - MSE & RMSE
zhlédnutí 896Před 2 lety
Neural Networks From Scratch - Lec 21 - Regression Losses - MSE & RMSE
Neural Networks From Scratch - Lec 20 - Regression Losses - MAE, MAPE & MBE
zhlédnutí 1,1KPřed 2 lety
Neural Networks From Scratch - Lec 20 - Regression Losses - MAE, MAPE & MBE
Neural Networks From Scratch - Lec 19 - Approaching Regression Problem with Neural Networks
zhlédnutí 803Před 2 lety
Neural Networks From Scratch - Lec 19 - Approaching Regression Problem with Neural Networks
Neural Networks From Scratch - Lec 18 - Typical Neural Network Training Setup
zhlédnutí 687Před 2 lety
Neural Networks From Scratch - Lec 18 - Typical Neural Network Training Setup
Neural Networks From Scratch - Lec 17 - Python Implementations of all Activation functions
zhlédnutí 986Před 2 lety
Neural Networks From Scratch - Lec 17 - Python Implementations of all Activation functions
Neural Networks From Scratch - Lec 16 - Summary of all Activation functions in 10 mins
zhlédnutí 964Před 2 lety
Neural Networks From Scratch - Lec 16 - Summary of all Activation functions in 10 mins
Neural Networks From Scratch - Lec 15 - GeLU Activation Function
zhlédnutí 5KPřed 2 lety
Neural Networks From Scratch - Lec 15 - GeLU Activation Function
Neural Networks From Scratch - Lec 14 - Mish Activation Function
zhlédnutí 1,2KPřed 2 lety
Neural Networks From Scratch - Lec 14 - Mish Activation Function
Neural Networks From Scratch - Lec 13 - Swish Activation Function
zhlédnutí 1,2KPřed 2 lety
Neural Networks From Scratch - Lec 13 - Swish Activation Function
Neural Networks From Scratch - Lec 12 - Softplus Activation Function
zhlédnutí 870Před 2 lety
Neural Networks From Scratch - Lec 12 - Softplus Activation Function

Komentáře

  • @daminirijhwani5792

    woah, i am happy i found this!

  • @daminirijhwani5792

    This is amazing could you do a transformer series!

  • @emoji4652
    @emoji4652 Před 12 dny

    Thanks

  • @emoji4652
    @emoji4652 Před 12 dny

    Glad! i found this legendary explanation

  • @emoji4652
    @emoji4652 Před 12 dny

    Best

  • @emoji4652
    @emoji4652 Před 12 dny

    Bro ❤ that's so better than 200 usdt lectures

  • @YashVardhanSingh-nw6sd

    hiii will u post yolo v5 ? this is the best playlist

  • @NulliusInVerba8
    @NulliusInVerba8 Před 22 dny

    thank you! clear and direct.

  • @wiputtuvaynond761
    @wiputtuvaynond761 Před 24 dny

    Thank you very much for the best explaination of yolo papers on youtube. I have a question of loss calculation on multiscale training. This affects the number of output grid (WxHXS) used in loss calculation when input image(WxH) size changes. How does the loss calculation maintain consistency for this training scheme?

  • @pratikpatil2866
    @pratikpatil2866 Před 26 dny

    could you please mention source of the mathematical explanations it would be great help.

  • @Maximos80
    @Maximos80 Před 28 dny

    The best explanations of YOLO on CZcams. Period. Thank you!🙏

  • @tramyole2049
    @tramyole2049 Před měsícem

    really great job sir! waiting for more lately versions

  • @consumeentertainment9310
    @consumeentertainment9310 Před měsícem

    You rock!!! It was very detailed. Clearly, you have out a lot of work into this. Thank you so much🙏🙏🙏🙏🙏🙏

  • @adityakrishnajaiswal8663
    @adityakrishnajaiswal8663 Před měsícem

    Just wow!

  • @tanishksingh8443
    @tanishksingh8443 Před měsícem

    The best explanation for this concept, kindly keep making such content. Thanks a lot.

  • @giabao2602
    @giabao2602 Před měsícem

    thanks for giving us perfect video, love you bro

  • @giabao2602
    @giabao2602 Před měsícem

    thanks for helping us a lot in learning, truly appreciate your work

  • @sumanthpichika5295
    @sumanthpichika5295 Před měsícem

    very detailed explanation, Thanks for making it more clear. I believe i didn't find any such video with the way you explained the things in deep. I have a doubt when you said total loss = obj loss+no obj loss, In the example you considered only 2 grid cells has an object which means obj loss is calculated for those 2 grid cells and remaining 47 grid cells falls under no obj loss right?

  • @adityaa8918
    @adityaa8918 Před měsícem

    Underrated. Keep going man!

  • @syedmoiezullahbukhari2676

    thank you i could have nerver learn and understand any better then your video

  • @MJBZG
    @MJBZG Před měsícem

    i still want to know how to code this stuff up? theory can be found everywhere but how to dive into doing it ourselves?

  • @prodbyprodigy
    @prodbyprodigy Před měsícem

    this is such a great vid

  • @MJBZG
    @MJBZG Před měsícem

    great explanation! thanks!

  • @TihanyiPéter-b3b
    @TihanyiPéter-b3b Před 2 měsíci

    Very cool project! I have only one question: based on the video, you achieve 71.6% accuracy for 10 epochs, but I can't go above 52-54%. However, you also only have 56.94% accuracy in your github project. What is the reason of this? I'm very new to DL/MI, that's why I'm asking. I tried so much that I went up to 20 epochs, but the end result was almost exactly the same as that of 10 epochs. However, the accuracy increased by half per epoch... Thank you very much for any suggestions!

    • @ohm7163
      @ohm7163 Před měsícem

      u can try increase learning rate like (0.1,0.5).. it will gradually increase the accuracy , but also note that extreme Learning rate also lead to skiping ..and dont run entire program every time if you run entire program it will reset the all the values so run only part of training part alone jupiter nb and no of times running the training part = reducing the loss, and a=increases the accuracy if accuracy is there might be logic error

  • @madhav5166
    @madhav5166 Před 2 měsíci

    Just like sir said the higher level feature map or smalle feature map are mainly for bigger object and lower level feature map or bigger feature map are for smaller object,due to this various scale are neded for differenet feature map , can anyone tell me in simple words what is 'scale' here

  • @prashantbhardwaj4831
    @prashantbhardwaj4831 Před 2 měsíci

    One of the best I have come across so far, Could you please continue the series.

  • @deeplearningexplained
    @deeplearningexplained Před 2 měsíci

    Great work on the explanation, very well summarized!

  • @benjaminperez5334
    @benjaminperez5334 Před 2 měsíci

    nice video maboy, the best theorical explanation of why numpy is that good

  • @fotoluminescencjastudiesai1239

    great video, now I finally understand it :) could you just please clarify why in 22:32 only 2 grid cells contain objects? the woman appears in a few other cells as well, so why only two?

    • @MLForNerds
      @MLForNerds Před 2 měsíci

      Wherever the object centroid falls, only those cells are considered

    • @fotoluminescencjastudiesai1239
      @fotoluminescencjastudiesai1239 Před 2 měsíci

      @@MLForNerds thank you!! just to make sure that I understand correctly - in this example, one cell has a centroid for the horse and one has a centroid for the person? also, are you planning on making a video on Yolo v7? :)

    • @MLForNerds
      @MLForNerds Před 2 měsíci

      Yes, you are right regarding object centers. I will continue this series and finish all yolo versions

  • @kvzui994
    @kvzui994 Před 2 měsíci

    IM EATING THIS UP THANK YOU

  • @user-mc7tg4pf3i
    @user-mc7tg4pf3i Před 2 měsíci

    Hello Sir First upon thanks for your all videos, Sir i request you please upload Video on YOLOv8 architecture and How to improve accuracy in yolov8 model.

  • @AryanKumarBaghel-cp1jv
    @AryanKumarBaghel-cp1jv Před 2 měsíci

    Fantastic explaination. Super clear

  • @saisingireddy2359
    @saisingireddy2359 Před 2 měsíci

    underrated af

  • @mayankbhatt1308
    @mayankbhatt1308 Před 3 měsíci

    WOW what a clear and simple explanation

  • @nayabwaris-pl8lj
    @nayabwaris-pl8lj Před 3 měsíci

    please make video soon on remaining yolo variants

  • @MJBZG
    @MJBZG Před 3 měsíci

    thanks!

  • @AndreiChegurovRobotics
    @AndreiChegurovRobotics Před 3 měsíci

    what a great material - so well structured. thx for your efforts - beautiful!!!

  • @AndreiChegurovRobotics
    @AndreiChegurovRobotics Před 3 měsíci

    Dear Author, thx for your great work! you are awesome!

  • @AndreiChegurovRobotics
    @AndreiChegurovRobotics Před 3 měsíci

    Best YoloV3 video!

  • @KhozemaPython
    @KhozemaPython Před 3 měsíci

    Excellent Explanation . Will not forget concept now till going to grave

  • @user-oq7ju6vp7j
    @user-oq7ju6vp7j Před 3 měsíci

    Hi! thank you for your wonderfull explanation! Unfortunately in the original paper there are many unclear moments. Your video helped me a lot. But i still have some questions. 1) "Grid cell is "responsible" if the center of bbox falls into it." In training data we have annotated bboxes. But in test data there are no annotated bboxes and therefore centers. So which grid cell will be "responsible" in that case? 2) if c < threshold, then we simply nullify all the values in the vector or we should train the model to nullify the vector on its own? 3) if only 2 grid cells (in your case) predict the coordinates of bboxes, what is the use of the other 47 grid cells (are the useless at all or not?) 4) How one small grid cell (64x64) predicts a box for an object that is a way bigger than this cell (450x450)? 5) Why you are telling that there are only 2 object cells, if the woman overlap at least 6 cells? Maybe you mean only 2 "responsible" cells?

  • @Expl0r3r_m1nd
    @Expl0r3r_m1nd Před 3 měsíci

    And, I have a question for the "Not 0 centered function" section : What if you use the activation function only up to the last hidden layer. You just don't put any to get the final result.. The weights between the hidden layer and the output might be negative and the backpropagation won't be changing the sign of it.. will it ?

  • @Expl0r3r_m1nd
    @Expl0r3r_m1nd Před 3 měsíci

    Yes it's aaawesome !! I was searching a beginner video for weeks but they are all only "beginner friendly" by name.. It comes really fast into complicated maths subjects and you don't get to understand the compution process behing it.. You explain everything :) it's really cool ! TY

  • @shahriarahmadfahim6457
    @shahriarahmadfahim6457 Před 4 měsíci

    Best explanation for YOLOs man. Keep up the good work!

  • @AndreiChegurovRobotics
    @AndreiChegurovRobotics Před 4 měsíci

    great material!

  • @luansouzasilva31
    @luansouzasilva31 Před 4 měsíci

    If only one grid cell is labeled as class X, how does it get the bbox for the entire object?

    • @MLForNerds
      @MLForNerds Před 4 měsíci

      Grid call is only for box centre, the box dimensions will be learned as regression parameters

  • @sherinvarughese3683
    @sherinvarughese3683 Před 4 měsíci

    Very good teaching... concept are clearly taught.. keep going

  • @mayankbhatt1308
    @mayankbhatt1308 Před 4 měsíci

    Bro how much serious effort u have put to make this video.....and still so less views....thank u so much....will watch all ur videos.

  • @kvnptl4400
    @kvnptl4400 Před 4 měsíci

    🌟Came from the YOLO series; one of the best CZcams videos with an easy-to-understand explanation of YOLOv3. Keep up the good work. If possible, make video on Vision Transformer (ViT) and then DETR. 🙏

  • @kvnptl4400
    @kvnptl4400 Před 4 měsíci

    🌟A very in-depth analysis of the paper. I would say this is one of the best easy to understand explanations of YOLOv1. Keep up the good work