How AI Learns (Backpropagation 101)

Sdílet
Vložit
  • čas přidán 13. 11. 2019
  • Explore the fundamental process of backpropagation in artificial intelligence (AI). This video show how neural networks learn and improve by adapting to data during each training phase. Backpropagation is crucial in calculating errors and updating the network's weights to enhance decision-making within the AI system. This tutorial breaks down the core mechanics of neural network training, making it easier to understand for individuals interested in AI, machine learning, and network training. By understanding backpropagation, viewers can better grasp how neural networks evolve to process information more accurately. Keywords: rosenblatt, AI, Artificial Intelligence, Neural Networks, Backpropagation, Machine Learning, Network Training, Data Adaptation, Error Calculation, Performance Tuning, Decision Making.

Komentáře • 132

  • @markheaney
    @markheaney Před 4 lety +39

    This is easily the best channel on CZcams.

  • @TimBorny
    @TimBorny Před 4 lety +73

    As always, worth the wait. You are a genius at distillation and visualization.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 6 měsíci +2

      just finished this series, please help me share it: czcams.com/video/OFS90-FX6pg/video.html

  • @robertbohrer7501
    @robertbohrer7501 Před 4 lety +37

    This is the best explanation of neural networks I've seen by far, and I've seen most of them.

  • @austinvw1988
    @austinvw1988 Před měsícem +2

    WOW!! This is the only video that I've watched that made me finally get it. The inclusion of the physical dimmer switch and weights in the neural net made me finally start to grasp this concept. Thank You! 👏

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před měsícem +1

      So so happy to hear this, glad I took the time. I need to get this video out there more

  • @TimBorny
    @TimBorny Před 4 lety +13

    Seriously impressive. As someone currently applying to masters degrees in science communication, you are an inspiration. While a personal inquiry within a public forum is generally not advisable, I'm compelled to wonder if you'd be willing to be available for a brief conversation.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +4

      I appreciate hearing this. You can reach me britjcruise@gmail.com

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 6 měsíci +1

      just finished this series, please help me share it: czcams.com/video/OFS90-FX6pg/video.html

  • @baechlio
    @baechlio Před 4 lety +4

    Yay!!! My favourite channel finally uploads again.. to be honest the quality of your videos makes the wait worth it

  • @yomanos
    @yomanos Před 4 lety +3

    Brilliant video, as always. The part on the explanation of a deep neural network was really well explained.

  • @iMamoMC
    @iMamoMC Před 4 lety +10

    This video was great! What an awesome introduction to deep learning ^^

  • @idiosinkrazijske.rutine
    @idiosinkrazijske.rutine Před 4 lety +8

    The highlight of this day

  • @KhaliliStudios
    @KhaliliStudios Před 4 lety +4

    I’m always very impressed at the novel approach to teaching these subjects - another hit, Brit!

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +1

      that you for your ongoing feedback. I worked super hard on this one,

  • @kriztoperurmeneta7089
    @kriztoperurmeneta7089 Před 4 lety +1

    This kind of content is a treasure.

  • @ccc3
    @ccc3 Před 4 lety +1

    Your videos are great at making someone more curious about a subject. They have the right balance of simplification and complexity.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +2

      appreciate the feedback that's what i'm looking to do with these videos. Stay tuned to the next in this series it took me a long while to write

  • @rj8875
    @rj8875 Před 4 lety +5

    After reading all day about tensorflow you just inspired me to go deeper on this subject. Thank you

  • @TheFirstObserver
    @TheFirstObserver Před 2 lety +1

    This is a well-done, visual representation of artificial neural networks and how they compare to biological ones. I will say, the only item I might add is that the real reason the "gradual" activation functions mentioned in the latter half of the video are so useful is because they are differentiable. The functions being differentiable are what truly allowed backpropagation to shine, as the chain rule is what allowed the error of a neuron to be determined by the error of the neurons following it, rather than calculating a neuron's error from the output directly each time.

  • @NoNTr1v1aL
    @NoNTr1v1aL Před rokem +1

    Absolutely brilliant video!

  • @raresmircea
    @raresmircea Před 4 lety +4

    This, along with everything else on this channel, is fantastic material for schools. I hope it gets noticed by teachers

    • @KipColeman
      @KipColeman Před 6 měsíci +1

      College IT professor here... we are noticing! :)

  • @zyugyzarc
    @zyugyzarc Před rokem +1

    now that's a brilliant explanation of neural networks. better than anything Ive ever seen.

  • @mehdia5176
    @mehdia5176 Před 4 lety +2

    Beautiful work coming from a beautiful biological neural network about the beauty of artificial neural networks.

  • @chris_1337
    @chris_1337 Před 4 lety +2

    Fantastic work!

  • @elektrisksitron9054
    @elektrisksitron9054 Před 4 lety +2

    Another amazing video!

  • @ssk081
    @ssk081 Před 3 lety

    Great explanation of why we use a smooth activation function

  • @poweruser64
    @poweruser64 Před 4 lety +1

    Wow.
    Thank you so much for this

  • @jayaganthan1
    @jayaganthan1 Před 2 lety

    Just wow. Awesome explanation.

  • @CYON4D
    @CYON4D Před 4 lety +1

    Excellent video as always.

  • @interspect_
    @interspect_ Před 4 lety +1

    Great video as always!!

  • @Aksahnsh
    @Aksahnsh Před 4 lety +2

    I just don't understand, why this channel is not popular.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +2

      I know, i kinda stopped asking myself. I know it's due to algorithm changes in some way. because my videos don't even go to subscribers much at all

    • @Aksahnsh
      @Aksahnsh Před 4 lety

      @@ArtOfTheProblem True, even I didn't got it in my recommendation feed. I just realized that why there is no new video from you from long time, have to open your channel to manually find it. Clicked on bell icon though now.

    • @ByteNishi
      @ByteNishi Před 4 lety +1

      @@ArtOfTheProblem Please, don't get disheartened. I really love your videos and eagerly wait for new ones :)

    • @roygalaasen
      @roygalaasen Před 2 lety

      There are no truer words like these. It baffles me as these videos are at least on the level of other highly popular science/math youtubers. It feels kind of unfair. Even the videos made 8+ years ago are pieces of masterful art. Did any of the other youtubers even exist back then? (I guess some did.)

    • @roygalaasen
      @roygalaasen Před 2 lety

      @@ByteNishi I am praying for the same. I am happy for the once a year schedule. At least there is something.
      Edit: I know it is a bit of exaggeration. There is at least 4 videos per year, which seems close to what 3b1b does nowadays as well.

  • @ByteNishi
    @ByteNishi Před 4 lety

    Love your videos, can you please post videos more often. Thanks, your videos are always worth the wait.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +5

      thank so much. I can't possibly post more often but what I can do is promise to continue for another 10 years :)

  • @midhunrajr372
    @midhunrajr372 Před 4 lety +1

    what a nice presentation..

  • @DaveMakes
    @DaveMakes Před 4 lety +2

    great work

  • @karolakkolo123
    @karolakkolo123 Před 4 lety +4

    Wow! The most amazing explanation on the internet probably. Will actual calculations be talked about in the series? (e.g. backpropagation calculus, etc) or will the series be mostly conceptual? (Any way I'm sure it will be interesting and of an unmatched quality)

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety

      great question and thank you. No more details on backprop calculations (there are lots of good videos of that) in order to focus on other key insights. stay tuned!

  • @JoshKings-tr2vc
    @JoshKings-tr2vc Před 17 dny +1

    This is a very well written video and explains it quite well. I have a question for anyone willing to answer; what would occur if we took a simple functioning neutral network and added another layer to it? Would it get better in confidence or in conceptualization or would it simply not have any noticeable effect?
    Along the same lines, if it even did a minor improvement (for better generalizations) would it be a more efficient way of training a deep neural net? Sort of like calculating the amount of ohms each resistor takes up in a circuit by breaking it down to simpler bite sized problems.
    Just things that tickle my fancy.

    • @JoshKings-tr2vc
      @JoshKings-tr2vc Před 17 dny

      That second question was confusing. All I’m saying is, if we have this huge neural net, why not break it down to smaller parts of it and optimize for confidence because adding more layers would supposedly make it better at generalization?

  • @srabansinha3430
    @srabansinha3430 Před 4 lety +1

    As a Medical Student studying Neural anatomy and Physiology , this is a whole new perspective to me !!! Keep teaching us More !!You are the Best teacher :)

  • @mridhulml9238
    @mridhulml9238 Před 2 lety

    Wow this is really really great...you are really good at explaining

  • @acidtears
    @acidtears Před 4 lety +3

    Great video! Do you have any idea how these types of Neural Networks would respond to visual illusions? I'm writing my thesis about Neural Networks and biological plausibility and realized that there seems to be a disconnect between human perception and the processing of Neural Networks. Either way, incredibly informative.

  • @lalaliri
    @lalaliri Před 3 lety

    amazing work! thank you

  • @ilovett
    @ilovett Před 4 lety +2

    This could be a Netflix series. Bravo.

  • @fungi42021
    @fungi42021 Před 3 lety

    always looking for new content to watch on this topic.. great channel

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 3 lety

      i'm so happy you found this series as it isn't ranking well yet. i have more videos coming out in this series soon

  • @harryb.234
    @harryb.234 Před měsícem

    Really cool. Haven't seen a better explanation

  • @yagomg7790
    @yagomg7790 Před 4 lety +1

    Best explanation on youtube. Keep it up

  • @shawnbibby
    @shawnbibby Před 5 měsíci +1

    The term Distributed Representation when compared to musical notes makes it seem like it has its own image Resonance or Signature Frequency. As if we really are seeing or feeling the totality of the image of the firing neurons.
    We seem to be addicted to understanding perceptions from a Human Point of View, imagine if we could begin to find translations to see them from Animal Point of Views, Different Sensory Combination Point of Views and different combinations of Layered Senses. The potential is infinite.
    I like the addition of the Linear Learning Machine versus one that forgets and uses Feelings. It seems that by combining both memory styles you would have more unique potentialities in the flavor pot of experiences, especially when the two interact with each other. Not to mention the infinite different perspectives they would each carry while traveling through time. Small and large epochs of time.
    I seem to keep coming back to the Encryption / Decryption videos on how it requires complete Randomness to create strong encryption and how the babies babbling was seemingly random in nature, which begs the question, was it truly random or could we simply just not see the pattern from our limited perspective?
    What is the scales and size of the pattern? And what conceptions and perspectives need to merge to simply find the Key to interpreting it?

  • @shawnbibby
    @shawnbibby Před 5 měsíci

    I would also love to see a video of all the terminologies used together and defined in a single video. Such as Bit, Node, Neuron, Layer, Weight, Deep learning, Entropy, Capacity, Memory etc. I am trying to write them down myself as a little glossary. There meanings are so much greater when they are grouped together.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 5 měsíci +1

      thank you! i was thinking of making a super edit of this series just need to scope it correctly...

  • @sumitlahiri209
    @sumitlahiri209 Před 4 lety +2

    Amazing Video. It was really worth the wait. I have watched all your videos. Just awesome I would say. Best channel for inspiring computer science enthusiasts

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +1

      that's really cool to hear you've watched them all. thanks for sharing

    • @sumitlahiri209
      @sumitlahiri209 Před 4 lety +1

      @@ArtOfTheProblem I watched all of them. They inspired me to take up computer science. I really love the video on Turing Machine. I share your videos in cirlces as well.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +1

      @@sumitlahiri209 you can offer my no higher compliment

  • @username4441
    @username4441 Před 4 lety +4

    11:49
    And the narration model took how long to-train?

  • @KDOERAK
    @KDOERAK Před 2 měsíci

    Simply excellent 👍

  • @solsticeprojekt1937
    @solsticeprojekt1937 Před 9 měsíci

    Hi! Three years late, but at 0:26 where you say "feelings", you describe something much better explained as "realizations". The answer to the "Why?" about this lies behind the saying "an image speaks a thousand words". The part that takes care of logic works in steps, sequentially and can analyse the "whole" of a realization, just like we can put feelings and ideas into words. This works both ways, of course, but the path from words to realizations is a much, much slower one.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 6 měsíci

      Took 2 years to finish this one, finally live would love your feedback: czcams.com/video/OFS90-FX6pg/video.html

  • @harryharpratap
    @harryharpratap Před 4 lety +2

    Biologically speaking, what are the "weights" inside our brains? What physical part of the brain do they represent?

  • @thisaccountisdead9060
    @thisaccountisdead9060 Před 4 lety +1

    I'm not an expert or anything. But I had just been looking at networks. I was interested in the erdos formula: -
    erdos number = ln(population size) / ln(average number of friends per person) = degrees of separation
    for example it is thought there is something like 6 degrees of separation and an average of 30 friends each person among the global population.
    But I was also looking at Pareto distributions as well: -
    1 - 1/Pareto index = ln(1 - P^n) / ln[1 - (1 - P)^n], where P relates to population of wealtheist and (1 - P) is the proportion of wealth they have.. for example if 20% of people have 80% of the wealth then P = 0.2 and (1 - P) = 0.8. n = 1 (but can be any number... if n = 3 it gives 1% of people with 50% wealth) and the Pareto Index would be 1.161.
    Whether it was a fluke I don't know? I did derive the formula as best I could rather than just guessing. But it seemed as though the following was true: -
    1 - 1/Pareto Index = 1/Erdos Number
    Meaning that the Pareto Index = ln(population size) / [ln(populationn size) - ln(average number of friends per person)]
    Suggesting that the more friends people had on average then the lower the wealth inequality would be. Which I thought was a fascinating idea...
    ...But it also seemed as though the wealtheist actually had the most 'friends' or 'connections'. So the poorest would have the least connections while the wealthiest would have the most connections - in effect poor people would be channeling their attention toward the wealtheist. Like the top 1% would have an average of around 2,000 connections each (*and a few million dollars) while the poorest would have as little as 1 or 2 connections each (*with just a few thousand dollars... *based on a share of $300 Trillion). Maybe in like a neural network the most dominant parts of the brain could be the most connected parts?
    As I say I am not an expert. I was just messing around with it.

  • @trainer1kali
    @trainer1kali Před 4 lety

    a message to the one's responsible for the choices of the background music to translate the mood: "you're pretty good".
    P.S. In fact - you are AWESOME.

  • @emanuelmma2
    @emanuelmma2 Před 3 měsíci +1

    Amazing Video.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 2 měsíci

      would love if you could help share my newest video: czcams.com/video/5EcQ1IcEMFQ/video.html

  • @zuhail1519
    @zuhail1519 Před rokem

    I want to mention here, I watched the video halfway and I must say, I am a complete noob when it comes to biology but without making things complicated for a person like me, You made it so incredibly clear to me to appreciate how amazing our brain works and generalizes stuff, especially with your example of the short-story (can you please mention that author name, I quite couldn't catch it and cc are not clear either). Thank you for making this content, I'm grateful. Jazakallah hu khayr

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před rokem +1

      thrilled to have you, i'm still working on the final video in this series so please stay tuned. Was it "Borges'?

    • @zuhail1519
      @zuhail1519 Před rokem

      @@ArtOfTheProblem Already have my seatbelt fastened !

  • @ahmadsalmankhan3200
    @ahmadsalmankhan3200 Před 6 měsíci +1

    Amazing

  • @iamacoder8331
    @iamacoder8331 Před rokem

    Very good content.

  • @slazy9219
    @slazy9219 Před rokem

    holy shit this is some next level explanation
    thank you so much!

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před rokem +1

      super glad you found it, still working on this series

  • @user-eh9jo9ep5r
    @user-eh9jo9ep5r Před 4 měsíci

    What input could be done to heal neurons to basic correct stages, for give correct outputs

  • @Arifi070
    @Arifi070 Před 4 lety +2

    Great work! However, although the artificial neural network was inspired from the working of our brains, visualizing the network inside a head, can give a wrong idea that the human brain works that way. In fact, it is not like a feed forward neural network.
    [Just a side note]

  • @KalimbaRlz
    @KalimbaRlz Před 3 lety

    excellent explained

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 3 lety

      thanks for feedback, have you watched the whole series?

    • @KalimbaRlz
      @KalimbaRlz Před 3 lety

      @@ArtOfTheProblem yes I did!, thank you for all the information

  • @user-eh9jo9ep5r
    @user-eh9jo9ep5r Před 4 měsíci

    What if one layer behaviour is different from expected, and not recognised as correct, but other layers are give output from sorts geomerical inputs on the level sense impulse, what can be done to filter inputs and recieve correct outputs

  • @ArtOfTheProblem
    @ArtOfTheProblem  Před 4 lety +5

    STAY TUNED: Next video will be on "History of RL | How AI Learned to Feel"
    SUBSCRIBE: www.youtube.com/@ArtOfTheProblem?sub_confirmation=1
    WATCH AI series: czcams.com/play/PLbg3ZX2pWlgKV8K6bFJr5dhM7oOClExUJ.html

  • @user-eh9jo9ep5r
    @user-eh9jo9ep5r Před 4 měsíci

    If network recieve input, give output, and answer isnt clear and recognised as not correct. Could it be recognised as network desieas , and if it so could be recognised as consequences influenced from other network or networks outputs

  • @robosergTV
    @robosergTV Před 3 lety

    Isn't universal approximation theorem the mathematical proof NN can solve and model any problem/function?

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 3 lety

      right but that is only in theory, in practice the number of neurons is impractical to make it "practically impossible" to impliment.

    • @robosergTV
      @robosergTV Před 3 lety

      @@ArtOfTheProblem true, but at the end of the video, you were talking something about "we still don't have a mathematical proof of how NN works" or something like that.

  • @lakeguy65616
    @lakeguy65616 Před 2 lety +1

    so adding hidden layers allows a NN to solve more complex problems. How many layers is too many? You are limited by the speed of the computer training the NN. I assume too many layers allow the NN to "memorize" instead of generalizing. Any other limits on the number of hidden layers?
    What about the number of neurons/nodes per layer? Is there a relationship between the number of inputs and the number of neurons/nodes in the network?
    What about the relationship between the number of rows in your dataset versus the number of columns? As I understand it, the number of rows imposes a limit on the number of columns. Adding rows to your dataset allows you to expand the number of columns too. Do you agree or have a different understanding?
    OUTSTANDING VIDEOS!
    John D Deatherage

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 2 lety +2

      super great questions. I hope others can chime in. just wanted to add that in 'theory' you only need one hidden layer if it was really really wide to solve any problem (see universal approximation theorem), but in practice that doesn't work. and yes if the network is "too deep" it will be too difficult to train, so you need a sweet spot. when it comes to how wide those layers need to be, the most interesting research to me is how 'narrow' you can make them to 'force' the network to abstract (compress/generalize) the information in the middle. you can also make the hidden layers very wide which will cause it to 'memorize' instead of generalize. i didn't quite follow your column / row question though

  • @abiakhil69
    @abiakhil69 Před 4 lety

    Consensus mechanism?

  • @columbus8myhw
    @columbus8myhw Před 4 lety

    Link to the Hinton lecture?

  • @Trombonauta
    @Trombonauta Před 4 měsíci

    1:13 Cajal is pronounced more closely to /kah'al/ than to that, FYI.

  • @KittyBoom360
    @KittyBoom360 Před 4 lety

    This might be more of a tangent to your great video, but my understanding is that intuition and logic aren't really distinct things. The former is merely more hidden in deep webs of logic while the latter is the surface or what is most obvious and intuitive. Ah, see the paradox? It's a false dichotomy resulting from awkward folk terms and their common definitions.
    I was always like the teacher's pet in college courses of logic and symbolic reasoning while majoring in philosophy maybe partly because anything that was labeled "counter-intuitive" was just something I would never accept until I could make it intuitive for me via study and understanding. But putting me and my possible ego aside, look at the example of a great mathematician such as Ramanujan and how he described his own process of doing math while in dream-like states. His gift of logic was indeed his gift in intuition, or vice versa, depending on your definitions.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety +1

      Yes I had a section in this video i cut which I kinda wish I left it. it was about how intuition is the foundation out of which logic grows. Kids won't learn with "words first" they learn with "sense first" - so mathematicians are of course guided by intuition and then they can later prove things with logic.

  • @bicates
    @bicates Před rokem

    Eureka!

  • @arty4679
    @arty4679 Před 4 lety

    Anyone knows the name of the Borjes story?

  • @abiakhil69
    @abiakhil69 Před 4 lety +1

    Sir any blockchain related videos in future?

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety

      have you seen my bitcoin video?

    • @abiakhil69
      @abiakhil69 Před 4 lety +1

      @@ArtOfTheProblem Yes sir. One of the best video YT.

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 4 lety

      @@abiakhil69 I do plan a follow up video, starting with ETH

    • @abiakhil69
      @abiakhil69 Před 4 lety

      @@ArtOfTheProblem great sir. Another best video coming . Waiting👍.

  • @fredsmith4134
    @fredsmith4134 Před 6 měsíci +1

    it's all a chain of cause and effect from start to finish, each level or layer sharpens and zero's in on the exact match and is refined until a result is locked in, the human brain compares past results to incoming stimuli, and the result is also linked by chains of associations with the result, like result it's a dog, associations : dogs are fury, playful, dangerous, have a master, wage there tail when happy, and so on, but associations are unique to each separate mind ????

  • @user-eh9jo9ep5r
    @user-eh9jo9ep5r Před 4 měsíci

    If sensory order was destroyed or noised or anything like this, something like network trafficking , what need to do for to safe all neural network

  • @Libertariun
    @Libertariun Před 6 měsíci

    14:45 ... can learn to configure THEIR connections ...

  • @kmachine5110
    @kmachine5110 Před 4 lety +1

    CZcams is a mind reader.

  • @AceHardy
    @AceHardy Před 4 lety +1

    👑

  • @fxtech-art8242
    @fxtech-art8242 Před rokem

    gpt4

  • @escapefelicity2913
    @escapefelicity2913 Před 4 lety

    Get rid of the background noise

    • @escapefelicity2913
      @escapefelicity2913 Před 4 lety +1

      For anything expository, any background sound is unhelpful.

  • @EvenStar303
    @EvenStar303 Před 6 měsíci +1

    After 35 seconds, you are already wrong.
    We do not think in sentences!!!
    Thinking is wordless.
    However, we are translating our thinking into words.
    But this is not necessary.
    The point being, is that languageing is only necessary if we want to communicate to another person.
    But thinking comes first, NOT as a result of sentences.
    If you get good at meditation and centering yourself, you can drive your car without verbalizing what you are doing.
    You can make decisions and act them out without verbalization, internally or externally!
    Language is only a descriptor, not the thinking faculty itself!!!

    • @ArtOfTheProblem
      @ArtOfTheProblem  Před 6 měsíci +1

      check out the whole series as I built up to this, i agree with you!

  • @vj.joseph
    @vj.joseph Před 6 měsíci

    you are wrong within the first 40 seconds.

  • @joe_cock
    @joe_cock Před 5 měsíci

    good sound design