Video není dostupné.
Omlouváme se.

Eigenvalues & Eigenvectors : Data Science Basics

Sdílet
Vložit
  • čas přidán 15. 08. 2024
  • So what are eigenvectors and why are they important?
    ---
    Like, Subscribe, and Hit that Bell to get all the latest videos from ritvikmath ~
    ---
    Check out my Medium:
    / ritvikmathematics

Komentáře • 135

  • @maditea
    @maditea Před 3 lety +107

    *me learning linear algebra for the first time even though i passed the class three or four years ago*

    • @EdeYOlorDSZs
      @EdeYOlorDSZs Před 3 lety

      Oof

    • @jonathanokorie9857
      @jonathanokorie9857 Před 2 lety +1

      lol

    • @spiderjerusalem4009
      @spiderjerusalem4009 Před rokem +1

      hate it. Seems more theory rather than intuition. Let alone super rigorous dumbed down overrated book such as axler's. All students would end up doing would be memorizing them for the sake of getting through all the complexity since the intuition is nowhere to be found

  • @batuhantekmen6607
    @batuhantekmen6607 Před 4 lety +72

    Last 2-3 mins are invaluable, I knew eigen values and eigenvectors but didnt know where to use them. Thanks very much!

    • @wirelessboogie
      @wirelessboogie Před 5 měsíci

      Same here, cheers to the author of the video!

  • @romainthomas8238
    @romainthomas8238 Před 2 lety +13

    as a native french speaker, I understand your videos better than most of any french courses I've read / watched ! Thanks a lot, you save me a lot of time and desperation :D

    • @ritvikmath
      @ritvikmath  Před 2 lety +7

      Je suis heureux d'aider! (Sorry I only took 3 years of french in high school 😁)

  • @tobydunbar1153
    @tobydunbar1153 Před 3 lety +21

    You have an easy, inclusive and coherent way of teaching. Great job!!👌

  • @visheshmp
    @visheshmp Před rokem +2

    You are amazing ... anyone who is watching this video please don't miss last few minutes.

  • @TheDroidMate
    @TheDroidMate Před 3 lety +11

    Just came to your vids by accident.. now I'm asking how I could not have got here earlier..!

  • @saraaltamirano
    @saraaltamirano Před 4 lety +4

    I have searched and searched for an explanation like this one, took me months to finally found an explanation that anyone can understand. You are a talented teacher, thank you!

    • @marcowen1506
      @marcowen1506 Před 3 lety

      there's a really old maths book by Bostock and Chandler that has a good explanation of this too.

  • @nexus1226
    @nexus1226 Před 4 lety +29

    Just came across your channel .. Your videos are absolutely amazing! I'm in a multivariate analysis course, where I need to refresh my linear algebra skills, so these videos are really helpful.

  • @anujasebastian8034
    @anujasebastian8034 Před 3 lety +1

    I've been learning eigen values and vectors solved a bunch of problems without even understanding what i was doing....Thanks a lot for that explanation!!!

  • @sepidehmalektaji3770
    @sepidehmalektaji3770 Před 2 lety +11

    And the best explanation eigenvector/value prize goes to.... this guy!... good job man ...great video

  • @siddhantrai7529
    @siddhantrai7529 Před 3 lety +3

    last 30 secs taught me more than last 3 months. Thanks you sir. Your way of teaching is impeccable. I am absolutely stunned by the last minute intution.MIND = BLOWN

  • @paulntalo1425
    @paulntalo1425 Před 3 lety +1

    You have not just made my day but my career. I have following you for two days and seems u just keep cracking the rocket science. Am doing an Msc Data Science. Thank u so much

  • @sirivilari6796
    @sirivilari6796 Před 4 lety +6

    Really did a good job man. Appreciate your time and valuable information

  • @sirginirgin4808
    @sirginirgin4808 Před 3 lety +2

    Many thanks. You summarised important chapter of linear Algebra in just less than 12 minutes.

  • @ryansolomon2778
    @ryansolomon2778 Před 3 lety +3

    How this man has not blown up bigger that someone like blackpenredpen is beyond me. I am in Calculus II right now, and this video made perfect sense to me.

  • @saulflores5052
    @saulflores5052 Před 3 lety +3

    Thank you! I took linear over a year ago and your explanations clear up so many questions I had.

  • @donalddavis8033
    @donalddavis8033 Před 2 lety +2

    This is way better than the explanation I had in my linear algebra course long ago!

  • @CaterpillarOGM
    @CaterpillarOGM Před rokem +1

    It would be useful to be able to like these videos more than once to express how appreciated they are for a newbie! Thank you!

  • @chathuraedirisuriya6535
    @chathuraedirisuriya6535 Před 4 lety +2

    Excellent explanation. Very useful in my mathematical modelling of infectious disease learning. Thank you

  • @komelmerchant6772
    @komelmerchant6772 Před 3 lety +2

    These are awesome videos! They really Intuitively connect theoretical concepts in linear algebra with application in ways that I was never explicitly taught! Keep up the great work!

  • @theodoresweger4948
    @theodoresweger4948 Před rokem +1

    You not only explained the math operations, but thanks for the insite of why we are doingit. Thanks for the enlightment.

  • @123gregery
    @123gregery Před 2 lety +1

    You are good. I knew some linear algebra but I couldn't get the "feel" of it. Watching this Data Science Basics series has changed it.

  • @marcogelsomini7655
    @marcogelsomini7655 Před 2 měsíci

    11:35 useful to think about the same "ratio", thank you boss

  • @arsemabes
    @arsemabes Před 3 lety +2

    this just made me so happy... THANK U!

  • @danieljulian4676
    @danieljulian4676 Před 8 měsíci

    Your presentation skills are top-notch. Since this is the first of your videos I've watched, I don't yet know whether you devote another video to other properties of eigenvectors. You stress the collinearity, but don't talk about the way the hypervolume of some set of vectors collapses. Maybe you do this in a video where you define the determinant. Maybe your mentioning the null space of the matrix covers this. At any rate, I'll say at this point that I'll probably find all your presentations worthwhile. Best wishes in growing your channel.

  • @user-or7ji5hv8y
    @user-or7ji5hv8y Před 3 lety +2

    Wow, I don’t know why professors rarely provide motivation like this.

  • @scottlivezey9479
    @scottlivezey9479 Před 3 lety +1

    I haven’t had a reason to dive into this kind of topic for over 20 years: only saw it during undergrad & grad school. But I enjoyed your technique of going through it.

  • @yingchen8028
    @yingchen8028 Před 3 lety +1

    I love your videos .. super helpful not just to refresh the knowledge but also understand it in a more intuitive way!! Thank you so much !

  • @minxxdia1132
    @minxxdia1132 Před 4 lety +1

    this video made it all look so so simple, thankyou very much!

  • @lilmoesk899
    @lilmoesk899 Před 4 lety +3

    great explanation. thanks.

  • @gello95
    @gello95 Před 4 lety +1

    I appreciate you making these videos as they have helped so much in understanding abstract machine learning/data science concepts! :) Cheers to you!

    • @sahil0094
      @sahil0094 Před 3 lety

      Cheers to me . I taught him

  • @Ivan-mp6ff
    @Ivan-mp6ff Před 3 měsíci +1

    What an amazing example in eigenvector. Help fish to find fish with same "figures". Surely has been used in dating apps😂

  • @blueis910
    @blueis910 Před 3 lety +1

    You're helping me a lot refreshing these concepts! So happy I found your channel!

  • @edphi
    @edphi Před rokem

    Best video on eigen

  • @user-xj4gg9jm3q
    @user-xj4gg9jm3q Před 2 lety

    this cleared up so much and importance of why we need eigenvectors, tysm!!

  • @mustafizurrahman5699
    @mustafizurrahman5699 Před rokem

    Awesome splendid mesmerising

  • @sali6989
    @sali6989 Před 2 lety

    thank you I love this topic it gives me a lot in the foundation of basic data science most likely in machine learning

  • @JasonBjörne89
    @JasonBjörne89 Před 3 lety

    Your videos are absolutely top-notch. Keep it up!

  • @agamchug595
    @agamchug595 Před 2 lety

    This is amazing! Thank you so much for making these videos.

  • @rajgopalmanoharan
    @rajgopalmanoharan Před 10 měsíci

    Thank you, really helpful, awesome explanation :)

  • @user-ku1qq4jo4p
    @user-ku1qq4jo4p Před rokem

    thank you, I learned faster and easier with your explanation, you rock!

  • @sarfrazjaved330
    @sarfrazjaved330 Před 2 lety

    This man is a real gem.

  • @jairomejia616
    @jairomejia616 Před rokem +1

    Take the last section of the video, knowing that eigen is a German word for "own" and you will never forget what is the importance of eigenvalues and eigenvectors.

  • @bocckoka
    @bocckoka Před 3 lety

    I think it's important to point out what an operator can do to a vector (A*x) in general, and then point out that these eigen directions are special, because here the operator's effect is just scaling. And this is useful, because...

  • @MichaelGoldenberg
    @MichaelGoldenberg Před 3 lety

    Very clean and clear presentation.

  • @nikolaosnikolaou1556
    @nikolaosnikolaou1556 Před rokem

    really nice explanation thank you!!

  • @chandrikasaha6301
    @chandrikasaha6301 Před 4 měsíci

    Which video to follow for the importance of invertibility?

  • @amjedbelgacem8218
    @amjedbelgacem8218 Před rokem

    You are a literally a Godsend and a savior, Machine Learning is becoming more clear with every video i watch of you fam, thank you!

  • @ernestanonde3218
    @ernestanonde3218 Před 2 lety

    just found you & I LOVE YOU

  • @sanketannadate4407
    @sanketannadate4407 Před 2 měsíci

    God Tier Explanation

  • @alinazem6662
    @alinazem6662 Před 10 měsíci

    Killed it. Period.

  • @danielwiczew
    @danielwiczew Před 4 lety +1

    Great video!

  • @MohamedMostafa-kg6gk
    @MohamedMostafa-kg6gk Před 3 lety +1

    Thank you for this great explanation .

  • @EdeYOlorDSZs
    @EdeYOlorDSZs Před 3 lety +1

    funny how the mathematical understanding behind it is very important to grasp, however we will never have to calculate the eigenvectors and values by hand after university.

  • @AG-dt7we
    @AG-dt7we Před 3 dny

    How does this property of a vector (eigen vector) remains in the same dimension even after transformation (by A) helps in some problem solving (related to ML)?

  • @mcwulf25
    @mcwulf25 Před 3 lety +2

    Good explanation of the math. But for 40 years I still struggle with what eigenvalues really are. Your fish example was better than most I have heard but I am still missing something vital.

    • @ritvikmath
      @ritvikmath  Před 3 lety +1

      It's definitely a tricky concept and I'm glad this video helped a little bit. Took me a long time to understand too. I think the easiest explanation is that an eigenvector is one where the matrix will map a vector to a multiple of itself (so that the input vector and the output vector both point in the same direction). Why does this matter? Because the same direction ensures the same ratios between each individual vector component which loosely means that the input and output vectors have the same proportions.

    • @mcwulf25
      @mcwulf25 Před 3 lety

      @@ritvikmath That helps. I have come across it in PCA and in quantum mechanics. (Yes, I am eclectic).
      Another question: do eigenvalues HAVE to be real?

  • @abhinavmishra9401
    @abhinavmishra9401 Před 3 lety +1

    This is amazing. You are amazing.

  • @user-ny5jw1sq9b
    @user-ny5jw1sq9b Před 11 měsíci

    WOW!! bless you man🌺

  • @theodoresweger4948
    @theodoresweger4948 Před 4 lety

    Thank you very well explained and I like the fish analogy. .

  • @luiswilbert2377
    @luiswilbert2377 Před rokem

    great job

  • @AnDr3s0
    @AnDr3s0 Před 3 lety +1

    Thanks man! Really good explanation! Keep it up!

  • @karthikeya9803
    @karthikeya9803 Před 4 lety

    Explanation is awesome

  • @unzamathematicstutormwanaumo

    You are good 👍

  • @PaddyMcCarthy2.1
    @PaddyMcCarthy2.1 Před rokem

    All teachers seem to fail at the stage where they include the identity matrix [02:07] why is that? The reason is, because they understand that putting the identity matrix in does not affect vector or lambda. But they never tell you this vital bit of information. And they still wonder why people fail to understand mathematics. They had to learn it themselves. but they are not including this vital piece of knowledge in their explanations. It now seems a trivial point to them, but for a student starting out, it is not trivial. In fact, for any student with a basic understanding of Algebra, they would wonder why I only one side of the equation Ax = λx being multiplied by I, the identity matrix, surely this breaks the rule of algebra which says whatever you multiply one side of the equation by, you must multiply the other side of the equation by. And yet here, we see only one side of the equation being multiplied by I, the identity matrix. Without any explanation as to why you can do that. It's time you start explaining why it is alright to multiply λx by I, the identity matrix on one side and not the other side of the equation: answer, because as any identity does, it does not change the number. hence the word identity. [02:23] let's subtract, Ax - λx = λIx, ready? Ax - λx -[λIx]= λIx - [λIx], okay how does that equal Ax - λIx = 0? well, λIx - [ λIx] equals zero, so the right hand side of the equation is fine. but what about the left hand side? What we have is Ax - λx -[λIx] . Okay so let's apply a little algebra: like terms can be added or subtracted. No like terms so, nothing can be subtracted here. Amazingly, ritvikmath seems to think these can be subtracted. Actually, in his calculations the term λx just magically disappers, so he is left with Ax - λIx = 0. He could have got to this result a different way.
    Let's start out with Ax = λx, then subtract λx from both sides (as laws of algebra suggest) that would give: Ax - λx = λx - λx. which results in, Ax - λx = 0. Now he has a choice to include the identity matrix Ax - λIx = 0. see, same result. Nothing magical, nothing disappears, every step accounted for. His main argument is right. And I look forward to his video on determinants of matrices, i.e. proving that a matrix is non-invertable.

  • @brownsugar85
    @brownsugar85 Před 4 lety

    These videos are amazing!

  • @andreo1030
    @andreo1030 Před 3 lety

    Thanks for your clear explanation

  • @prakashb1278
    @prakashb1278 Před 3 lety

    This helps a lot!! Thank you so much

  • @VictorOrdu
    @VictorOrdu Před rokem

    Lovely! I have a question: if the scalar (eigenvalue) is negative, when multiplied is the vector's direction not changed by 180 degrees?

  • @marcosfuenmayor563
    @marcosfuenmayor563 Před 3 lety

    amazing !!

  • @himanihasani
    @himanihasani Před 4 lety

    great explanation!!

  • @kewtomrao
    @kewtomrao Před 2 lety

    How did u assume the shape of X was (2,1)?
    Was it because of two eigen values?

  • @umehmoses8118
    @umehmoses8118 Před rokem

    Thank you

  • @paulbrown5839
    @paulbrown5839 Před 3 lety

    Good video! Thanks.

  • @SoreneSorene
    @SoreneSorene Před 2 lety

    Hey, special thanks for that last application example 😊

  • @rezaerabbi2492
    @rezaerabbi2492 Před 3 lety

    Could you upload a video on Hat Matrix?

  • @Fat_Cat_Fly
    @Fat_Cat_Fly Před 3 lety +1

    soooo good!

  • @ranjbar
    @ranjbar Před rokem

    my man gave the fish a mohawk :))) thanks for the content though. much love

  • @djangoworldwide7925
    @djangoworldwide7925 Před rokem +1

    Evals and Evecs are everywhere in DS

  • @NishantAroraarora007
    @NishantAroraarora007 Před 3 lety +1

    I have a quick question, at 1:01 you mention that lambda is such that it is a real number, can't this be extended to imaginary numbers as well? Btw, Thanks for you great work!

  • @matthewchunk3689
    @matthewchunk3689 Před 4 lety +2

    2:51 Thank you. A teacher who melds big pictures with equations

  • @nandakumarcheiro
    @nandakumarcheiro Před 3 lety

    In aerodrums screens eigen vector and eigenvalues for different landing planes may be manipulated with out collision by having graphics accordingly correct? That might have been a better explanation.

  • @beaudjangles
    @beaudjangles Před 4 lety

    Fantastic

  • @user-or7ji5hv8y
    @user-or7ji5hv8y Před 3 lety

    Does matrix A have to be square?

  • @saadelmadafri8050
    @saadelmadafri8050 Před 4 lety +1

    great fish !

  • @paingzinkyaw331
    @paingzinkyaw331 Před 3 lety +1

    I just subscribed!

  • @nicholasnelson1005
    @nicholasnelson1005 Před rokem

    Didn't really go over finding the Eigenvector 😕 just solved the system of equations and left it be.

  • @neurite001
    @neurite001 Před 4 lety

    Talking about fishy vectors... 8:31

  • @robertpenoyer9998
    @robertpenoyer9998 Před 3 lety

    Math and engineering classes always seem to treat Ax = λx as an abstraction. I wish someone would say at the beginning of the discussion that Ax = λx means that an eigenvector is a vector that points in the same direction after it's been operated on by A.

    • @s25412
      @s25412 Před 3 lety

      they could point in different direction though, is lamda is a negative number

    • @robertpenoyer9998
      @robertpenoyer9998 Před 3 lety

      @@s25412 My comment about direction was a generality. Of course, A might transform x so that it points in the opposite direction, but the eigenvector will point along the same line as it was pointing before being operated on by A. A scalar multiple of an eigenvector is also an eigenvector.

    • @s25412
      @s25412 Před 3 lety

      @@robertpenoyer9998 Thank you

  • @sahirbansal7027
    @sahirbansal7027 Před 4 lety

    hey, can we subtract the mean of each column from the column so as to make it zero mean before calculating the cov matrix. and in some textbooks it is divided by n-1 instead of n. why is that? Thanks

    • @neel_in_germany
      @neel_in_germany Před 4 lety

      I think because with (n-1) the estimator is unbiased...

    • @paingzinkyaw331
      @paingzinkyaw331 Před 3 lety

      It is because of the difference between "population" and "sample" if you use for population then the accuracy must be considered so that we use n-1 it's for more accuracy.

  • @reemalshanbari
    @reemalshanbari Před 3 lety

    so we always gonna use only one eigenvalues, am I right?

  • @kulbhushansingh9131
    @kulbhushansingh9131 Před 5 měsíci

    the lecture got interesting after he said, who cares!!!

  • @user-mw3lh5kk3v
    @user-mw3lh5kk3v Před 3 lety

    新たに定理を発見しました。

  • @mlaursen
    @mlaursen Před měsícem

    Why couldn’t you have been my teacher when I was studying eigenvectors. Sigh.

  • @SNSaadu1999
    @SNSaadu1999 Před 4 lety

    does Ax = LAMDA X holds for all x?

  • @nicoleluo6692
    @nicoleluo6692 Před rokem

    🌹🌹🌹

  • @user-gu9ur6hi2v
    @user-gu9ur6hi2v Před 4 dny

    I'm crying

  • @premstein16
    @premstein16 Před 4 lety

    Hi, could you please do the computation for eigen value -2 and eager to know how to plug in x1 and x2

  • @terrym2007
    @terrym2007 Před 5 měsíci

    Actually, the concepts are foundational....

  • @Ahmad_Alhasanat
    @Ahmad_Alhasanat Před 3 lety

    Wondering who hit dislike!!

  • @iwasforcedtomakethis8818
    @iwasforcedtomakethis8818 Před 3 lety +1

    wtf is aroung his neck

    • @caiobustani5223
      @caiobustani5223 Před 2 lety

      probably protection for a recently done tattoo...just guessing here since I've noticed too!