mathematicalmonk
mathematicalmonk
  • 262
  • 9 897 070
(ML 2.5) Generalizations for trees (CART)
Other "impurity" quantities (entropy and Gini index), and generalizations of decision trees for classification and regression using the CART approach.
A playlist of these Machine Learning videos is available here:
czcams.com/users/my_playlists?p=D0F06AA0D2E8FFBA
zhlédnutí: 34 314

Video

(ML 2.4) Growing a classification tree (CART)
zhlédnutí 37KPřed 11 lety
How to build a decision tree for classification using the CART approach. A playlist of these Machine Learning videos is available here: czcams.com/users/my_playlists?p=D0F06AA0D2E8FFBA
(IC 5.14) Finite-precision arithmetic coding - Decoder
zhlédnutí 10KPřed 12 lety
Pseudocode for the arithmetic coding decoder, using finite-precision. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.13) Finite-precision arithmetic coding - Encoder
zhlédnutí 6KPřed 12 lety
Pseudocode for the arithmetic coding encoder, using finite-precision. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.12) Finite-precision arithmetic coding - Setup
zhlédnutí 6KPřed 12 lety
Pre-defining the quantities that will be needed in the finite-precision algorithm. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.11) Finite-precision arithmetic coding - Rescaling
zhlédnutí 6KPřed 12 lety
We integrate the rescaling operations into the infinite-precision encoder, as a precursor to the finite-precision encoder. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.10) Generalizing arithmetic coding to non-i.i.d. models
zhlédnutí 3,8KPřed 12 lety
Arithmetic coding can accommodate essentially any probabilistic model of the source, in a very natural way. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.9) Computational complexity of arithmetic coding
zhlédnutí 6KPřed 12 lety
Arithmetic coding is linear time in the length of the source message and the encoded message. Since the encoded message length is near optimal on average, the expected time is near optimal. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.8) Near optimality of arithmetic coding
zhlédnutí 3,8KPřed 12 lety
The expected encoded length of the entire message is within 2 bits of the ideal encoded length (the entropy), assuming infinite precision. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.7) Decoder for arithmetic coding (infinite-precision)
zhlédnutí 7KPřed 12 lety
Pseudocode for the arithmetic coding algorithm, assuming addition and multiplication can be done exactly (i.e. with infinite precision). Later we modify this to work with finite precision. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.4) Why the interval needs to be completely contained
zhlédnutí 6KPřed 12 lety
To ensure unique decodeability, it's necessary that the interval [a,b) contain the whole interval corresponding to the encoded binary sequence, rather than just containing the number corresponding to the binary sequence. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.6) Encoder for arithmetic coding (infinite-precision)
zhlédnutí 7KPřed 12 lety
Pseudocode for the arithmetic coding algorithm, assuming addition and multiplication can be done exactly (i.e. with infinite precision). Later we modify this to work with finite precision. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.5) Rescaling operations for arithmetic coding
zhlédnutí 8KPřed 12 lety
Certain rescaling operations are convenient for the infinite-precision algorithm, and are critical for the finite-precision algorithm. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.3) Arithmetic coding - Example #2
zhlédnutí 18KPřed 12 lety
A simple example to illustrate the basic idea of arithmetic coding. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.2) Arithmetic coding - Example #1
zhlédnutí 85KPřed 12 lety
A simple example to illustrate the basic idea of arithmetic coding. A playlist of these videos is available at: czcams.com/play/PLE125425EC837021F.html
(IC 5.1) Arithmetic coding - introduction
zhlédnutí 59KPřed 12 lety
(IC 5.1) Arithmetic coding - introduction
(IC 4.12) Optimality of Huffman codes (part 7) - existence
zhlédnutí 1,8KPřed 12 lety
(IC 4.12) Optimality of Huffman codes (part 7) - existence
(IC 4.13) Not every optimal prefix code is Huffman
zhlédnutí 3,2KPřed 12 lety
(IC 4.13) Not every optimal prefix code is Huffman
(IC 4.11) Optimality of Huffman codes (part 6) - induction
zhlédnutí 2,4KPřed 12 lety
(IC 4.11) Optimality of Huffman codes (part 6) - induction
(IC 4.10) Optimality of Huffman codes (part 5) - extension lemma
zhlédnutí 2,1KPřed 12 lety
(IC 4.10) Optimality of Huffman codes (part 5) - extension lemma
(IC 4.9) Optimality of Huffman codes (part 4) - extension and contraction
zhlédnutí 2KPřed 12 lety
(IC 4.9) Optimality of Huffman codes (part 4) - extension and contraction
(IC 4.8) Optimality of Huffman codes (part 3) - sibling codes
zhlédnutí 1,8KPřed 12 lety
(IC 4.8) Optimality of Huffman codes (part 3) - sibling codes
(IC 4.7) Optimality of Huffman codes (part 2) - weak siblings
zhlédnutí 2,8KPřed 12 lety
(IC 4.7) Optimality of Huffman codes (part 2) - weak siblings
(IC 4.6) Optimality of Huffman codes (part 1) - inverse ordering
zhlédnutí 6KPřed 12 lety
(IC 4.6) Optimality of Huffman codes (part 1) - inverse ordering
(IC 4.5) An issue with Huffman coding
zhlédnutí 3,9KPřed 12 lety
(IC 4.5) An issue with Huffman coding
(IC 4.4) Weighted minimization with Huffman coding
zhlédnutí 4KPřed 12 lety
(IC 4.4) Weighted minimization with Huffman coding
(IC 4.3) B-ary Huffman codes
zhlédnutí 10KPřed 12 lety
(IC 4.3) B-ary Huffman codes
(IC 4.2) Huffman coding - more examples
zhlédnutí 20KPřed 12 lety
(IC 4.2) Huffman coding - more examples
(IC 4.1) Huffman coding - introduction and example
zhlédnutí 119KPřed 12 lety
(IC 4.1) Huffman coding - introduction and example
(IC 3.10) Relative entropy as the mismatch inefficiency
zhlédnutí 5KPřed 12 lety
(IC 3.10) Relative entropy as the mismatch inefficiency

Komentáře

  • @yethusithole4695
    @yethusithole4695 Před 8 dny

    Practical to follow, thanks.

  • @PRiKoL1ST1
    @PRiKoL1ST1 Před 19 dny

    Why we need "smaller" sigma algebras and not use always powerset as sigma algebra?

  • @sintiahernandez3421
    @sintiahernandez3421 Před 26 dny

    ❤ que se encuentra en el ❤❤❤😂❤❤❤❤❤😂llppku 😂 mmslllllMMKkKKKKkzkzkzzkdkskskskskkjkkksskkskkkksksksjJjJKKakSaLaalaeewee esa🇧🇴🚩🇧🇴🇧🇳🇧🇾🇧🇳🇧🇳🇧🇷🇧🇾🇧🇾🇧🇾🇧🇾🇧🇾🇧🇾🇧🇷🇧🇪🇧🇪🇧🇪🇧🇪🇦🇺🟧🚂🚋🚋🚅🚃🏹 m

  • @cagataykaydr3015
    @cagataykaydr3015 Před měsícem

    Man, 12 years ago you did a better job than recent articles and lessons. I swear I understood the algorithm in less than 5 minutes, which I'm looking for descriptive contents to understand for days. I can't thank you enough, I'm trying to write a new implementation of an image format! If I become Linus Torvalds someday, I'll make sure people will now you helped me a lot haha!

  • @isaac5990
    @isaac5990 Před měsícem

    why do we need the probability mass function? wouldn't the steps be the same for encoding and decoding for an even distribution?

  • @schoenwettersl
    @schoenwettersl Před 2 měsíci

    very nice video. thank you so much!

  • @lancezhang892
    @lancezhang892 Před 3 měsíci

    Hello ,I have one question,why are mu and theta not hidden variables?

  • @sina48289
    @sina48289 Před 3 měsíci

    I pressed the like button only for the sounds you created while drawing the lines

  • @HongzhiSun
    @HongzhiSun Před 3 měsíci

    The whole series of the lectures are so great and I like them very very much! As you mentioned in one of your lecture, there are three big theorems in information theory: source coding theorem, rate distortion theorem and channel coding theorem. It is my best wish that you can provide video lectures for the contents of the channel coding and rate distortion theorems. Thank you so much for the wonderful lectures!

  • @user-xr9br3jj9q
    @user-xr9br3jj9q Před 3 měsíci

    I have returned to this derivation several times over that past many years. This is a clear calculation of the gradient of the log likelihood function for logistic regression. Thank you!

  • @skadoosh296
    @skadoosh296 Před 3 měsíci

    if C_j is a vector of counts of each category, what does alpha mean?

  • @Kayi_Alp07
    @Kayi_Alp07 Před 3 měsíci

    Sir your tutor is amazing. But many students are not benefiting from it. So for you to help them find your tutor please first add profile to your youtube channel and for your videos add thumbnails on it.

  • @provocateach
    @provocateach Před 4 měsíci

    This is like the best algorithm ever designed. So many nice properties, from being able to derive the length of the source string from the interval while decoding, to this elegant entropy property. Utterly beautiful.

  • @Nikifuj908
    @Nikifuj908 Před 4 měsíci

    You had me at “this is overhead”. Number version is bloat, just like Microsoft telemetry

  • @alicetang8009
    @alicetang8009 Před 4 měsíci

    Sorry the way you pronounce theta and data really confuses me 😶‍🌫

  • @lordcasper3357
    @lordcasper3357 Před 4 měsíci

    this is so good man

  • @merterdem9345
    @merterdem9345 Před 4 měsíci

    why is the differential equal to the term inside the infinte sum?

  • @ebubechukwuenwerem9145
    @ebubechukwuenwerem9145 Před 4 měsíci

    Just one example with real values and data would have helped alot

  • @andrashorvath2411
    @andrashorvath2411 Před 5 měsíci

    Fantastic presentation, easy to follow and giving great intuition, thank you.

  • @ThePlacehole
    @ThePlacehole Před 6 měsíci

    It's videos like these that make you wish youtube had a 3x playback speed...

  • @cypherecon5989
    @cypherecon5989 Před 6 měsíci

    The first column of the desing matrix should contain only 1s, shouldnt it?

  • @Im_a_weirdo666
    @Im_a_weirdo666 Před 7 měsíci

    What did he say

  • @mohsinjunaid8454
    @mohsinjunaid8454 Před 7 měsíci

    super intuitive Thanks

  • @asmaaasmaa2395
    @asmaaasmaa2395 Před 7 měsíci

    Why B=2?? In example a!

  • @gennadyshutkov1912
    @gennadyshutkov1912 Před 7 měsíci

    Hello, I have watched the whole series of lectures about arithmetic encoding and I'm very thrilled about this algorithm and your way of explaining things. You are a great professor and mathematician. Thank you for your lectures! Unfortunately my finite precision implementation works incorrectly on large input data (maybe due to round off error). So maybe you cold share your implementation of the algorithm. It would be a great help! Thank you in advance!

    • @HongzhiSun
      @HongzhiSun Před 3 měsíci

      Set the probability distribution so that each p(x_i) has the form 1/2**m with m>=0. It seems the algorithm work for long sequence in this case. I guess this can get rid of the round off error issue but I haven't proved it yet. In addition, any places like "< half" or ">half" or ">quarter and <3*quarter" should use <= and >= instead.

  • @amirreza08
    @amirreza08 Před 8 měsíci

    It was one of the best explanations, so informative and helpful. Thank you!

  • @pedrocolangelo5844
    @pedrocolangelo5844 Před 8 měsíci

    Your enthusiastic way of teaching is so inspiring. Thank you for sharing this great video!

  • @Celdorsc2
    @Celdorsc2 Před 8 měsíci

    With regrets I have to say, these are hardly good videos to learn this topic from. It is not the first video that makes me completely confused about where things come from or why certain terms appear in the main or side equations. It's not well organised IMO. I have made a few attempts to come back to the playlist and give it a another go. First I thought I was missing basics. Now, I am convinced the videos are problems. The idea is great, the effort is still appreciated, but sadly it's hard to notice or understand things if they are not explained.

  • @adelecutler1195
    @adelecutler1195 Před 9 měsíci

    You cite a study but you do not cite the inventor of Random Forests, Leo Breiman.

  • @damoonrobatian9371
    @damoonrobatian9371 Před 9 měsíci

    I like these videos but this one was very confusing!

  • @denemesistemi
    @denemesistemi Před 9 měsíci

    yes this seems good, i agree on that topic

  • @dineshds5124
    @dineshds5124 Před 9 měsíci

    ありがとう、先生👍

  • @philip2.042
    @philip2.042 Před 9 měsíci

    Good explanation of Poisson and Exponential rv

  • @leukosnanos
    @leukosnanos Před 10 měsíci

    I strongly disagree about the naming you are talking at the beginning. Logistic regression ultimately serves as a classification method, but it fits a logistic (or sigmoid) line to the data so it could be thought as a regression process.

  • @stokesarisnet
    @stokesarisnet Před 10 měsíci

    Didn't quite follow the step where "the joint was summed over c". What does that mean?

  • @rebeccacarroll8385
    @rebeccacarroll8385 Před 10 měsíci

    A little dissapointed that Arianna Rosenbluth wasn't included in the list of creators. She was also the first to use the method. Great explanation of MCMC!

  • @QT-yt4db
    @QT-yt4db Před 10 měsíci

    I like this name "mathematicalmonk" a lot. Because I think you must have a monk's heart to love mathematics...

  • @BillHaug
    @BillHaug Před 10 měsíci

    thank you

  • @bhardwajsatyam
    @bhardwajsatyam Před 10 měsíci

    Speaking from personal experience, you explain these concepts so much better than some professor with an h-index of 60 at a well-known university.

  • @michellemaps8103
    @michellemaps8103 Před 11 měsíci

    sir, may you please make a video for gamma random variables expected values

  • @pan19682
    @pan19682 Před 11 měsíci

    It was perfect exposition of the subject many thanks

  • @JReuben111
    @JReuben111 Před 11 měsíci

    This vid makes references to watching other vids before proceeding to the next one !

  • @JReuben111
    @JReuben111 Před 11 měsíci

    I must admit, I got lost in the details here ... 😢

  • @MachineCarlo
    @MachineCarlo Před 11 měsíci

    for the continuous case, of what use is it?

  • @ikhlasulkamal5245
    @ikhlasulkamal5245 Před rokem

    Can you please come back to youtube sir, your videos are great. Sadly im 10 years late to see your videos

  • @khaledsrrr
    @khaledsrrr Před rokem

    thanks for ur efforts

  • @user-pg8xw5qm1e
    @user-pg8xw5qm1e Před rokem

    Thanks for the video, but in fact you don't need to factor both C and D, just factoring one of them is enough! As Schur did in his paper (1911).

  • @jiaqint961
    @jiaqint961 Před rokem

    Thanks so much, after looking at many explanation on this topic, I finally got it after watching this video.

  • @fsaldan1
    @fsaldan1 Před rokem

    In the shell game X and Y are not independent like in your equations. In the shell game there us an additional equation: X + Y + Z = 1.

  • @KAUSTUBHCHAKRABORTY

    @mathematicalmonk at 1:26 you write P(A, B, C) =P(A|C) P(B|C) P(C). How do you prove it?