Large Language Models: How Large is Large Enough?

Sdílet
Vložit
  • čas přidán 14. 12. 2023
  • Explore IBM watsonx → ibm.biz/IBM-watsonx
    When it comes to large language models, one of the first things you may read about is the enormous size of the model data or the number of parameters. But is "bigger" always better? In this video, Kip Yego, Program Marketing Manager explains that the truthful answer is "It depends". So what are the factors that will drive this decision? Kip breaks them down and explains one-by-one how to decide.
    Get started for free on IBM Cloud → ibm.biz/sign-up-now
    Subscribe to see more videos like this in the future → ibm.biz/subscribe-now
    #llm #ai #generativeai #ml

Komentáře • 21

  • @thatdudewiththething
    @thatdudewiththething Před 7 měsíci +4

    These videos are fantastic!
    Thank you so much for making them available :D

  • @ttjordan81
    @ttjordan81 Před 7 měsíci +2

    Thank you, this is the information I was searching for. I was explaining the concept in theory to someone. The idea was to use smaller models that are trained for specific domains. By eliminating or reduce all the other domains, the model should perform better and reduce messy results.

  • @dominiquecoladon8343
    @dominiquecoladon8343 Před 7 měsíci +1

    Well done video, Get kip to more of these please.

  • @LaurenFrazier-ch4kn
    @LaurenFrazier-ch4kn Před 7 měsíci

    Great video, super informative!

  • @YvesNewman
    @YvesNewman Před 7 měsíci

    Great video Kip!
    At the moment it seems that bigger equals better. Time to change that perception accordingly

    • @sherpya
      @sherpya Před 7 měsíci

      already the trend, see mixture of experts concept

  • @donson3326
    @donson3326 Před měsícem

    0:16 🤣 You tell me.

  • @tyrojames9937
    @tyrojames9937 Před 7 měsíci

    INTERESTING. 😀

  • @7rich79
    @7rich79 Před 7 měsíci

    Thank you, that was informative. One question I have is how you determine domain specificity, and perhaps potential lost opportunity?
    For example, using financial services tasks as in your example. If you ask someone working in finance about what insights they'd be looking for, tax or perhaps transfer pricing may not be what they consider as part of their domain. However, transfer pricing and tax could have a huge impact on what finance should consider when taking decisions. How do you ensure the domain specificity is not too narrow?

    • @julioberas2106
      @julioberas2106 Před 6 měsíci

      I believe anything remotely related to the domains should be included in the training data. He didn't talk about the training data size, but I believe it should still be very big (but smaller than a general one)

  • @gjjakobsen
    @gjjakobsen Před 7 měsíci

    The MBA in me says, beyond some point, the trade-off isn't worth it. Then again, that's probably what they said about the Apollo mission.

  • @fortuneosho8137
    @fortuneosho8137 Před 7 měsíci

    It really depends

  • @nirmal7103
    @nirmal7103 Před 7 měsíci

    How can we find a domain specific models or how to train them?

    • @ttjordan81
      @ttjordan81 Před 7 měsíci +1

      I think that's the next business idea, lol... At this point, pick an industry, and create specific domain model! It's a race! Also, specific domain Vector Databases will be needed!

  • @IsaacFoster..
    @IsaacFoster.. Před 7 měsíci

    My llm's so large, it reaches almost every 1 and 0 it can write on; you can literally call it a "wipe"

  • @Alice8000
    @Alice8000 Před 4 měsíci

    5:38 THANK YOU BRO. Definitely feel more confident after hearing that.

  • @warsin8641
    @warsin8641 Před 7 měsíci

    The Bloke

  • @deathlife2414
    @deathlife2414 Před 7 měsíci

    Lets go phi. chroot chroot chroot

  • @aberobwohl
    @aberobwohl Před 7 měsíci

    I see no point whatsoever in comparing a domain specific finetuned model to a non finetuned model to draw conclusions or suggest any insights doing this.

  • @Alice8000
    @Alice8000 Před 4 měsíci

    bro you got worse handwriting than me!!! Good info though. lol

  • @TheBiffsterLife
    @TheBiffsterLife Před 5 měsíci

    Kip, that’s a very poor analogy.