ChatGPT: Zero to Hero

Sdílet
Vložit
  • čas přidán 28. 08. 2024

Komentáře • 15

  • @Rm-no6jr
    @Rm-no6jr Před 10 měsíci

    Your channel deserves more. Thanks a lot

  • @amiralioghli8622
    @amiralioghli8622 Před 10 měsíci

    Thank you, sir, for sharing valuable information through your CZcams channel. Once again, I have a request: please create a series on how to apply Transformers to time series tasks such as anomaly detection, forecasting, or classification. Working on just one of these tasks would be sufficient for us. I have followed numerous articles, short notes, and videos regarding the application of Transformers to time series data, but it is still not clear to me. I am a beginner on this Transformer journey, and there are no useful videos available on CZcams overall.

  • @user-zt2vq8ne1l
    @user-zt2vq8ne1l Před 6 měsíci

    Thank you for sharing this valuable Content, Great Channel

  • @chrisogonas
    @chrisogonas Před 11 měsíci

    Great picture of how the GPTs work and what they are. Awesome 👍

  • @victle
    @victle Před 10 měsíci

    Mind-blowing that all of this good stuff is free. Great video!

  • @user-wc7em8kf9d
    @user-wc7em8kf9d Před 10 měsíci

    Thks mate. I love the summary around 2:30!

  • @khoshsirat
    @khoshsirat Před 4 měsíci

    Great video, but I think the "Likert scale" part around 31:00 is not correct. They scale the ratings for each rater separately, and they use those questions to detect sensitive topics and filter them.

  • @barni_7762
    @barni_7762 Před 11 měsíci

    Watching this while finetuning llama 2 :D
    I think this may be the first gpt tutorial featuring RLHF

  • @DaTruAndi
    @DaTruAndi Před 11 měsíci

    Awesome video. A few comments.
    Is the architecture for the rewards model OpenAI used actually publicly documented? Can we be sure it is a GPT model? (I believe you mention it).
    I would love a contrastive deeper look at BERT vs GPT following what you mention. You mention it when you talk about stacking them up but it could make sense to talk about the what’s and whys a bit more.
    At 43:41 you say that the table is generated anew as being the reason for different generations even for the same prompt, but wouldn’t the table be the same, just the sampling strategy would result in the table being processed resulting in a different token?
    Overall:
    Maybe as a video title “making of the tasty ChatGPT sausage” would have been better :)
    The title may set wrong expectations for folks who casually discover your great content.
    From the title alone many people may expect to become heroes in using ChatGPT.

  • @r.alexander9075
    @r.alexander9075 Před 6 měsíci

    I have a question:
    How does the GPT architecture provide outputs without an encoder? Doesnt the cross attention module need KV pairs from the encoder?

  • @sharangkulkarni1759
    @sharangkulkarni1759 Před 4 dny

    but why sft i.e. first step needed, we can do with it ofcourse it helps but not necessary right ? lets do 2 and 3 only .

  • @neetpride5919
    @neetpride5919 Před 10 měsíci

    Is there ANY open-source repo that includes a tool for manually rating the results of your own ChatGPT like in steps 2 and 3 of this video? I want to *actually* train a TNN from scratch and I have the time to do it.
    Why do so few people talk about this aspect of ChatGPT?

  • @sam_joshua_s
    @sam_joshua_s Před 11 měsíci +1

    can you make video about deepspeed coding implementation

  • @prashlovessamosa
    @prashlovessamosa Před 11 měsíci +1

    Bahi is on Steroids after 100k.