Llama 3 BREAKS the industry !!! | Llama3 fully Tested

Sdílet
Vložit
  • čas přidán 11. 07. 2024
  • FULLY Tested Llama 3, the flagship model from Meta.
    Benchmark of GPT-4 vs GPT-4 Turbo vs Llama 3.
    _______ 👇 Links 👇 _______
    lmstudio.ai/
    🤝 Discord: / discord
    💼 𝗟𝗶𝗻𝗸𝗲𝗱𝗜𝗻: / reda-marzouk-rpa
    📸 𝗜𝗻𝘀𝘁𝗮𝗴𝗿𝗮𝗺: / redamarzouk.rpa
    🤖 𝗬𝗼𝘂𝗧𝘂𝗯𝗲: / @redamarzouk
    www.automation-campus.com/
    _______ 👇 Content👇 _______
    00:00 Introduction to Llama3
    00:30 All you need to know about Llama3
    02:31 Math and Reasoning 1
    02:45 Math and Reasoning 2
    02:59 Math and Reasoning 3
    03:17 Math and Reasoning 4
    04:06 Logic and Reasoning 1
    05:04 Logic and Reasoning 2
    05:36 Logic and Reasoning 3
    06:01 Logic and Reasoning 4
    07:09 Logic and Reasoning 5
    07:31 Logic and Reasoning 6
    08:37 Coding 1
    09:28 Coding 2 (minesweeper game)
    11:41 Dilemma
    13:19 Final Results (Tiny_Benchmark)
  • Věda a technologie

Komentáře • 11

  • @leeme179
    @leeme179 Před 2 měsíci +3

    Amazing video, Llama 3 seems way better than expected, I hope Meta would release 8 B for code, and 8 B for Python, I don't think Meta would do something like 400 B for code but that would crazy if they did, even as is Llama 3 would help the community a lot.

    • @redamarzouk
      @redamarzouk  Před 2 měsíci +1

      Thank you.
      Yeah I find it hard to imagine them making the 400B open, it would change the open source community.
      Also an 8B for code will be a beast for local use.

  • @macoson
    @macoson Před 2 měsíci

    Great set of benchmark questions! I had some surprising results when I tried out the question about Sally's sisters. Oddly enough, Llama 3-70B missed it, while Llama 3-8B nailed it every time. Both GPT-4 Turbo and GPT-3.5 Turbo got it right as well. Interestingly, Clade 3 Opus got it mixed up and said she has two sisters. Really intriguing to see the differences!

    • @redamarzouk
      @redamarzouk  Před 2 měsíci

      It's very surprising for Llama3-8B to answer correctly when 70B can't.
      I just tested it again in another platform (Meta.ai) where we have the 70B model and it's worked just fine.
      Where did you test the models? was the 8B llama hosted locally?

  • @RocketLR
    @RocketLR Před 2 měsíci +2

    The 7B models are pretty shitty tbh :(
    They are fast but wrong most of the times.

    • @redamarzouk
      @redamarzouk  Před 2 měsíci +1

      Yeah I used llama3-70B in the video and I tried the 8B one in my machine as well, sadly the smaller models are no match to the bigger ones no matter how much fine tuning we make.
      this is why it's important to have specialized models (Llama3-8B coding for example) and that model will be good at a specific thing

  • @kamalkamals
    @kamalkamals Před měsícem

    nop it s not better that GPT

    • @redamarzouk
      @redamarzouk  Před měsícem

      you're right now it's not, these models are beating each other like there is no tomorrow, to this date GPT-4o is the one at the top.

    • @kamalkamals
      @kamalkamals Před měsícem

      @@redamarzouk before gpt 4 omni, gpt 4 turbo still better, the only best point with llama is free model :)